diff --git contrib/src/test/results/clientpositive/fileformat_base64.q.out contrib/src/test/results/clientpositive/fileformat_base64.q.out index 2b3ae49..6c941de 100644 --- contrib/src/test/results/clientpositive/fileformat_base64.q.out +++ contrib/src/test/results/clientpositive/fileformat_base64.q.out @@ -44,10 +44,8 @@ PREHOOK: query: DESCRIBE EXTENDED base64_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED base64_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None +key int +value string #### A masked pattern was here #### PREHOOK: query: FROM src diff --git contrib/src/test/results/clientpositive/serde_s3.q.out contrib/src/test/results/clientpositive/serde_s3.q.out index 8f4335a..e1ecc4b 100644 --- contrib/src/test/results/clientpositive/serde_s3.q.out +++ contrib/src/test/results/clientpositive/serde_s3.q.out @@ -15,25 +15,23 @@ PREHOOK: query: DESCRIBE s3log PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE s3log POSTHOOK: type: DESCTABLE -# col_name data_type comment - #### A masked pattern was here #### -bucketname string from deserializer -rdatetime string from deserializer -rip string from deserializer -requester string from deserializer -requestid string from deserializer -operation string from deserializer -rkey string from deserializer -requesturi string from deserializer -httpstatus int from deserializer -errorcode string from deserializer -bytessent int from deserializer -objsize int from deserializer -totaltime int from deserializer -turnaroundtime int from deserializer -referer string from deserializer -useragent string from deserializer +bucketname string from deserializer +rdatetime string from deserializer +rip string from deserializer +requester string from deserializer +requestid string from deserializer +operation string from deserializer +rkey string from deserializer +requesturi string from deserializer +httpstatus int from deserializer +errorcode string from deserializer +bytessent int from deserializer +objsize int from deserializer +totaltime int from deserializer +turnaroundtime int from deserializer +referer string from deserializer +useragent string from deserializer PREHOOK: query: LOAD DATA LOCAL INPATH '../contrib/data/files/s3.log' INTO TABLE s3log PREHOOK: type: LOAD PREHOOK: Output: default@s3log diff --git hbase-handler/src/test/results/positive/hbase_queries.q.out hbase-handler/src/test/results/positive/hbase_queries.q.out index ae961c4..0d74717 100644 --- hbase-handler/src/test/results/positive/hbase_queries.q.out +++ hbase-handler/src/test/results/positive/hbase_queries.q.out @@ -17,10 +17,8 @@ PREHOOK: query: DESCRIBE EXTENDED hbase_table_1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED hbase_table_1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int from deserializer -value string from deserializer +key int from deserializer +value string from deserializer #### A masked pattern was here #### PREHOOK: query: select * from hbase_table_1 diff --git hwi/src/test/org/apache/hadoop/hive/hwi/TestHWISessionManager.java hwi/src/test/org/apache/hadoop/hive/hwi/TestHWISessionManager.java index adcbd83..8f05dc8 100644 --- hwi/src/test/org/apache/hadoop/hive/hwi/TestHWISessionManager.java +++ hwi/src/test/org/apache/hadoop/hive/hwi/TestHWISessionManager.java @@ -121,15 +121,10 @@ public final void testHiveDriver() throws Exception { ArrayList> searchBlockRes = searchItem.getResultBucket(); - // "describe [table_name]" result format - // first line should be format name: - // "# col_name data_type comment" - // second line is empty - // the following lines contain the values - String resLine = searchBlockRes.get(0).get(2); + String resLine = searchBlockRes.get(0).get(0); assertEquals(true, resLine.contains("key")); assertEquals(true, resLine.contains("int")); - String resLine2 = searchBlockRes.get(0).get(3); + String resLine2 = searchBlockRes.get(0).get(1); assertEquals(true, resLine2.contains("value")); assertEquals(true, resLine2.contains("string")); diff --git jdbc/src/test/org/apache/hadoop/hive/jdbc/TestJdbcDriver.java jdbc/src/test/org/apache/hadoop/hive/jdbc/TestJdbcDriver.java index 639418f..acaead7 100644 --- jdbc/src/test/org/apache/hadoop/hive/jdbc/TestJdbcDriver.java +++ jdbc/src/test/org/apache/hadoop/hive/jdbc/TestJdbcDriver.java @@ -767,23 +767,17 @@ public void testDescribeTable() throws SQLException { ResultSet res = stmt.executeQuery("describe " + tableName); - // "describe [table_name]" result format - // first line should be format name: - // "# col_name data_type comment" - // second line is empty - // the following lines contain the values res.next(); - assertEquals(true, res.getString(1).contains("col_name")); - assertEquals(true, res.getString(2).contains("data_type")); - assertEquals(true, res.getString(3).contains("comment")); + assertEquals("Column name 'under_col' not found", "under_col", res.getString(1)); + assertEquals("Column type 'under_col' for column under_col not found", "int", res + .getString(2)); res.next(); - res.next(); - assertEquals(true, res.getString(1).contains("under_col")); - assertEquals(true, res.getString(2).contains("int")); - res.next(); - assertEquals(true, res.getString(1).contains("value")); - assertEquals(true, res.getString(2).contains("string")); + assertEquals("Column name 'value' not found", "value", res.getString(1)); + assertEquals("Column type 'string' for column key not found", "string", res + .getString(2)); + assertFalse("More results found than expected", res.next()); + } public void testDatabaseMetaData() throws SQLException { diff --git ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java index 147febb..ac6bc80 100644 --- ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java +++ ql/src/java/org/apache/hadoop/hive/ql/exec/DDLTask.java @@ -2323,7 +2323,7 @@ public int showColumns(Hive db, ShowColumnsDesc showCols) List cols = table.getCols(); cols.addAll(table.getPartCols()); - outStream.writeBytes(MetaDataFormatUtils.getAllColumnsInformation(cols)); + outStream.writeBytes(MetaDataFormatUtils.displayColsUnformatted(cols)); ((FSDataOutputStream) outStream).close(); outStream = null; } catch (IOException e) { diff --git ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/MetaDataFormatUtils.java ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/MetaDataFormatUtils.java index bc046aa..15d5b6a 100644 --- ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/MetaDataFormatUtils.java +++ ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/MetaDataFormatUtils.java @@ -69,6 +69,23 @@ public static String getAllColumnsInformation(List cols) { return columnInformation.toString(); } + /* + Displaying columns unformatted for backward compatibility. + */ + public static String displayColsUnformatted(List cols) { + StringBuilder colBuffer = new StringBuilder(DEFAULT_STRINGBUILDER_SIZE); + for (FieldSchema col : cols) { + colBuffer.append(col.getName()); + colBuffer.append(FIELD_DELIM); + colBuffer.append(col.getType()); + colBuffer.append(FIELD_DELIM); + colBuffer.append(col.getComment() == null ? "" : col.getComment()); + colBuffer.append(LINE_DELIM); + } + return colBuffer.toString(); + } + + public static String getAllColumnsInformation(List cols, List partCols) { StringBuilder columnInformation = new StringBuilder(DEFAULT_STRINGBUILDER_SIZE); formatColumnsHeader(columnInformation); diff --git ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/TextMetaDataFormatter.java ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/TextMetaDataFormatter.java index c0ba66f..03b2667 100644 --- ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/TextMetaDataFormatter.java +++ ql/src/java/org/apache/hadoop/hive/ql/metadata/formatting/TextMetaDataFormatter.java @@ -139,16 +139,25 @@ public void describeTable(DataOutputStream outStream, throws HiveException { try { if (colPath.equals(tableName)) { - List partCols = tbl.isPartitioned() ? tbl.getPartCols() : null; - outStream.writeBytes( - isPretty ? - MetaDataPrettyFormatUtils.getAllColumnsInformation( - cols, partCols, prettyOutputNumCols) - : - MetaDataFormatUtils.getAllColumnsInformation(cols, partCols) - ); + if (!isFormatted) { + outStream.writeBytes(MetaDataFormatUtils.displayColsUnformatted(cols)); + } + else { + List partCols = tbl.isPartitioned() ? tbl.getPartCols() : null; + outStream.writeBytes( + isPretty ? + MetaDataPrettyFormatUtils.getAllColumnsInformation( + cols, partCols, prettyOutputNumCols) + : + MetaDataFormatUtils.getAllColumnsInformation(cols, partCols) + ); + } } else { - outStream.writeBytes(MetaDataFormatUtils.getAllColumnsInformation(cols)); + if(!isFormatted){ + outStream.writeBytes(MetaDataFormatUtils.displayColsUnformatted(cols)); + }else { + outStream.writeBytes(MetaDataFormatUtils.getAllColumnsInformation(cols)); + } } if (tableName.equals(colPath)) { @@ -176,7 +185,13 @@ public void describeTable(DataOutputStream outStream, // show table information outStream.writeBytes("Detailed Table Information"); outStream.write(separator); - outStream.writeBytes(tbl.getTTable().toString()); + + //get the parameters map sorted in table copy + org.apache.hadoop.hive.metastore.api.Table tcopy = tbl.getTTable().deepCopy(); + tcopy.setParameters(new TreeMap(tcopy.getParameters())); + + outStream.writeBytes(tcopy.toString()); + outStream.write(separator); outStream.write(terminator); } diff --git ql/src/test/queries/clientpositive/describe_comment_indent.q ql/src/test/queries/clientpositive/describe_comment_indent.q deleted file mode 100644 index 310b694..0000000 --- ql/src/test/queries/clientpositive/describe_comment_indent.q +++ /dev/null @@ -1,14 +0,0 @@ --- test comment indent processing for multi-line comments - -CREATE TABLE test_table( - col1 INT COMMENT 'col1 one line comment', - col2 STRING COMMENT 'col2 -two lines comment', - col3 STRING COMMENT 'col3 -three lines -comment') -COMMENT 'table comment -two lines'; - -DESCRIBE test_table; -DESCRIBE FORMATTED test_table; diff --git ql/src/test/results/beelinepositive/bucket_groupby.q.out ql/src/test/results/beelinepositive/bucket_groupby.q.out index 6dacd37..f8a1cd2 100644 --- ql/src/test/results/beelinepositive/bucket_groupby.q.out +++ ql/src/test/results/beelinepositive/bucket_groupby.q.out @@ -488,32 +488,27 @@ select key, count(1) from (select value as key, key as value from clustergroupby ' expr: value' ' type: string' ' outputColumnNames: _col0' -' Select Operator' -' expressions:' +' Group By Operator' +' aggregations:' +' expr: count(1)' +' bucketGroup: false' +' keys:' ' expr: _col0' ' type: string' -' outputColumnNames: _col0' -' Group By Operator' -' aggregations:' -' expr: count(1)' -' bucketGroup: false' -' keys:' +' mode: hash' +' outputColumnNames: _col0, _col1' +' Reduce Output Operator' +' key expressions:' ' expr: _col0' ' type: string' -' mode: hash' -' outputColumnNames: _col0, _col1' -' Reduce Output Operator' -' key expressions:' -' expr: _col0' -' type: string' -' sort order: +' -' Map-reduce partition columns:' -' expr: _col0' -' type: string' -' tag: -1' -' value expressions:' -' expr: _col1' -' type: bigint' +' sort order: +' +' Map-reduce partition columns:' +' expr: _col0' +' type: string' +' tag: -1' +' value expressions:' +' expr: _col1' +' type: bigint' ' Reduce Operator Tree:' ' Group By Operator' ' aggregations:' @@ -544,7 +539,7 @@ select key, count(1) from (select value as key, key as value from clustergroupby ' limit: 10' '' '' -75 rows selected +70 rows selected >>> select key, count(1) from (select value as key, key as value from clustergroupby where ds='101')subq group by key limit 10; 'key','_c1' 'val_0','3' diff --git ql/src/test/results/beelinepositive/bucket_map_join_1.q.out ql/src/test/results/beelinepositive/bucket_map_join_1.q.out index 367fb1f..ab2489b 100644 --- ql/src/test/results/beelinepositive/bucket_map_join_1.q.out +++ ql/src/test/results/beelinepositive/bucket_map_join_1.q.out @@ -26,8 +26,6 @@ No rows affected No rows affected >>> set hive.optimize.bucketmapjoin.sortedmerge = true; No rows affected ->>> set hive.input.format = org.apache.hadoop.hive.ql.io.BucketizedHiveInputFormat; -No rows affected >>> >>> -- The tables are bucketed in same columns in different order, >>> -- but sorted in different column orders @@ -40,13 +38,12 @@ select /*+ mapjoin(b) */ count(*) from table1 a join table2 b on a.key=b.key and ' (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME table1) a) (TOK_TABREF (TOK_TABNAME table2) b) (and (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL b) key)) (= (. (TOK_TABLE_OR_COL a) value) (. (TOK_TABLE_OR_COL b) value))))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_MAPJOIN (TOK_HINTARGLIST b))) (TOK_SELEXPR (TOK_FUNCTIONSTAR count)))))' '' 'STAGE DEPENDENCIES:' -' Stage-4 is a root stage' -' Stage-1 depends on stages: Stage-4' -' Stage-2 depends on stages: Stage-1' +' Stage-3 is a root stage' +' Stage-1 depends on stages: Stage-3' ' Stage-0 is a root stage' '' 'STAGE PLANS:' -' Stage: Stage-4' +' Stage: Stage-3' ' Map Reduce Local Work' ' Alias -> Map Local Tables:' ' b ' @@ -85,21 +82,19 @@ select /*+ mapjoin(b) */ count(*) from table1 a join table2 b on a.key=b.key and ' 0 [Column[key], Column[value]]' ' 1 [Column[key], Column[value]]' ' Position of Big Table: 0' -' File Output Operator' -' compressed: false' -' GlobalTableId: 0' -' directory: file:!!{hive.exec.local.scratchdir}!!' -' NumFilesPerFileSink: 1' -' table:' -' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' -' properties:' -' columns ' -' columns.types ' -' escape.delim \' -' TotalFiles: 1' -' GatherStats: false' -' MultiFileSpray: false' +' Select Operator' +' Group By Operator' +' aggregations:' +' expr: count()' +' bucketGroup: false' +' mode: hash' +' outputColumnNames: _col0' +' Reduce Output Operator' +' sort order: ' +' tag: -1' +' value expressions:' +' expr: _col0' +' type: bigint' ' Local Work:' ' Map Reduce Local Work' ' Needs Tagging: false' @@ -156,47 +151,6 @@ select /*+ mapjoin(b) */ count(*) from table1 a join table2 b on a.key=b.key and ' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' ' name: bucket_map_join_1.table1' ' name: bucket_map_join_1.table1' -' Truncated Path -> Alias:' -' /bucket_map_join_1.db/table1 [a]' -'' -' Stage: Stage-2' -' Map Reduce' -' Alias -> Map Operator Tree:' -' file:!!{hive.exec.local.scratchdir}!! ' -' Select Operator' -' Select Operator' -' Group By Operator' -' aggregations:' -' expr: count()' -' bucketGroup: false' -' mode: hash' -' outputColumnNames: _col0' -' Reduce Output Operator' -' sort order: ' -' tag: -1' -' value expressions:' -' expr: _col0' -' type: bigint' -' Needs Tagging: false' -' Path -> Alias:' -' file:!!{hive.exec.local.scratchdir}!! [file:!!{hive.exec.local.scratchdir}!!]' -' Path -> Partition:' -' file:!!{hive.exec.local.scratchdir}!! ' -' Partition' -' base file name: -mr-10002' -' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' -' properties:' -' columns ' -' columns.types ' -' escape.delim \' -' ' -' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' -' properties:' -' columns ' -' columns.types ' -' escape.delim \' ' Reduce Operator Tree:' ' Group By Operator' ' aggregations:' @@ -227,14 +181,14 @@ select /*+ mapjoin(b) */ count(*) from table1 a join table2 b on a.key=b.key and ' GatherStats: false' ' MultiFileSpray: false' ' Truncated Path -> Alias:' -' file:!!{hive.exec.local.scratchdir}!! [file:!!{hive.exec.local.scratchdir}!!]' +' /bucket_map_join_1.db/table1 [a]' '' ' Stage: Stage-0' ' Fetch Operator' ' limit: -1' '' '' -198 rows selected +154 rows selected >>> >>> select /*+ mapjoin(b) */ count(*) from table1 a join table2 b on a.key=b.key and a.value=b.value; '_c1' diff --git ql/src/test/results/beelinepositive/bucketcontext_1.q.out ql/src/test/results/beelinepositive/bucketcontext_1.q.out index 977596c..d2ed6da 100644 --- ql/src/test/results/beelinepositive/bucketcontext_1.q.out +++ ql/src/test/results/beelinepositive/bucketcontext_1.q.out @@ -36,13 +36,12 @@ No rows affected ' (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME bucket_small) a) (TOK_TABREF (TOK_TABNAME bucket_big) b) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL b) key)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_MAPJOIN (TOK_HINTARGLIST a))) (TOK_SELEXPR (TOK_FUNCTIONSTAR count)))))' '' 'STAGE DEPENDENCIES:' -' Stage-4 is a root stage' -' Stage-1 depends on stages: Stage-4' -' Stage-2 depends on stages: Stage-1' +' Stage-3 is a root stage' +' Stage-1 depends on stages: Stage-3' ' Stage-0 is a root stage' '' 'STAGE PLANS:' -' Stage: Stage-4' +' Stage: Stage-3' ' Map Reduce Local Work' ' Alias -> Map Local Tables:' ' a ' @@ -95,21 +94,20 @@ No rows affected ' 0 [Column[key]]' ' 1 [Column[key]]' ' Position of Big Table: 1' -' File Output Operator' -' compressed: false' -' GlobalTableId: 0' -' directory: file:!!{hive.exec.local.scratchdir}!!' -' NumFilesPerFileSink: 1' -' table:' -' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' -' properties:' -' columns ' -' columns.types ' -' escape.delim \' -' TotalFiles: 1' -' GatherStats: false' -' MultiFileSpray: false' +' BucketMapJoin: true' +' Select Operator' +' Group By Operator' +' aggregations:' +' expr: count()' +' bucketGroup: false' +' mode: hash' +' outputColumnNames: _col0' +' Reduce Output Operator' +' sort order: ' +' tag: -1' +' value expressions:' +' expr: _col0' +' type: bigint' ' Local Work:' ' Map Reduce Local Work' ' Needs Tagging: false' @@ -125,7 +123,6 @@ No rows affected ' partition values:' ' ds 2008-04-08' ' properties:' -' SORTBUCKETCOLSPREFIX TRUE' ' bucket_count 4' ' bucket_field_name key' ' columns key,value' @@ -135,7 +132,6 @@ No rows affected ' location !!{hive.metastore.warehouse.dir}!!/bucketcontext_1.db/bucket_big/ds=2008-04-08' ' name bucketcontext_1.bucket_big' ' numFiles 4' -' numPartitions 2' ' numRows 0' ' partition_columns ds' ' rawDataSize 0' @@ -179,7 +175,6 @@ No rows affected ' partition values:' ' ds 2008-04-09' ' properties:' -' SORTBUCKETCOLSPREFIX TRUE' ' bucket_count 4' ' bucket_field_name key' ' columns key,value' @@ -189,7 +184,6 @@ No rows affected ' location !!{hive.metastore.warehouse.dir}!!/bucketcontext_1.db/bucket_big/ds=2008-04-09' ' name bucketcontext_1.bucket_big' ' numFiles 4' -' numPartitions 2' ' numRows 0' ' partition_columns ds' ' rawDataSize 0' @@ -225,48 +219,6 @@ No rows affected ' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' ' name: bucketcontext_1.bucket_big' ' name: bucketcontext_1.bucket_big' -' Truncated Path -> Alias:' -' /bucketcontext_1.db/bucket_big/ds=2008-04-08 [b]' -' /bucketcontext_1.db/bucket_big/ds=2008-04-09 [b]' -'' -' Stage: Stage-2' -' Map Reduce' -' Alias -> Map Operator Tree:' -' file:!!{hive.exec.local.scratchdir}!! ' -' Select Operator' -' Select Operator' -' Group By Operator' -' aggregations:' -' expr: count()' -' bucketGroup: false' -' mode: hash' -' outputColumnNames: _col0' -' Reduce Output Operator' -' sort order: ' -' tag: -1' -' value expressions:' -' expr: _col0' -' type: bigint' -' Needs Tagging: false' -' Path -> Alias:' -' file:!!{hive.exec.local.scratchdir}!! [file:!!{hive.exec.local.scratchdir}!!]' -' Path -> Partition:' -' file:!!{hive.exec.local.scratchdir}!! ' -' Partition' -' base file name: -mr-10002' -' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' -' properties:' -' columns ' -' columns.types ' -' escape.delim \' -' ' -' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' -' properties:' -' columns ' -' columns.types ' -' escape.delim \' ' Reduce Operator Tree:' ' Group By Operator' ' aggregations:' @@ -297,14 +249,15 @@ No rows affected ' GatherStats: false' ' MultiFileSpray: false' ' Truncated Path -> Alias:' -' file:!!{hive.exec.local.scratchdir}!! [file:!!{hive.exec.local.scratchdir}!!]' +' /bucketcontext_1.db/bucket_big/ds=2008-04-08 [b]' +' /bucketcontext_1.db/bucket_big/ds=2008-04-09 [b]' '' ' Stage: Stage-0' ' Fetch Operator' ' limit: -1' '' '' -272 rows selected +225 rows selected >>> select /* + MAPJOIN(a) */ count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key; '_c1' '928' @@ -319,7 +272,6 @@ No rows affected '' 'STAGE DEPENDENCIES:' ' Stage-1 is a root stage' -' Stage-2 depends on stages: Stage-1' ' Stage-0 is a root stage' '' 'STAGE PLANS:' @@ -341,21 +293,19 @@ No rows affected ' 0 [Column[key]]' ' 1 [Column[key]]' ' Position of Big Table: 1' -' File Output Operator' -' compressed: false' -' GlobalTableId: 0' -' directory: file:!!{hive.exec.local.scratchdir}!!' -' NumFilesPerFileSink: 1' -' table:' -' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' -' properties:' -' columns ' -' columns.types ' -' escape.delim \' -' TotalFiles: 1' -' GatherStats: false' -' MultiFileSpray: false' +' Select Operator' +' Group By Operator' +' aggregations:' +' expr: count()' +' bucketGroup: false' +' mode: hash' +' outputColumnNames: _col0' +' Reduce Output Operator' +' sort order: ' +' tag: -1' +' value expressions:' +' expr: _col0' +' type: bigint' ' Needs Tagging: false' ' Path -> Alias:' ' !!{hive.metastore.warehouse.dir}!!/bucketcontext_1.db/bucket_big/ds=2008-04-08 [b]' @@ -369,7 +319,6 @@ No rows affected ' partition values:' ' ds 2008-04-08' ' properties:' -' SORTBUCKETCOLSPREFIX TRUE' ' bucket_count 4' ' bucket_field_name key' ' columns key,value' @@ -379,7 +328,6 @@ No rows affected ' location !!{hive.metastore.warehouse.dir}!!/bucketcontext_1.db/bucket_big/ds=2008-04-08' ' name bucketcontext_1.bucket_big' ' numFiles 4' -' numPartitions 2' ' numRows 0' ' partition_columns ds' ' rawDataSize 0' @@ -423,7 +371,6 @@ No rows affected ' partition values:' ' ds 2008-04-09' ' properties:' -' SORTBUCKETCOLSPREFIX TRUE' ' bucket_count 4' ' bucket_field_name key' ' columns key,value' @@ -433,7 +380,6 @@ No rows affected ' location !!{hive.metastore.warehouse.dir}!!/bucketcontext_1.db/bucket_big/ds=2008-04-09' ' name bucketcontext_1.bucket_big' ' numFiles 4' -' numPartitions 2' ' numRows 0' ' partition_columns ds' ' rawDataSize 0' @@ -469,48 +415,6 @@ No rows affected ' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' ' name: bucketcontext_1.bucket_big' ' name: bucketcontext_1.bucket_big' -' Truncated Path -> Alias:' -' /bucketcontext_1.db/bucket_big/ds=2008-04-08 [b]' -' /bucketcontext_1.db/bucket_big/ds=2008-04-09 [b]' -'' -' Stage: Stage-2' -' Map Reduce' -' Alias -> Map Operator Tree:' -' file:!!{hive.exec.local.scratchdir}!! ' -' Select Operator' -' Select Operator' -' Group By Operator' -' aggregations:' -' expr: count()' -' bucketGroup: false' -' mode: hash' -' outputColumnNames: _col0' -' Reduce Output Operator' -' sort order: ' -' tag: -1' -' value expressions:' -' expr: _col0' -' type: bigint' -' Needs Tagging: false' -' Path -> Alias:' -' file:!!{hive.exec.local.scratchdir}!! [file:!!{hive.exec.local.scratchdir}!!]' -' Path -> Partition:' -' file:!!{hive.exec.local.scratchdir}!! ' -' Partition' -' base file name: -mr-10002' -' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' -' properties:' -' columns ' -' columns.types ' -' escape.delim \' -' ' -' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' -' properties:' -' columns ' -' columns.types ' -' escape.delim \' ' Reduce Operator Tree:' ' Group By Operator' ' aggregations:' @@ -541,14 +445,15 @@ No rows affected ' GatherStats: false' ' MultiFileSpray: false' ' Truncated Path -> Alias:' -' file:!!{hive.exec.local.scratchdir}!! [file:!!{hive.exec.local.scratchdir}!!]' +' /bucketcontext_1.db/bucket_big/ds=2008-04-08 [b]' +' /bucketcontext_1.db/bucket_big/ds=2008-04-09 [b]' '' ' Stage: Stage-0' ' Fetch Operator' ' limit: -1' '' '' -234 rows selected +186 rows selected >>> select /* + MAPJOIN(a) */ count(*) FROM bucket_small a JOIN bucket_big b ON a.key = b.key; '_c1' '928' diff --git ql/src/test/results/beelinepositive/create_default_prop.q.out ql/src/test/results/beelinepositive/create_default_prop.q.out index 4773ce0..febc3b9 100644 --- ql/src/test/results/beelinepositive/create_default_prop.q.out +++ ql/src/test/results/beelinepositive/create_default_prop.q.out @@ -19,7 +19,7 @@ No rows affected 'col_name','data_type','comment' 'a','string','' '','NULL','NULL' -'Detailed Table Information','Table(tableName:table_p2, dbName:create_default_prop, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:a, type:string, comment:null)], location:!!{hive.metastore.warehouse.dir}!!/create_default_prop.db/table_p2, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{}), storedAsSubDirectories:false), partitionKeys:[], parameters:{transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)','' +'Detailed Table Information','Table(tableName:table_p2, dbName:create_default_prop, owner:!!{user.name}!!, createTime:!!UNIXTIME!!, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:a, type:string, comment:null)], location:!!{hive.metastore.warehouse.dir}!!/create_default_prop.db/table_p2, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[], skewedColValues:[], skewedColValueLocationMaps:{}), storedAsSubDirectories:false), partitionKeys:[], parameters:{p3=v3, transient_lastDdlTime=!!UNIXTIME!!}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)','' 3 rows selected >>> >>> CREATE TABLE table_p3 AS SELECT * FROM table_p1; diff --git ql/src/test/results/beelinepositive/create_skewed_table1.q.out ql/src/test/results/beelinepositive/create_skewed_table1.q.out index f6eb558..f2bf7b6 100644 --- ql/src/test/results/beelinepositive/create_skewed_table1.q.out +++ ql/src/test/results/beelinepositive/create_skewed_table1.q.out @@ -2,8 +2,6 @@ Saving all output to "!!{outputDirectory}!!/create_skewed_table1.q.raw". Enter " >>> !run !!{qFileDirectory}!!/create_skewed_table1.q >>> set hive.mapred.supports.subdirectories=true; No rows affected ->>> set hive.internal.ddl.list.bucketing.enable=true; -No rows affected >>> CREATE TABLE list_bucket_single (key STRING, value STRING) SKEWED BY (key) ON ('1','5','6'); No rows affected >>> CREATE TABLE list_bucket_single_2 (key STRING, value STRING) SKEWED BY (key) ON ((1),(5),(6)); diff --git ql/src/test/results/beelinepositive/create_view.q.out ql/src/test/results/beelinepositive/create_view.q.out index 622a914..3f7aab4 100644 --- ql/src/test/results/beelinepositive/create_view.q.out +++ ql/src/test/results/beelinepositive/create_view.q.out @@ -84,6 +84,7 @@ CREATE VIEW view0(valoo) AS SELECT upper(value) FROM src WHERE key=86; ' Create View Operator:' ' Create View' ' if not exists: false' +' is alter view as select: false' ' or replace: false' ' columns: valoo string' ' expanded text: SELECT `_c0` AS `valoo` FROM (SELECT upper(`src`.`value`) FROM `create_view`.`src` WHERE `src`.`key`=86) `view0`' @@ -91,7 +92,7 @@ CREATE VIEW view0(valoo) AS SELECT upper(value) FROM src WHERE key=86; ' original text: SELECT upper(value) FROM src WHERE key=86' '' '' -18 rows selected +19 rows selected >>> >>> -- make sure EXPLAIN works with a query which references a view >>> EXPLAIN @@ -122,26 +123,19 @@ SELECT * from view2 where key=18; ' expr: value' ' type: string' ' outputColumnNames: _col0, _col1' -' Select Operator' -' expressions:' -' expr: _col0' -' type: string' -' expr: _col1' -' type: string' -' outputColumnNames: _col0, _col1' -' File Output Operator' -' compressed: false' -' GlobalTableId: 0' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' File Output Operator' +' compressed: false' +' GlobalTableId: 0' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' '' ' Stage: Stage-0' ' Fetch Operator' ' limit: -1' '' '' -44 rows selected +37 rows selected >>> >>> SHOW TABLES 'view.*'; 'tab_name' diff --git ql/src/test/results/beelinepositive/describe_table_json.q.out ql/src/test/results/beelinepositive/describe_table_json.q.out index 84d2e64..8a5811c 100644 --- ql/src/test/results/beelinepositive/describe_table_json.q.out +++ ql/src/test/results/beelinepositive/describe_table_json.q.out @@ -31,7 +31,7 @@ No rows affected >>> >>> DESCRIBE extended jsontable; 'col_name','data_type','comment' -'{"columns":[{"name":"key","type":"int"},{"name":"value","type":"string"}],"tableInfo":{"owner":"!!{user.name}!!","parameters":{"id":"jsontable","last_modified_by":"!!{user.name}!!","last_modified_time":"!!UNIXTIME!!","transient_lastDdlTime":"!!UNIXTIME!!","comment":"json table"},"tableName":"jsontable","dbName":"describe_table_json","tableType":"MANAGED_TABLE","sd":{"location":"!!{hive.metastore.warehouse.dir}!!/describe_table_json.db/jsontable","parameters":{},"inputFormat":"org.apache.hadoop.mapred.TextInputFormat","outputFormat":"org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat","cols":[{"name":"key","type":"int","comment":null,"setName":true,"setType":true,"setComment":false},{"name":"value","type":"string","comment":null,"setName":true,"setType":true,"setComment":false}],"skewedInfo":{"skewedColNames":[],"skewedColValues":[],"skewedColValueLocationMaps":{},"skewedColNamesSize":0,"skewedColNamesIterator":[],"setSkewedColNames":true,"skewedColValuesSize":0,"skewedColValuesIterator":[],"setSkewedColValues":true,"skewedColValueLocationMapsSize":0,"setSkewedColValueLocationMaps":true},"serdeInfo":{"name":null,"parameters":{"serialization.format":"1"},"serializationLib":"org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe","setName":false,"parametersSize":1,"setParameters":true,"setSerializationLib":true},"numBuckets":-1,"bucketCols":[],"compressed":false,"sortCols":[],"storedAsSubDirectories":false,"parametersSize":0,"setParameters":true,"colsSize":2,"colsIterator":[{"name":"key","type":"int","comment":null,"setName":true,"setType":true,"setComment":false},{"name":"value","type":"string","comment":null,"setName":true,"setType":true,"setComment":false}],"setCols":true,"setLocation":true,"setInputFormat":true,"setOutputFormat":true,"setCompressed":true,"setNumBuckets":true,"setSerdeInfo":true,"bucketColsSize":0,"bucketColsIterator":[],"setBucketCols":true,"sortColsSize":0,"sortColsIterator":[],"setSortCols":true,"setSkewedInfo":true,"setStoredAsSubDirectories":true},"partitionKeys":[],"createTime":!!UNIXTIME!!,"privileges":null,"viewOriginalText":null,"lastAccessTime":0,"retention":0,"viewExpandedText":null,"partitionKeysSize":0,"setDbName":true,"setCreateTime":true,"setLastAccessTime":true,"setSd":true,"parametersSize":5,"setParameters":true,"setTableName":true,"setOwner":true,"setRetention":true,"partitionKeysIterator":[],"setPartitionKeys":true,"setViewOriginalText":false,"setViewExpandedText":false,"setTableType":true,"setPrivileges":false}}','NULL','NULL' +'{"columns":[{"name":"key","type":"int"},{"name":"value","type":"string"}],"tableInfo":{"owner":"!!{user.name}!!","parameters":{"id":"jsontable","last_modified_by":"!!{user.name}!!","last_modified_time":"!!UNIXTIME!!","transient_lastDdlTime":"!!UNIXTIME!!","comment":"json table"},"tableName":"jsontable","dbName":"describe_table_json","tableType":"MANAGED_TABLE","sd":{"location":"!!{hive.metastore.warehouse.dir}!!/describe_table_json.db/jsontable","parameters":{},"inputFormat":"org.apache.hadoop.mapred.TextInputFormat","outputFormat":"org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat","cols":[{"name":"key","type":"int","comment":null,"setName":true,"setType":true,"setComment":false},{"name":"value","type":"string","comment":null,"setName":true,"setType":true,"setComment":false}],"skewedInfo":{"skewedColNames":[],"skewedColValues":[],"skewedColValueLocationMaps":{},"skewedColNamesSize":0,"skewedColNamesIterator":[],"setSkewedColNames":true,"skewedColValuesSize":0,"skewedColValuesIterator":[],"setSkewedColValues":true,"skewedColValueLocationMapsSize":0,"setSkewedColValueLocationMaps":true},"serdeInfo":{"name":null,"parameters":{"serialization.format":"1"},"serializationLib":"org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe","parametersSize":1,"setParameters":true,"setName":false,"setSerializationLib":true},"parametersSize":0,"setParameters":true,"numBuckets":-1,"bucketCols":[],"compressed":false,"sortCols":[],"storedAsSubDirectories":false,"colsSize":2,"colsIterator":[{"name":"key","type":"int","comment":null,"setName":true,"setType":true,"setComment":false},{"name":"value","type":"string","comment":null,"setName":true,"setType":true,"setComment":false}],"setCols":true,"setLocation":true,"setInputFormat":true,"setOutputFormat":true,"setCompressed":true,"setNumBuckets":true,"setSerdeInfo":true,"bucketColsSize":0,"bucketColsIterator":[],"setBucketCols":true,"sortColsSize":0,"sortColsIterator":[],"setSortCols":true,"setSkewedInfo":true,"setStoredAsSubDirectories":true},"partitionKeys":[],"createTime":!!UNIXTIME!!,"setDbName":true,"setTableName":true,"setCreateTime":true,"lastAccessTime":0,"setLastAccessTime":true,"setSd":true,"parametersSize":5,"setParameters":true,"privileges":null,"setPrivileges":false,"viewOriginalText":null,"retention":0,"viewExpandedText":null,"partitionKeysSize":0,"setOwner":true,"setRetention":true,"partitionKeysIterator":[],"setPartitionKeys":true,"setViewOriginalText":false,"setViewExpandedText":false,"setTableType":true}}','NULL','NULL' 1 row selected >>> >>> DROP TABLE jsontable; diff --git ql/src/test/results/beelinepositive/groupby1.q.out ql/src/test/results/beelinepositive/groupby1.q.out index beb9177..f1f23f5 100644 --- ql/src/test/results/beelinepositive/groupby1.q.out +++ ql/src/test/results/beelinepositive/groupby1.q.out @@ -94,26 +94,19 @@ FROM src INSERT OVERWRITE TABLE dest_g1 SELECT src.key, sum(substr(src.value,5)) ' outputColumnNames: _col0, _col1' ' Select Operator' ' expressions:' -' expr: _col0' -' type: string' +' expr: UDFToInteger(_col0)' +' type: int' ' expr: _col1' ' type: double' ' outputColumnNames: _col0, _col1' -' Select Operator' -' expressions:' -' expr: UDFToInteger(_col0)' -' type: int' -' expr: _col1' -' type: double' -' outputColumnNames: _col0, _col1' -' File Output Operator' -' compressed: false' -' GlobalTableId: 1' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: groupby1.dest_g1' +' File Output Operator' +' compressed: false' +' GlobalTableId: 1' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' name: groupby1.dest_g1' '' ' Stage: Stage-0' ' Move Operator' @@ -129,7 +122,7 @@ FROM src INSERT OVERWRITE TABLE dest_g1 SELECT src.key, sum(substr(src.value,5)) ' Stats-Aggr Operator' '' '' -115 rows selected +108 rows selected >>> >>> set fs.default.name=file:///; No rows affected diff --git ql/src/test/results/beelinepositive/groupby10.q.out ql/src/test/results/beelinepositive/groupby10.q.out index e429512..de79530 100644 --- ql/src/test/results/beelinepositive/groupby10.q.out +++ ql/src/test/results/beelinepositive/groupby10.q.out @@ -121,28 +121,19 @@ INSERT OVERWRITE TABLE dest2 SELECT INPUT.key, sum(substr(INPUT.value,5)), sum(d ' expressions:' ' expr: _col0' ' type: int' -' expr: _col1' -' type: bigint' -' expr: _col2' -' type: bigint' +' expr: UDFToInteger(_col1)' +' type: int' +' expr: UDFToInteger(_col2)' +' type: int' ' outputColumnNames: _col0, _col1, _col2' -' Select Operator' -' expressions:' -' expr: _col0' -' type: int' -' expr: UDFToInteger(_col1)' -' type: int' -' expr: UDFToInteger(_col2)' -' type: int' -' outputColumnNames: _col0, _col1, _col2' -' File Output Operator' -' compressed: false' -' GlobalTableId: 1' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: groupby10.dest1' +' File Output Operator' +' compressed: false' +' GlobalTableId: 1' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' name: groupby10.dest1' '' ' Stage: Stage-0' ' Move Operator' @@ -190,28 +181,19 @@ INSERT OVERWRITE TABLE dest2 SELECT INPUT.key, sum(substr(INPUT.value,5)), sum(d ' expressions:' ' expr: _col0' ' type: int' -' expr: _col1' -' type: double' -' expr: _col2' -' type: double' +' expr: UDFToInteger(_col1)' +' type: int' +' expr: UDFToInteger(_col2)' +' type: int' ' outputColumnNames: _col0, _col1, _col2' -' Select Operator' -' expressions:' -' expr: _col0' -' type: int' -' expr: UDFToInteger(_col1)' -' type: int' -' expr: UDFToInteger(_col2)' -' type: int' -' outputColumnNames: _col0, _col1, _col2' -' File Output Operator' -' compressed: false' -' GlobalTableId: 2' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: groupby10.dest2' +' File Output Operator' +' compressed: false' +' GlobalTableId: 2' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' name: groupby10.dest2' '' ' Stage: Stage-1' ' Move Operator' @@ -227,7 +209,7 @@ INSERT OVERWRITE TABLE dest2 SELECT INPUT.key, sum(substr(INPUT.value,5)), sum(d ' Stats-Aggr Operator' '' '' -205 rows selected +187 rows selected >>> >>> FROM INPUT INSERT OVERWRITE TABLE dest1 SELECT INPUT.key, count(substr(INPUT.value,5)), count(distinct substr(INPUT.value,5)) GROUP BY INPUT.key @@ -389,28 +371,19 @@ INSERT OVERWRITE TABLE dest2 SELECT INPUT.key, sum(substr(INPUT.value,5)), sum(d ' expressions:' ' expr: _col0' ' type: int' -' expr: _col1' -' type: bigint' -' expr: _col2' -' type: bigint' +' expr: UDFToInteger(_col1)' +' type: int' +' expr: UDFToInteger(_col2)' +' type: int' ' outputColumnNames: _col0, _col1, _col2' -' Select Operator' -' expressions:' -' expr: _col0' -' type: int' -' expr: UDFToInteger(_col1)' -' type: int' -' expr: UDFToInteger(_col2)' -' type: int' -' outputColumnNames: _col0, _col1, _col2' -' File Output Operator' -' compressed: false' -' GlobalTableId: 1' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: groupby10.dest1' +' File Output Operator' +' compressed: false' +' GlobalTableId: 1' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' name: groupby10.dest1' '' ' Stage: Stage-0' ' Move Operator' @@ -458,28 +431,19 @@ INSERT OVERWRITE TABLE dest2 SELECT INPUT.key, sum(substr(INPUT.value,5)), sum(d ' expressions:' ' expr: _col0' ' type: int' -' expr: _col1' -' type: double' -' expr: _col2' -' type: double' +' expr: UDFToInteger(_col1)' +' type: int' +' expr: UDFToInteger(_col2)' +' type: int' ' outputColumnNames: _col0, _col1, _col2' -' Select Operator' -' expressions:' -' expr: _col0' -' type: int' -' expr: UDFToInteger(_col1)' -' type: int' -' expr: UDFToInteger(_col2)' -' type: int' -' outputColumnNames: _col0, _col1, _col2' -' File Output Operator' -' compressed: false' -' GlobalTableId: 2' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: groupby10.dest2' +' File Output Operator' +' compressed: false' +' GlobalTableId: 2' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' name: groupby10.dest2' '' ' Stage: Stage-1' ' Move Operator' @@ -495,7 +459,7 @@ INSERT OVERWRITE TABLE dest2 SELECT INPUT.key, sum(substr(INPUT.value,5)), sum(d ' Stats-Aggr Operator' '' '' -205 rows selected +187 rows selected >>> >>> FROM INPUT INSERT OVERWRITE TABLE dest1 SELECT INPUT.key, count(substr(INPUT.value,5)), count(distinct substr(INPUT.value,5)) GROUP BY INPUT.key @@ -549,4 +513,252 @@ No rows selected '409','409','409' '484','484','484' 20 rows selected +>>> +>>> -- HIVE-3852 Multi-groupby optimization fails when same distinct column is used twice or more +>>> EXPLAIN +FROM INPUT +INSERT OVERWRITE TABLE dest1 SELECT INPUT.key, sum(distinct substr(INPUT.value,5)), count(distinct substr(INPUT.value,5)) GROUP BY INPUT.key +INSERT OVERWRITE TABLE dest2 SELECT INPUT.key, sum(distinct substr(INPUT.value,5)), avg(distinct substr(INPUT.value,5)) GROUP BY INPUT.key; +'Explain' +'ABSTRACT SYNTAX TREE:' +' (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME INPUT))) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME dest1))) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL INPUT) key)) (TOK_SELEXPR (TOK_FUNCTIONDI sum (TOK_FUNCTION substr (. (TOK_TABLE_OR_COL INPUT) value) 5))) (TOK_SELEXPR (TOK_FUNCTIONDI count (TOK_FUNCTION substr (. (TOK_TABLE_OR_COL INPUT) value) 5)))) (TOK_GROUPBY (. (TOK_TABLE_OR_COL INPUT) key))) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME dest2))) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL INPUT) key)) (TOK_SELEXPR (TOK_FUNCTIONDI sum (TOK_FUNCTION substr (. (TOK_TABLE_OR_COL INPUT) value) 5))) (TOK_SELEXPR (TOK_FUNCTIONDI avg (TOK_FUNCTION substr (. (TOK_TABLE_OR_COL INPUT) value) 5)))) (TOK_GROUPBY (. (TOK_TABLE_OR_COL INPUT) key))))' +'' +'STAGE DEPENDENCIES:' +' Stage-2 is a root stage' +' Stage-3 depends on stages: Stage-2' +' Stage-0 depends on stages: Stage-3' +' Stage-4 depends on stages: Stage-0' +' Stage-5 depends on stages: Stage-2' +' Stage-1 depends on stages: Stage-5' +' Stage-6 depends on stages: Stage-1' +'' +'STAGE PLANS:' +' Stage: Stage-2' +' Map Reduce' +' Alias -> Map Operator Tree:' +' input ' +' TableScan' +' alias: input' +' Reduce Output Operator' +' key expressions:' +' expr: substr(value, 5)' +' type: string' +' sort order: +' +' Map-reduce partition columns:' +' expr: substr(value, 5)' +' type: string' +' tag: -1' +' value expressions:' +' expr: key' +' type: int' +' Reduce Operator Tree:' +' Forward' +' Group By Operator' +' aggregations:' +' expr: sum(DISTINCT KEY._col0)' +' expr: count(DISTINCT KEY._col0)' +' bucketGroup: false' +' keys:' +' expr: VALUE._col0' +' type: int' +' mode: hash' +' outputColumnNames: _col0, _col1, _col2' +' File Output Operator' +' compressed: false' +' GlobalTableId: 0' +' table:' +' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' +' Group By Operator' +' aggregations:' +' expr: sum(DISTINCT KEY._col0)' +' expr: avg(DISTINCT KEY._col0)' +' bucketGroup: false' +' keys:' +' expr: VALUE._col0' +' type: int' +' mode: hash' +' outputColumnNames: _col0, _col1, _col2' +' File Output Operator' +' compressed: false' +' GlobalTableId: 0' +' table:' +' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' +'' +' Stage: Stage-3' +' Map Reduce' +' Alias -> Map Operator Tree:' +' file:!!{hive.exec.local.scratchdir}!! ' +' Reduce Output Operator' +' key expressions:' +' expr: _col0' +' type: int' +' sort order: +' +' Map-reduce partition columns:' +' expr: _col0' +' type: int' +' tag: -1' +' value expressions:' +' expr: _col1' +' type: double' +' expr: _col2' +' type: bigint' +' Reduce Operator Tree:' +' Group By Operator' +' aggregations:' +' expr: sum(VALUE._col0)' +' expr: count(VALUE._col1)' +' bucketGroup: false' +' keys:' +' expr: KEY._col0' +' type: int' +' mode: final' +' outputColumnNames: _col0, _col1, _col2' +' Select Operator' +' expressions:' +' expr: _col0' +' type: int' +' expr: UDFToInteger(_col1)' +' type: int' +' expr: UDFToInteger(_col2)' +' type: int' +' outputColumnNames: _col0, _col1, _col2' +' File Output Operator' +' compressed: false' +' GlobalTableId: 1' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' name: groupby10.dest1' +'' +' Stage: Stage-0' +' Move Operator' +' tables:' +' replace: true' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' name: groupby10.dest1' +'' +' Stage: Stage-4' +' Stats-Aggr Operator' +'' +' Stage: Stage-5' +' Map Reduce' +' Alias -> Map Operator Tree:' +' file:!!{hive.exec.local.scratchdir}!! ' +' Reduce Output Operator' +' key expressions:' +' expr: _col0' +' type: int' +' sort order: +' +' Map-reduce partition columns:' +' expr: _col0' +' type: int' +' tag: -1' +' value expressions:' +' expr: _col1' +' type: double' +' expr: _col2' +' type: struct' +' Reduce Operator Tree:' +' Group By Operator' +' aggregations:' +' expr: sum(VALUE._col0)' +' expr: avg(VALUE._col1)' +' bucketGroup: false' +' keys:' +' expr: KEY._col0' +' type: int' +' mode: final' +' outputColumnNames: _col0, _col1, _col2' +' Select Operator' +' expressions:' +' expr: _col0' +' type: int' +' expr: UDFToInteger(_col1)' +' type: int' +' expr: UDFToInteger(_col2)' +' type: int' +' outputColumnNames: _col0, _col1, _col2' +' File Output Operator' +' compressed: false' +' GlobalTableId: 2' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' name: groupby10.dest2' +'' +' Stage: Stage-1' +' Move Operator' +' tables:' +' replace: true' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' name: groupby10.dest2' +'' +' Stage: Stage-6' +' Stats-Aggr Operator' +'' +'' +187 rows selected +>>> +>>> FROM INPUT +INSERT OVERWRITE TABLE dest1 SELECT INPUT.key, sum(distinct substr(INPUT.value,5)), count(distinct substr(INPUT.value,5)) GROUP BY INPUT.key +INSERT OVERWRITE TABLE dest2 SELECT INPUT.key, sum(distinct substr(INPUT.value,5)), avg(distinct substr(INPUT.value,5)) GROUP BY INPUT.key; +'_col0','_col1','_col2' +No rows selected +>>> +>>> SELECT * from dest1; +'key','val1','val2' +'27','27','1' +'66','66','1' +'86','86','1' +'98','98','1' +'128','128','1' +'150','150','1' +'165','165','1' +'193','193','1' +'213','427','2' +'224','224','1' +'238','717','3' +'255','255','1' +'265','265','1' +'273','273','1' +'278','278','1' +'311','311','1' +'369','369','1' +'401','401','1' +'409','409','1' +'484','484','1' +20 rows selected +>>> SELECT * from dest2; +'key','val1','val2' +'27','27','27' +'66','66','66' +'86','86','86' +'98','98','98' +'128','128','128' +'150','150','150' +'165','165','165' +'193','193','193' +'213','427','213' +'224','224','224' +'238','717','239' +'255','255','255' +'265','265','265' +'273','273','273' +'278','278','278' +'311','311','311' +'369','369','369' +'401','401','401' +'409','409','409' +'484','484','484' +20 rows selected >>> !record diff --git ql/src/test/results/beelinepositive/input_part1.q.out ql/src/test/results/beelinepositive/input_part1.q.out index c691efc..362f523 100644 --- ql/src/test/results/beelinepositive/input_part1.q.out +++ ql/src/test/results/beelinepositive/input_part1.q.out @@ -35,8 +35,8 @@ INSERT OVERWRITE TABLE dest1 SELECT srcpart.key, srcpart.value, srcpart.hr, srcp ' type: boolean' ' Select Operator' ' expressions:' -' expr: key' -' type: string' +' expr: UDFToInteger(key)' +' type: int' ' expr: value' ' type: string' ' expr: hr' @@ -44,43 +44,32 @@ INSERT OVERWRITE TABLE dest1 SELECT srcpart.key, srcpart.value, srcpart.hr, srcp ' expr: ds' ' type: string' ' outputColumnNames: _col0, _col1, _col2, _col3' -' Select Operator' -' expressions:' -' expr: UDFToInteger(_col0)' -' type: int' -' expr: _col1' -' type: string' -' expr: _col2' -' type: string' -' expr: _col3' -' type: string' -' outputColumnNames: _col0, _col1, _col2, _col3' -' File Output Operator' -' compressed: false' -' GlobalTableId: 1' -' directory: pfile:!!{hive.exec.scratchdir}!!' -' NumFilesPerFileSink: 1' -' Stats Publishing Key Prefix: pfile:!!{hive.exec.scratchdir}!!' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' properties:' -' bucket_count -1' -' columns key,value,hr,ds' -' columns.types int:string:string:string' -' file.inputformat org.apache.hadoop.mapred.TextInputFormat' -' file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' location !!{hive.metastore.warehouse.dir}!!/input_part1.db/dest1' -' name input_part1.dest1' -' serialization.ddl struct dest1 { i32 key, string value, string hr, string ds}' -' serialization.format 1' -' serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' transient_lastDdlTime !!UNIXTIME!!' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: input_part1.dest1' -' TotalFiles: 1' -' GatherStats: true' -' MultiFileSpray: false' +' File Output Operator' +' compressed: false' +' GlobalTableId: 1' +' directory: pfile:!!{hive.exec.scratchdir}!!' +' NumFilesPerFileSink: 1' +' Stats Publishing Key Prefix: pfile:!!{hive.exec.scratchdir}!!' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' properties:' +' bucket_count -1' +' columns key,value,hr,ds' +' columns.types int:string:string:string' +' file.inputformat org.apache.hadoop.mapred.TextInputFormat' +' file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' location !!{hive.metastore.warehouse.dir}!!/input_part1.db/dest1' +' name input_part1.dest1' +' serialization.ddl struct dest1 { i32 key, string value, string hr, string ds}' +' serialization.format 1' +' serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' transient_lastDdlTime !!UNIXTIME!!' +' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' name: input_part1.dest1' +' TotalFiles: 1' +' GatherStats: true' +' MultiFileSpray: false' ' Needs Tagging: false' ' Path -> Alias:' ' !!{hive.metastore.warehouse.dir}!!/input_part1.db/srcpart/ds=2008-04-08/hr=12 [srcpart]' @@ -102,7 +91,6 @@ INSERT OVERWRITE TABLE dest1 SELECT srcpart.key, srcpart.value, srcpart.hr, srcp ' location !!{hive.metastore.warehouse.dir}!!/input_part1.db/srcpart/ds=2008-04-08/hr=12' ' name input_part1.srcpart' ' numFiles 1' -' numPartitions 4' ' numRows 0' ' partition_columns ds/hr' ' rawDataSize 0' @@ -329,7 +317,7 @@ INSERT OVERWRITE TABLE dest1 SELECT srcpart.key, srcpart.value, srcpart.hr, srcp ' destination: pfile:!!{hive.exec.scratchdir}!!' '' '' -322 rows selected +310 rows selected >>> >>> FROM srcpart INSERT OVERWRITE TABLE dest1 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12'; diff --git ql/src/test/results/beelinepositive/join_view.q.out ql/src/test/results/beelinepositive/join_view.q.out index 998d1cb..a7674f1 100644 --- ql/src/test/results/beelinepositive/join_view.q.out +++ ql/src/test/results/beelinepositive/join_view.q.out @@ -89,28 +89,19 @@ No rows selected ' expr: _col7' ' type: string' ' outputColumnNames: _col0, _col1, _col2' -' Select Operator' -' expressions:' -' expr: _col0' -' type: string' -' expr: _col1' -' type: int' -' expr: _col2' -' type: string' -' outputColumnNames: _col0, _col1, _col2' -' File Output Operator' -' compressed: false' -' GlobalTableId: 0' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' File Output Operator' +' compressed: false' +' GlobalTableId: 0' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' '' ' Stage: Stage-0' ' Fetch Operator' ' limit: -1' '' '' -90 rows selected +81 rows selected >>> >>> drop view v; No rows affected diff --git ql/src/test/results/beelinepositive/louter_join_ppr.q.out ql/src/test/results/beelinepositive/louter_join_ppr.q.out index 6c91d42..7617256 100644 --- ql/src/test/results/beelinepositive/louter_join_ppr.q.out +++ ql/src/test/results/beelinepositive/louter_join_ppr.q.out @@ -138,7 +138,6 @@ WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25; ' location !!{hive.metastore.warehouse.dir}!!/louter_join_ppr.db/srcpart/ds=2008-04-08/hr=11' ' name louter_join_ppr.srcpart' ' numFiles 1' -' numPartitions 4' ' numRows 0' ' partition_columns ds/hr' ' rawDataSize 0' @@ -189,7 +188,6 @@ WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25; ' location !!{hive.metastore.warehouse.dir}!!/louter_join_ppr.db/srcpart/ds=2008-04-08/hr=12' ' name louter_join_ppr.srcpart' ' numFiles 1' -' numPartitions 4' ' numRows 0' ' partition_columns ds/hr' ' rawDataSize 0' @@ -275,7 +273,7 @@ WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25; ' limit: -1' '' '' -263 rows selected +261 rows selected >>> >>> FROM src a @@ -438,7 +436,6 @@ WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25; ' location !!{hive.metastore.warehouse.dir}!!/louter_join_ppr.db/srcpart/ds=2008-04-08/hr=11' ' name louter_join_ppr.srcpart' ' numFiles 1' -' numPartitions 4' ' numRows 0' ' partition_columns ds/hr' ' rawDataSize 0' @@ -489,7 +486,6 @@ WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25; ' location !!{hive.metastore.warehouse.dir}!!/louter_join_ppr.db/srcpart/ds=2008-04-08/hr=12' ' name louter_join_ppr.srcpart' ' numFiles 1' -' numPartitions 4' ' numRows 0' ' partition_columns ds/hr' ' rawDataSize 0' @@ -540,7 +536,6 @@ WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25; ' location !!{hive.metastore.warehouse.dir}!!/louter_join_ppr.db/srcpart/ds=2008-04-09/hr=11' ' name louter_join_ppr.srcpart' ' numFiles 1' -' numPartitions 4' ' numRows 0' ' partition_columns ds/hr' ' rawDataSize 0' @@ -591,7 +586,6 @@ WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25; ' location !!{hive.metastore.warehouse.dir}!!/louter_join_ppr.db/srcpart/ds=2008-04-09/hr=12' ' name louter_join_ppr.srcpart' ' numFiles 1' -' numPartitions 4' ' numRows 0' ' partition_columns ds/hr' ' rawDataSize 0' @@ -684,7 +678,7 @@ WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25; ' limit: -1' '' '' -376 rows selected +372 rows selected >>> >>> FROM srcpart a @@ -848,7 +842,6 @@ WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND b.ds = '2008-0 ' location !!{hive.metastore.warehouse.dir}!!/louter_join_ppr.db/srcpart/ds=2008-04-08/hr=11' ' name louter_join_ppr.srcpart' ' numFiles 1' -' numPartitions 4' ' numRows 0' ' partition_columns ds/hr' ' rawDataSize 0' @@ -899,7 +892,6 @@ WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND b.ds = '2008-0 ' location !!{hive.metastore.warehouse.dir}!!/louter_join_ppr.db/srcpart/ds=2008-04-08/hr=12' ' name louter_join_ppr.srcpart' ' numFiles 1' -' numPartitions 4' ' numRows 0' ' partition_columns ds/hr' ' rawDataSize 0' @@ -950,7 +942,6 @@ WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND b.ds = '2008-0 ' location !!{hive.metastore.warehouse.dir}!!/louter_join_ppr.db/srcpart/ds=2008-04-09/hr=11' ' name louter_join_ppr.srcpart' ' numFiles 1' -' numPartitions 4' ' numRows 0' ' partition_columns ds/hr' ' rawDataSize 0' @@ -1001,7 +992,6 @@ WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND b.ds = '2008-0 ' location !!{hive.metastore.warehouse.dir}!!/louter_join_ppr.db/srcpart/ds=2008-04-09/hr=12' ' name louter_join_ppr.srcpart' ' numFiles 1' -' numPartitions 4' ' numRows 0' ' partition_columns ds/hr' ' rawDataSize 0' @@ -1089,7 +1079,7 @@ WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND b.ds = '2008-0 ' limit: -1' '' '' -371 rows selected +367 rows selected >>> >>> FROM src a @@ -1250,7 +1240,6 @@ WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND a.ds = '2008-0 ' location !!{hive.metastore.warehouse.dir}!!/louter_join_ppr.db/srcpart/ds=2008-04-08/hr=11' ' name louter_join_ppr.srcpart' ' numFiles 1' -' numPartitions 4' ' numRows 0' ' partition_columns ds/hr' ' rawDataSize 0' @@ -1301,7 +1290,6 @@ WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND a.ds = '2008-0 ' location !!{hive.metastore.warehouse.dir}!!/louter_join_ppr.db/srcpart/ds=2008-04-08/hr=12' ' name louter_join_ppr.srcpart' ' numFiles 1' -' numPartitions 4' ' numRows 0' ' partition_columns ds/hr' ' rawDataSize 0' @@ -1387,7 +1375,7 @@ WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND a.ds = '2008-0 ' limit: -1' '' '' -265 rows selected +263 rows selected >>> >>> FROM srcpart a diff --git ql/src/test/results/beelinepositive/mapjoin1.q.out ql/src/test/results/beelinepositive/mapjoin1.q.out index 370635b..6292aba 100644 --- ql/src/test/results/beelinepositive/mapjoin1.q.out +++ ql/src/test/results/beelinepositive/mapjoin1.q.out @@ -9,4 +9,615 @@ JOIN src b ON a.key = b.key where a.ds is not null; 'sum_a' '1114788.0' 1 row selected +>>> +>>> set hive.outerjoin.supports.filters=true; +No rows affected +>>> +>>> -- const filter on outer join +>>> EXPLAIN +SELECT /*+ MAPJOIN(a) */ * FROM src a RIGHT OUTER JOIN src b on a.key=b.key AND true limit 10; +'Explain' +'ABSTRACT SYNTAX TREE:' +' (TOK_QUERY (TOK_FROM (TOK_RIGHTOUTERJOIN (TOK_TABREF (TOK_TABNAME src) a) (TOK_TABREF (TOK_TABNAME src) b) (AND (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL b) key)) true))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_MAPJOIN (TOK_HINTARGLIST a))) (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_LIMIT 10)))' +'' +'STAGE DEPENDENCIES:' +' Stage-3 is a root stage' +' Stage-1 depends on stages: Stage-3' +' Stage-0 is a root stage' +'' +'STAGE PLANS:' +' Stage: Stage-3' +' Map Reduce Local Work' +' Alias -> Map Local Tables:' +' a ' +' Fetch Operator' +' limit: -1' +' Alias -> Map Local Operator Tree:' +' a ' +' TableScan' +' alias: a' +' HashTable Sink Operator' +' condition expressions:' +' 0 {key} {value}' +' 1 {key} {value}' +' filter predicates:' +' 0 ' +' 1 {true}' +' handleSkewJoin: false' +' keys:' +' 0 [Column[key]]' +' 1 [Column[key]]' +' Position of Big Table: 1' +'' +' Stage: Stage-1' +' Map Reduce' +' Alias -> Map Operator Tree:' +' b ' +' TableScan' +' alias: b' +' Map Join Operator' +' condition map:' +' Right Outer Join0 to 1' +' condition expressions:' +' 0 {key} {value}' +' 1 {key} {value}' +' filter predicates:' +' 0 ' +' 1 {true}' +' handleSkewJoin: false' +' keys:' +' 0 [Column[key]]' +' 1 [Column[key]]' +' outputColumnNames: _col0, _col1, _col4, _col5' +' Position of Big Table: 1' +' Select Operator' +' expressions:' +' expr: _col0' +' type: string' +' expr: _col1' +' type: string' +' expr: _col4' +' type: string' +' expr: _col5' +' type: string' +' outputColumnNames: _col0, _col1, _col2, _col3' +' Limit' +' File Output Operator' +' compressed: false' +' GlobalTableId: 0' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' Local Work:' +' Map Reduce Local Work' +'' +' Stage: Stage-0' +' Fetch Operator' +' limit: 10' +'' +'' +79 rows selected +>>> SELECT /*+ MAPJOIN(a) */ * FROM src a RIGHT OUTER JOIN src b on a.key=b.key AND true limit 10; +'key','value','key','value' +'238','val_238','238','val_238' +'238','val_238','238','val_238' +'86','val_86','86','val_86' +'311','val_311','311','val_311' +'311','val_311','311','val_311' +'311','val_311','311','val_311' +'27','val_27','27','val_27' +'165','val_165','165','val_165' +'165','val_165','165','val_165' +'409','val_409','409','val_409' +10 rows selected +>>> +>>> -- func filter on outer join +>>> EXPLAIN +SELECT /*+ MAPJOIN(a) */ * FROM src a RIGHT OUTER JOIN src b on a.key=b.key AND b.key * 10 < '1000' limit 10; +'Explain' +'ABSTRACT SYNTAX TREE:' +' (TOK_QUERY (TOK_FROM (TOK_RIGHTOUTERJOIN (TOK_TABREF (TOK_TABNAME src) a) (TOK_TABREF (TOK_TABNAME src) b) (AND (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL b) key)) (< (* (. (TOK_TABLE_OR_COL b) key) 10) '1000')))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_MAPJOIN (TOK_HINTARGLIST a))) (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_LIMIT 10)))' +'' +'STAGE DEPENDENCIES:' +' Stage-3 is a root stage' +' Stage-1 depends on stages: Stage-3' +' Stage-0 is a root stage' +'' +'STAGE PLANS:' +' Stage: Stage-3' +' Map Reduce Local Work' +' Alias -> Map Local Tables:' +' a ' +' Fetch Operator' +' limit: -1' +' Alias -> Map Local Operator Tree:' +' a ' +' TableScan' +' alias: a' +' HashTable Sink Operator' +' condition expressions:' +' 0 {key} {value}' +' 1 {key} {value}' +' filter predicates:' +' 0 ' +' 1 {((key * 10) < '1000')}' +' handleSkewJoin: false' +' keys:' +' 0 [Column[key]]' +' 1 [Column[key]]' +' Position of Big Table: 1' +'' +' Stage: Stage-1' +' Map Reduce' +' Alias -> Map Operator Tree:' +' b ' +' TableScan' +' alias: b' +' Map Join Operator' +' condition map:' +' Right Outer Join0 to 1' +' condition expressions:' +' 0 {key} {value}' +' 1 {key} {value}' +' filter predicates:' +' 0 ' +' 1 {((key * 10) < '1000')}' +' handleSkewJoin: false' +' keys:' +' 0 [Column[key]]' +' 1 [Column[key]]' +' outputColumnNames: _col0, _col1, _col4, _col5' +' Position of Big Table: 1' +' Select Operator' +' expressions:' +' expr: _col0' +' type: string' +' expr: _col1' +' type: string' +' expr: _col4' +' type: string' +' expr: _col5' +' type: string' +' outputColumnNames: _col0, _col1, _col2, _col3' +' Limit' +' File Output Operator' +' compressed: false' +' GlobalTableId: 0' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' Local Work:' +' Map Reduce Local Work' +'' +' Stage: Stage-0' +' Fetch Operator' +' limit: 10' +'' +'' +79 rows selected +>>> SELECT /*+ MAPJOIN(a) */ * FROM src a RIGHT OUTER JOIN src b on a.key=b.key AND b.key * 10 < '1000' limit 10; +'key','value','key','value' +'NULL','NULL','238','val_238' +'NULL','NULL','238','val_238' +'86','val_86','86','val_86' +'NULL','NULL','311','val_311' +'NULL','NULL','311','val_311' +'NULL','NULL','311','val_311' +'27','val_27','27','val_27' +'NULL','NULL','165','val_165' +'NULL','NULL','165','val_165' +'NULL','NULL','409','val_409' +10 rows selected +>>> +>>> -- field filter on outer join +>>> EXPLAIN +SELECT /*+ MAPJOIN(a) */ * FROM src a RIGHT OUTER JOIN +(select key, named_struct('key', key, 'value', value) as kv from src) b on a.key=b.key AND b.kv.key > 200 limit 10; +'Explain' +'ABSTRACT SYNTAX TREE:' +' (TOK_QUERY (TOK_FROM (TOK_RIGHTOUTERJOIN (TOK_TABREF (TOK_TABNAME src) a) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_FUNCTION named_struct 'key' (TOK_TABLE_OR_COL key) 'value' (TOK_TABLE_OR_COL value)) kv)))) b) (AND (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL b) key)) (> (. (. (TOK_TABLE_OR_COL b) kv) key) 200)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_MAPJOIN (TOK_HINTARGLIST a))) (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_LIMIT 10)))' +'' +'STAGE DEPENDENCIES:' +' Stage-3 is a root stage' +' Stage-1 depends on stages: Stage-3' +' Stage-0 is a root stage' +'' +'STAGE PLANS:' +' Stage: Stage-3' +' Map Reduce Local Work' +' Alias -> Map Local Tables:' +' a ' +' Fetch Operator' +' limit: -1' +' Alias -> Map Local Operator Tree:' +' a ' +' TableScan' +' alias: a' +' HashTable Sink Operator' +' condition expressions:' +' 0 {key} {value}' +' 1 {_col0} {_col1}' +' filter predicates:' +' 0 ' +' 1 {(_col1.key > 200)}' +' handleSkewJoin: false' +' keys:' +' 0 [Column[key]]' +' 1 [Column[_col0]]' +' Position of Big Table: 1' +'' +' Stage: Stage-1' +' Map Reduce' +' Alias -> Map Operator Tree:' +' b:src ' +' TableScan' +' alias: src' +' Select Operator' +' expressions:' +' expr: key' +' type: string' +' expr: named_struct('key',key,'value',value)' +' type: struct' +' outputColumnNames: _col0, _col1' +' Map Join Operator' +' condition map:' +' Right Outer Join0 to 1' +' condition expressions:' +' 0 {key} {value}' +' 1 {_col0} {_col1}' +' filter predicates:' +' 0 ' +' 1 {(_col1.key > 200)}' +' handleSkewJoin: false' +' keys:' +' 0 [Column[key]]' +' 1 [Column[_col0]]' +' outputColumnNames: _col0, _col1, _col4, _col5' +' Position of Big Table: 1' +' Select Operator' +' expressions:' +' expr: _col0' +' type: string' +' expr: _col1' +' type: string' +' expr: _col4' +' type: string' +' expr: _col5' +' type: struct' +' outputColumnNames: _col0, _col1, _col2, _col3' +' Limit' +' File Output Operator' +' compressed: false' +' GlobalTableId: 0' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' Local Work:' +' Map Reduce Local Work' +'' +' Stage: Stage-0' +' Fetch Operator' +' limit: 10' +'' +'' +86 rows selected +>>> SELECT /*+ MAPJOIN(a) */ * FROM src a RIGHT OUTER JOIN +(select key, named_struct('key', key, 'value', value) as kv from src) b on a.key=b.key AND b.kv.key > 200 limit 10; +'key','value','key','kv' +'238','val_238','238','{"key":"238","value":"val_238"}' +'238','val_238','238','{"key":"238","value":"val_238"}' +'NULL','NULL','86','{"key":"86","value":"val_86"}' +'311','val_311','311','{"key":"311","value":"val_311"}' +'311','val_311','311','{"key":"311","value":"val_311"}' +'311','val_311','311','{"key":"311","value":"val_311"}' +'NULL','NULL','27','{"key":"27","value":"val_27"}' +'NULL','NULL','165','{"key":"165","value":"val_165"}' +'NULL','NULL','165','{"key":"165","value":"val_165"}' +'409','val_409','409','{"key":"409","value":"val_409"}' +10 rows selected +>>> +>>> set hive.outerjoin.supports.filters=false; +No rows affected +>>> +>>> EXPLAIN +SELECT /*+ MAPJOIN(a) */ * FROM src a RIGHT OUTER JOIN src b on a.key=b.key AND true limit 10; +'Explain' +'ABSTRACT SYNTAX TREE:' +' (TOK_QUERY (TOK_FROM (TOK_RIGHTOUTERJOIN (TOK_TABREF (TOK_TABNAME src) a) (TOK_TABREF (TOK_TABNAME src) b) (AND (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL b) key)) true))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_MAPJOIN (TOK_HINTARGLIST a))) (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_LIMIT 10)))' +'' +'STAGE DEPENDENCIES:' +' Stage-3 is a root stage' +' Stage-1 depends on stages: Stage-3' +' Stage-0 is a root stage' +'' +'STAGE PLANS:' +' Stage: Stage-3' +' Map Reduce Local Work' +' Alias -> Map Local Tables:' +' a ' +' Fetch Operator' +' limit: -1' +' Alias -> Map Local Operator Tree:' +' a ' +' TableScan' +' alias: a' +' Filter Operator' +' predicate:' +' expr: true' +' type: boolean' +' HashTable Sink Operator' +' condition expressions:' +' 0 {key} {value}' +' 1 {key} {value}' +' handleSkewJoin: false' +' keys:' +' 0 [Column[key]]' +' 1 [Column[key]]' +' Position of Big Table: 1' +'' +' Stage: Stage-1' +' Map Reduce' +' Alias -> Map Operator Tree:' +' b ' +' TableScan' +' alias: b' +' Filter Operator' +' predicate:' +' expr: true' +' type: boolean' +' Map Join Operator' +' condition map:' +' Right Outer Join0 to 1' +' condition expressions:' +' 0 {key} {value}' +' 1 {key} {value}' +' handleSkewJoin: false' +' keys:' +' 0 [Column[key]]' +' 1 [Column[key]]' +' outputColumnNames: _col0, _col1, _col4, _col5' +' Position of Big Table: 1' +' Select Operator' +' expressions:' +' expr: _col0' +' type: string' +' expr: _col1' +' type: string' +' expr: _col4' +' type: string' +' expr: _col5' +' type: string' +' outputColumnNames: _col0, _col1, _col2, _col3' +' Limit' +' File Output Operator' +' compressed: false' +' GlobalTableId: 0' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' Local Work:' +' Map Reduce Local Work' +'' +' Stage: Stage-0' +' Fetch Operator' +' limit: 10' +'' +'' +81 rows selected +>>> SELECT /*+ MAPJOIN(a) */ * FROM src a RIGHT OUTER JOIN src b on a.key=b.key AND true limit 10; +'key','value','key','value' +'238','val_238','238','val_238' +'238','val_238','238','val_238' +'86','val_86','86','val_86' +'311','val_311','311','val_311' +'311','val_311','311','val_311' +'311','val_311','311','val_311' +'27','val_27','27','val_27' +'165','val_165','165','val_165' +'165','val_165','165','val_165' +'409','val_409','409','val_409' +10 rows selected +>>> +>>> EXPLAIN +SELECT /*+ MAPJOIN(a) */ * FROM src a RIGHT OUTER JOIN src b on a.key=b.key AND b.key * 10 < '1000' limit 10; +'Explain' +'ABSTRACT SYNTAX TREE:' +' (TOK_QUERY (TOK_FROM (TOK_RIGHTOUTERJOIN (TOK_TABREF (TOK_TABNAME src) a) (TOK_TABREF (TOK_TABNAME src) b) (AND (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL b) key)) (< (* (. (TOK_TABLE_OR_COL b) key) 10) '1000')))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_MAPJOIN (TOK_HINTARGLIST a))) (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_LIMIT 10)))' +'' +'STAGE DEPENDENCIES:' +' Stage-3 is a root stage' +' Stage-1 depends on stages: Stage-3' +' Stage-0 is a root stage' +'' +'STAGE PLANS:' +' Stage: Stage-3' +' Map Reduce Local Work' +' Alias -> Map Local Tables:' +' a ' +' Fetch Operator' +' limit: -1' +' Alias -> Map Local Operator Tree:' +' a ' +' TableScan' +' alias: a' +' Filter Operator' +' predicate:' +' expr: ((key * 10) < '1000')' +' type: boolean' +' HashTable Sink Operator' +' condition expressions:' +' 0 {key} {value}' +' 1 {key} {value}' +' handleSkewJoin: false' +' keys:' +' 0 [Column[key]]' +' 1 [Column[key]]' +' Position of Big Table: 1' +'' +' Stage: Stage-1' +' Map Reduce' +' Alias -> Map Operator Tree:' +' b ' +' TableScan' +' alias: b' +' Filter Operator' +' predicate:' +' expr: ((key * 10) < '1000')' +' type: boolean' +' Map Join Operator' +' condition map:' +' Right Outer Join0 to 1' +' condition expressions:' +' 0 {key} {value}' +' 1 {key} {value}' +' handleSkewJoin: false' +' keys:' +' 0 [Column[key]]' +' 1 [Column[key]]' +' outputColumnNames: _col0, _col1, _col4, _col5' +' Position of Big Table: 1' +' Select Operator' +' expressions:' +' expr: _col0' +' type: string' +' expr: _col1' +' type: string' +' expr: _col4' +' type: string' +' expr: _col5' +' type: string' +' outputColumnNames: _col0, _col1, _col2, _col3' +' Limit' +' File Output Operator' +' compressed: false' +' GlobalTableId: 0' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' Local Work:' +' Map Reduce Local Work' +'' +' Stage: Stage-0' +' Fetch Operator' +' limit: 10' +'' +'' +81 rows selected +>>> SELECT /*+ MAPJOIN(a) */ * FROM src a RIGHT OUTER JOIN src b on a.key=b.key AND b.key * 10 < '1000' limit 10; +'key','value','key','value' +'86','val_86','86','val_86' +'27','val_27','27','val_27' +'98','val_98','98','val_98' +'98','val_98','98','val_98' +'66','val_66','66','val_66' +'37','val_37','37','val_37' +'37','val_37','37','val_37' +'15','val_15','15','val_15' +'15','val_15','15','val_15' +'82','val_82','82','val_82' +10 rows selected +>>> +>>> EXPLAIN +SELECT /*+ MAPJOIN(a) */ * FROM src a RIGHT OUTER JOIN +(select key, named_struct('key', key, 'value', value) as kv from src) b on a.key=b.key AND b.kv.key > 200 limit 10; +'Explain' +'ABSTRACT SYNTAX TREE:' +' (TOK_QUERY (TOK_FROM (TOK_RIGHTOUTERJOIN (TOK_TABREF (TOK_TABNAME src) a) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_FUNCTION named_struct 'key' (TOK_TABLE_OR_COL key) 'value' (TOK_TABLE_OR_COL value)) kv)))) b) (AND (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL b) key)) (> (. (. (TOK_TABLE_OR_COL b) kv) key) 200)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_MAPJOIN (TOK_HINTARGLIST a))) (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_LIMIT 10)))' +'' +'STAGE DEPENDENCIES:' +' Stage-3 is a root stage' +' Stage-1 depends on stages: Stage-3' +' Stage-0 is a root stage' +'' +'STAGE PLANS:' +' Stage: Stage-3' +' Map Reduce Local Work' +' Alias -> Map Local Tables:' +' a ' +' Fetch Operator' +' limit: -1' +' Alias -> Map Local Operator Tree:' +' a ' +' TableScan' +' alias: a' +' HashTable Sink Operator' +' condition expressions:' +' 0 {key} {value}' +' 1 {_col0} {_col1}' +' handleSkewJoin: false' +' keys:' +' 0 [Column[key]]' +' 1 [Column[_col0]]' +' Position of Big Table: 1' +'' +' Stage: Stage-1' +' Map Reduce' +' Alias -> Map Operator Tree:' +' b:src ' +' TableScan' +' alias: src' +' Select Operator' +' expressions:' +' expr: key' +' type: string' +' expr: named_struct('key',key,'value',value)' +' type: struct' +' outputColumnNames: _col0, _col1' +' Filter Operator' +' predicate:' +' expr: (_col1.key > 200)' +' type: boolean' +' Map Join Operator' +' condition map:' +' Right Outer Join0 to 1' +' condition expressions:' +' 0 {key} {value}' +' 1 {_col0} {_col1}' +' handleSkewJoin: false' +' keys:' +' 0 [Column[key]]' +' 1 [Column[_col0]]' +' outputColumnNames: _col0, _col1, _col4, _col5' +' Position of Big Table: 1' +' Select Operator' +' expressions:' +' expr: _col0' +' type: string' +' expr: _col1' +' type: string' +' expr: _col4' +' type: string' +' expr: _col5' +' type: struct' +' outputColumnNames: _col0, _col1, _col2, _col3' +' Limit' +' File Output Operator' +' compressed: false' +' GlobalTableId: 0' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' Local Work:' +' Map Reduce Local Work' +'' +' Stage: Stage-0' +' Fetch Operator' +' limit: 10' +'' +'' +84 rows selected +>>> SELECT /*+ MAPJOIN(a) */ * FROM src a RIGHT OUTER JOIN +(select key, named_struct('key', key, 'value', value) as kv from src) b on a.key=b.key AND b.kv.key > 200 limit 10; +'key','value','key','kv' +'238','val_238','238','{"key":"238","value":"val_238"}' +'238','val_238','238','{"key":"238","value":"val_238"}' +'311','val_311','311','{"key":"311","value":"val_311"}' +'311','val_311','311','{"key":"311","value":"val_311"}' +'311','val_311','311','{"key":"311","value":"val_311"}' +'409','val_409','409','{"key":"409","value":"val_409"}' +'409','val_409','409','{"key":"409","value":"val_409"}' +'409','val_409','409','{"key":"409","value":"val_409"}' +'255','val_255','255','{"key":"255","value":"val_255"}' +'255','val_255','255','{"key":"255","value":"val_255"}' +10 rows selected >>> !record diff --git ql/src/test/results/beelinepositive/mapjoin_subquery2.q.out ql/src/test/results/beelinepositive/mapjoin_subquery2.q.out index 9a813bb..adc919c 100644 --- ql/src/test/results/beelinepositive/mapjoin_subquery2.q.out +++ ql/src/test/results/beelinepositive/mapjoin_subquery2.q.out @@ -26,45 +26,44 @@ No rows affected >>> load data local inpath '../data/files/z.txt' INTO TABLE z; No rows affected >>> ->>> SELECT subq.key1, subq.value1, subq.key2, subq.value2, z.id, z.name -FROM -(SELECT x.id as key1, x.name as value1, y.id as key2, y.name as value2 -FROM y JOIN x ON (x.id = y.id)) subq -JOIN z ON (subq.key1 = z.id); -'key1','value1','key2','value2','id','name' -'2','Joe','2','Tie','2','Tie' -'2','Hank','2','Tie','2','Tie' -2 rows selected +>>> set hive.auto.convert.join=true; +No rows affected +>>> set hive.auto.convert.join.noconditionaltask=true; +No rows affected +>>> set hive.auto.convert.join.noconditionaltask.size=10000; +No rows affected +>>> +>>> -- Since the inputs are small, it should be automatically converted to mapjoin >>> >>> EXPLAIN -SELECT /*+ MAPJOIN(z) */ subq.key1, subq.value1, subq.key2, subq.value2, z.id, z.name +SELECT subq.key1, subq.value1, subq.key2, subq.value2, z.id, z.name FROM -(SELECT /*+ MAPJOIN(x) */ x.id as key1, x.name as value1, y.id as key2, y.name as value2 +(SELECT x.id as key1, x.name as value1, y.id as key2, y.name as value2 FROM y JOIN x ON (x.id = y.id)) subq JOIN z ON (subq.key1 = z.id); 'Explain' 'ABSTRACT SYNTAX TREE:' -' (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME y)) (TOK_TABREF (TOK_TABNAME x)) (= (. (TOK_TABLE_OR_COL x) id) (. (TOK_TABLE_OR_COL y) id)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_MAPJOIN (TOK_HINTARGLIST x))) (TOK_SELEXPR (. (TOK_TABLE_OR_COL x) id) key1) (TOK_SELEXPR (. (TOK_TABLE_OR_COL x) name) value1) (TOK_SELEXPR (. (TOK_TABLE_OR_COL y) id) key2) (TOK_SELEXPR (. (TOK_TABLE_OR_COL y) name) value2)))) subq) (TOK_TABREF (TOK_TABNAME z)) (= (. (TOK_TABLE_OR_COL subq) key1) (. (TOK_TABLE_OR_COL z) id)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_MAPJOIN (TOK_HINTARGLIST z))) (TOK_SELEXPR (. (TOK_TABLE_OR_COL subq) key1)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL subq) value1)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL subq) key2)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL subq) value2)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL z) id)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL z) name)))))' +' (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_TABREF (TOK_TABNAME y)) (TOK_TABREF (TOK_TABNAME x)) (= (. (TOK_TABLE_OR_COL x) id) (. (TOK_TABLE_OR_COL y) id)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL x) id) key1) (TOK_SELEXPR (. (TOK_TABLE_OR_COL x) name) value1) (TOK_SELEXPR (. (TOK_TABLE_OR_COL y) id) key2) (TOK_SELEXPR (. (TOK_TABLE_OR_COL y) name) value2)))) subq) (TOK_TABREF (TOK_TABNAME z)) (= (. (TOK_TABLE_OR_COL subq) key1) (. (TOK_TABLE_OR_COL z) id)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL subq) key1)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL subq) value1)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL subq) key2)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL subq) value2)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL z) id)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL z) name)))))' '' 'STAGE DEPENDENCIES:' -' Stage-4 is a root stage' -' Stage-1 depends on stages: Stage-4' +' Stage-7 is a root stage' +' Stage-6 depends on stages: Stage-7' ' Stage-0 is a root stage' '' 'STAGE PLANS:' -' Stage: Stage-4' +' Stage: Stage-7' ' Map Reduce Local Work' ' Alias -> Map Local Tables:' -' subq:x ' +' subq:y ' ' Fetch Operator' ' limit: -1' ' z ' ' Fetch Operator' ' limit: -1' ' Alias -> Map Local Operator Tree:' -' subq:x ' +' subq:y ' ' TableScan' -' alias: x' +' alias: y' ' HashTable Sink Operator' ' condition expressions:' ' 0 {id} {name}' @@ -73,7 +72,7 @@ JOIN z ON (subq.key1 = z.id); ' keys:' ' 0 [Column[id]]' ' 1 [Column[id]]' -' Position of Big Table: 0' +' Position of Big Table: 1' ' z ' ' TableScan' ' alias: z' @@ -87,12 +86,12 @@ JOIN z ON (subq.key1 = z.id); ' 1 [Column[id]]' ' Position of Big Table: 0' '' -' Stage: Stage-1' +' Stage: Stage-6' ' Map Reduce' ' Alias -> Map Operator Tree:' -' subq:y ' +' subq:x ' ' TableScan' -' alias: y' +' alias: x' ' Map Join Operator' ' condition map:' ' Inner Join 0 to 1' @@ -104,29 +103,18 @@ JOIN z ON (subq.key1 = z.id); ' 0 [Column[id]]' ' 1 [Column[id]]' ' outputColumnNames: _col0, _col1, _col4, _col5' -' Position of Big Table: 0' +' Position of Big Table: 1' ' Select Operator' ' expressions:' -' expr: _col0' +' expr: _col5' ' type: int' -' expr: _col1' -' type: string' ' expr: _col4' ' type: string' -' expr: _col5' +' expr: _col0' ' type: int' -' outputColumnNames: _col0, _col1, _col4, _col5' -' Select Operator' -' expressions:' -' expr: _col5' -' type: int' -' expr: _col4' -' type: string' -' expr: _col0' -' type: int' -' expr: _col1' -' type: string' -' outputColumnNames: _col0, _col1, _col2, _col3' +' expr: _col1' +' type: string' +' outputColumnNames: _col0, _col1, _col2, _col3' ' Map Join Operator' ' condition map:' ' Inner Join 0 to 1' @@ -154,27 +142,12 @@ JOIN z ON (subq.key1 = z.id); ' expr: _col5' ' type: string' ' outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5' -' Select Operator' -' expressions:' -' expr: _col0' -' type: int' -' expr: _col1' -' type: string' -' expr: _col2' -' type: int' -' expr: _col3' -' type: string' -' expr: _col4' -' type: int' -' expr: _col5' -' type: string' -' outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5' -' File Output Operator' -' compressed: false' -' GlobalTableId: 0' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' File Output Operator' +' compressed: false' +' GlobalTableId: 0' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' ' Local Work:' ' Map Reduce Local Work' '' @@ -183,11 +156,11 @@ JOIN z ON (subq.key1 = z.id); ' limit: -1' '' '' -140 rows selected +114 rows selected >>> ->>> SELECT /*+ MAPJOIN(z) */ subq.key1, subq.value1, subq.key2, subq.value2, z.id, z.name +>>> SELECT subq.key1, subq.value1, subq.key2, subq.value2, z.id, z.name FROM -(SELECT /*+ MAPJOIN(x) */ x.id as key1, x.name as value1, y.id as key2, y.name as value2 +(SELECT x.id as key1, x.name as value1, y.id as key2, y.name as value2 FROM y JOIN x ON (x.id = y.id)) subq JOIN z ON (subq.key1 = z.id); 'key1','value1','key2','value2','id','name' diff --git ql/src/test/results/beelinepositive/merge1.q.out ql/src/test/results/beelinepositive/merge1.q.out index 4f5d4a6..3bc5890 100644 --- ql/src/test/results/beelinepositive/merge1.q.out +++ ql/src/test/results/beelinepositive/merge1.q.out @@ -68,26 +68,19 @@ select key, count(1) from src group by key; ' outputColumnNames: _col0, _col1' ' Select Operator' ' expressions:' -' expr: _col0' -' type: string' -' expr: _col1' -' type: bigint' +' expr: UDFToInteger(_col0)' +' type: int' +' expr: UDFToInteger(_col1)' +' type: int' ' outputColumnNames: _col0, _col1' -' Select Operator' -' expressions:' -' expr: UDFToInteger(_col0)' -' type: int' -' expr: UDFToInteger(_col1)' -' type: int' -' outputColumnNames: _col0, _col1' -' File Output Operator' -' compressed: false' -' GlobalTableId: 1' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: merge1.dest1' +' File Output Operator' +' compressed: false' +' GlobalTableId: 1' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' name: merge1.dest1' '' ' Stage: Stage-7' ' Conditional Operator' @@ -144,7 +137,7 @@ select key, count(1) from src group by key; ' destination: pfile:!!{hive.exec.scratchdir}!!' '' '' -134 rows selected +127 rows selected >>> >>> insert overwrite table dest1 select key, count(1) from src group by key; diff --git ql/src/test/results/beelinepositive/parallel.q.out ql/src/test/results/beelinepositive/parallel.q.out index 5532f84..4420a19 100644 --- ql/src/test/results/beelinepositive/parallel.q.out +++ ql/src/test/results/beelinepositive/parallel.q.out @@ -81,19 +81,12 @@ insert overwrite table src_b select s.key, s.value group by s.key, s.value; ' expr: _col1' ' type: string' ' outputColumnNames: _col0, _col1' -' Select Operator' -' expressions:' -' expr: _col0' -' type: string' -' expr: _col1' -' type: string' -' outputColumnNames: _col0, _col1' -' File Output Operator' -' compressed: false' -' GlobalTableId: 0' -' table:' -' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' +' File Output Operator' +' compressed: false' +' GlobalTableId: 0' +' table:' +' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' '' ' Stage: Stage-3' ' Map Reduce' @@ -190,7 +183,7 @@ insert overwrite table src_b select s.key, s.value group by s.key, s.value; ' Stats-Aggr Operator' '' '' -173 rows selected +166 rows selected >>> >>> from (select key, value from src group by key, value) s insert overwrite table src_a select s.key, s.value group by s.key, s.value diff --git ql/src/test/results/beelinepositive/ppd1.q.out ql/src/test/results/beelinepositive/ppd1.q.out index 96c72ab..b28f191 100644 --- ql/src/test/results/beelinepositive/ppd1.q.out +++ ql/src/test/results/beelinepositive/ppd1.q.out @@ -26,28 +26,24 @@ SELECT src.key as c3 from src where src.key > '2'; ' predicate:' ' expr: (key > '2')' ' type: boolean' -' Filter Operator' -' predicate:' -' expr: (key > '2')' -' type: boolean' -' Select Operator' -' expressions:' -' expr: key' -' type: string' -' outputColumnNames: _col0' -' File Output Operator' -' compressed: false' -' GlobalTableId: 0' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' Select Operator' +' expressions:' +' expr: key' +' type: string' +' outputColumnNames: _col0' +' File Output Operator' +' compressed: false' +' GlobalTableId: 0' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' '' ' Stage: Stage-0' ' Fetch Operator' ' limit: -1' '' '' -39 rows selected +35 rows selected >>> >>> SELECT src.key as c3 from src where src.key > '2'; 'c3' diff --git ql/src/test/results/beelinepositive/ppd_constant_expr.q.out ql/src/test/results/beelinepositive/ppd_constant_expr.q.out index 3f00589..b419314 100644 --- ql/src/test/results/beelinepositive/ppd_constant_expr.q.out +++ ql/src/test/results/beelinepositive/ppd_constant_expr.q.out @@ -36,28 +36,19 @@ INSERT OVERWRITE TABLE ppd_constant_expr SELECT 4 + NULL, src1.key - NULL, NULL ' expressions:' ' expr: (4 + null)' ' type: int' -' expr: (key - null)' +' expr: UDFToInteger((key - null))' +' type: int' +' expr: UDFToDouble((null + null))' ' type: double' -' expr: (null + null)' -' type: tinyint' ' outputColumnNames: _col0, _col1, _col2' -' Select Operator' -' expressions:' -' expr: _col0' -' type: int' -' expr: UDFToInteger(_col1)' -' type: int' -' expr: UDFToDouble(_col2)' -' type: double' -' outputColumnNames: _col0, _col1, _col2' -' File Output Operator' -' compressed: false' -' GlobalTableId: 1' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: ppd_constant_expr.ppd_constant_expr' +' File Output Operator' +' compressed: false' +' GlobalTableId: 1' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' name: ppd_constant_expr.ppd_constant_expr' '' ' Stage: Stage-7' ' Conditional Operator' @@ -114,7 +105,7 @@ INSERT OVERWRITE TABLE ppd_constant_expr SELECT 4 + NULL, src1.key - NULL, NULL ' destination: pfile:!!{hive.exec.scratchdir}!!' '' '' -102 rows selected +93 rows selected >>> >>> FROM src1 INSERT OVERWRITE TABLE ppd_constant_expr SELECT 4 + NULL, src1.key - NULL, NULL + NULL; @@ -181,28 +172,19 @@ INSERT OVERWRITE TABLE ppd_constant_expr SELECT 4 + NULL, src1.key - NULL, NULL ' expressions:' ' expr: (4 + null)' ' type: int' -' expr: (key - null)' +' expr: UDFToInteger((key - null))' +' type: int' +' expr: UDFToDouble((null + null))' ' type: double' -' expr: (null + null)' -' type: tinyint' ' outputColumnNames: _col0, _col1, _col2' -' Select Operator' -' expressions:' -' expr: _col0' -' type: int' -' expr: UDFToInteger(_col1)' -' type: int' -' expr: UDFToDouble(_col2)' -' type: double' -' outputColumnNames: _col0, _col1, _col2' -' File Output Operator' -' compressed: false' -' GlobalTableId: 1' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: ppd_constant_expr.ppd_constant_expr' +' File Output Operator' +' compressed: false' +' GlobalTableId: 1' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' name: ppd_constant_expr.ppd_constant_expr' '' ' Stage: Stage-7' ' Conditional Operator' @@ -259,7 +241,7 @@ INSERT OVERWRITE TABLE ppd_constant_expr SELECT 4 + NULL, src1.key - NULL, NULL ' destination: pfile:!!{hive.exec.scratchdir}!!' '' '' -102 rows selected +93 rows selected >>> >>> FROM src1 INSERT OVERWRITE TABLE ppd_constant_expr SELECT 4 + NULL, src1.key - NULL, NULL + NULL; diff --git ql/src/test/results/beelinepositive/quote1.q.out ql/src/test/results/beelinepositive/quote1.q.out index fe9add6..d5fd523 100644 --- ql/src/test/results/beelinepositive/quote1.q.out +++ ql/src/test/results/beelinepositive/quote1.q.out @@ -33,26 +33,19 @@ INSERT OVERWRITE TABLE dest1 PARTITION(`table`='2008-04-08') SELECT src.key as ` ' type: boolean' ' Select Operator' ' expressions:' -' expr: key' -' type: string' +' expr: UDFToInteger(key)' +' type: int' ' expr: value' ' type: string' ' outputColumnNames: _col0, _col1' -' Select Operator' -' expressions:' -' expr: UDFToInteger(_col0)' -' type: int' -' expr: _col1' -' type: string' -' outputColumnNames: _col0, _col1' -' File Output Operator' -' compressed: false' -' GlobalTableId: 1' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: quote1.dest1' +' File Output Operator' +' compressed: false' +' GlobalTableId: 1' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' name: quote1.dest1' '' ' Stage: Stage-7' ' Conditional Operator' @@ -111,7 +104,7 @@ INSERT OVERWRITE TABLE dest1 PARTITION(`table`='2008-04-08') SELECT src.key as ` ' destination: pfile:!!{hive.exec.scratchdir}!!' '' '' -104 rows selected +97 rows selected >>> >>> EXPLAIN SELECT `int`.`location`, `int`.`type`, `int`.`table` FROM dest1 `int` WHERE `int`.`table` = '2008-04-08'; diff --git ql/src/test/results/beelinepositive/sample1.q.out ql/src/test/results/beelinepositive/sample1.q.out index ee1fe87..a7ce7e3 100644 --- ql/src/test/results/beelinepositive/sample1.q.out +++ ql/src/test/results/beelinepositive/sample1.q.out @@ -37,8 +37,8 @@ WHERE s.ds='2008-04-08' and s.hr='11'; ' type: boolean' ' Select Operator' ' expressions:' -' expr: key' -' type: string' +' expr: UDFToInteger(key)' +' type: int' ' expr: value' ' type: string' ' expr: ds' @@ -46,43 +46,32 @@ WHERE s.ds='2008-04-08' and s.hr='11'; ' expr: hr' ' type: string' ' outputColumnNames: _col0, _col1, _col2, _col3' -' Select Operator' -' expressions:' -' expr: UDFToInteger(_col0)' -' type: int' -' expr: _col1' -' type: string' -' expr: _col2' -' type: string' -' expr: _col3' -' type: string' -' outputColumnNames: _col0, _col1, _col2, _col3' -' File Output Operator' -' compressed: false' -' GlobalTableId: 1' -' directory: pfile:!!{hive.exec.scratchdir}!!' -' NumFilesPerFileSink: 1' -' Stats Publishing Key Prefix: pfile:!!{hive.exec.scratchdir}!!' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' properties:' -' bucket_count -1' -' columns key,value,dt,hr' -' columns.types int:string:string:string' -' file.inputformat org.apache.hadoop.mapred.TextInputFormat' -' file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' location !!{hive.metastore.warehouse.dir}!!/sample1.db/dest1' -' name sample1.dest1' -' serialization.ddl struct dest1 { i32 key, string value, string dt, string hr}' -' serialization.format 1' -' serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' transient_lastDdlTime !!UNIXTIME!!' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: sample1.dest1' -' TotalFiles: 1' -' GatherStats: true' -' MultiFileSpray: false' +' File Output Operator' +' compressed: false' +' GlobalTableId: 1' +' directory: pfile:!!{hive.exec.scratchdir}!!' +' NumFilesPerFileSink: 1' +' Stats Publishing Key Prefix: pfile:!!{hive.exec.scratchdir}!!' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' properties:' +' bucket_count -1' +' columns key,value,dt,hr' +' columns.types int:string:string:string' +' file.inputformat org.apache.hadoop.mapred.TextInputFormat' +' file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' location !!{hive.metastore.warehouse.dir}!!/sample1.db/dest1' +' name sample1.dest1' +' serialization.ddl struct dest1 { i32 key, string value, string dt, string hr}' +' serialization.format 1' +' serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' transient_lastDdlTime !!UNIXTIME!!' +' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' name: sample1.dest1' +' TotalFiles: 1' +' GatherStats: true' +' MultiFileSpray: false' ' Needs Tagging: false' ' Path -> Alias:' ' !!{hive.metastore.warehouse.dir}!!/sample1.db/srcpart/ds=2008-04-08/hr=11 [s]' @@ -104,7 +93,6 @@ WHERE s.ds='2008-04-08' and s.hr='11'; ' location !!{hive.metastore.warehouse.dir}!!/sample1.db/srcpart/ds=2008-04-08/hr=11' ' name sample1.srcpart' ' numFiles 1' -' numPartitions 4' ' numRows 0' ' partition_columns ds/hr' ' rawDataSize 0' @@ -331,7 +319,7 @@ WHERE s.ds='2008-04-08' and s.hr='11'; ' destination: pfile:!!{hive.exec.scratchdir}!!' '' '' -322 rows selected +310 rows selected >>> >>> INSERT OVERWRITE TABLE dest1 SELECT s.* FROM srcpart TABLESAMPLE (BUCKET 1 OUT OF 1 ON rand()) s diff --git ql/src/test/results/beelinepositive/semijoin.q.out ql/src/test/results/beelinepositive/semijoin.q.out index 38da95b..a9e519a 100644 --- ql/src/test/results/beelinepositive/semijoin.q.out +++ ql/src/test/results/beelinepositive/semijoin.q.out @@ -723,27 +723,22 @@ No rows selected ' expr: key' ' type: int' ' outputColumnNames: _col0' -' Select Operator' -' expressions:' +' Group By Operator' +' bucketGroup: false' +' keys:' ' expr: _col0' ' type: int' +' mode: hash' ' outputColumnNames: _col0' -' Group By Operator' -' bucketGroup: false' -' keys:' +' Reduce Output Operator' +' key expressions:' ' expr: _col0' ' type: int' -' mode: hash' -' outputColumnNames: _col0' -' Reduce Output Operator' -' key expressions:' -' expr: _col0' -' type: int' -' sort order: +' -' Map-reduce partition columns:' -' expr: _col0' -' type: int' -' tag: 1' +' sort order: +' +' Map-reduce partition columns:' +' expr: _col0' +' type: int' +' tag: 1' ' Reduce Operator Tree:' ' Join Operator' ' condition map:' @@ -792,7 +787,7 @@ No rows selected ' limit: -1' '' '' -108 rows selected +103 rows selected >>> select a.value from t1 a left semi join (select key from t3 where key > 5) b on a.key = b.key sort by a.value; 'value' 'val_10' @@ -843,31 +838,24 @@ No rows selected ' expr: value' ' type: string' ' outputColumnNames: _col0, _col1' -' Select Operator' -' expressions:' +' Group By Operator' +' bucketGroup: false' +' keys:' ' expr: _col0' ' type: int' ' expr: _col1' ' type: string' +' mode: hash' ' outputColumnNames: _col0, _col1' -' Group By Operator' -' bucketGroup: false' -' keys:' +' Reduce Output Operator' +' key expressions:' ' expr: _col0' ' type: int' -' expr: _col1' -' type: string' -' mode: hash' -' outputColumnNames: _col0, _col1' -' Reduce Output Operator' -' key expressions:' -' expr: _col0' -' type: int' -' sort order: +' -' Map-reduce partition columns:' -' expr: _col0' -' type: int' -' tag: 1' +' sort order: +' +' Map-reduce partition columns:' +' expr: _col0' +' type: int' +' tag: 1' ' Reduce Operator Tree:' ' Join Operator' ' condition map:' @@ -916,7 +904,7 @@ No rows selected ' limit: -1' '' '' -114 rows selected +107 rows selected >>> select a.value from t1 a left semi join (select key , value from t2 where key > 5) b on a.key = b.key and b.value <= 'val_20' sort by a.value ; 'value' No rows selected @@ -964,27 +952,22 @@ No rows selected ' expr: key' ' type: int' ' outputColumnNames: _col0' -' Select Operator' -' expressions:' +' Group By Operator' +' bucketGroup: false' +' keys:' ' expr: _col0' ' type: int' +' mode: hash' ' outputColumnNames: _col0' -' Group By Operator' -' bucketGroup: false' -' keys:' +' Reduce Output Operator' +' key expressions:' ' expr: _col0' ' type: int' -' mode: hash' -' outputColumnNames: _col0' -' Reduce Output Operator' -' key expressions:' -' expr: _col0' -' type: int' -' sort order: +' -' Map-reduce partition columns:' -' expr: _col0' -' type: int' -' tag: 1' +' sort order: +' +' Map-reduce partition columns:' +' expr: _col0' +' type: int' +' tag: 1' ' Reduce Operator Tree:' ' Join Operator' ' condition map:' @@ -1039,7 +1022,7 @@ No rows selected ' limit: -1' '' '' -116 rows selected +111 rows selected >>> select * from t2 a left semi join (select key , value from t1 where key > 2) b on a.key = b.key sort by a.key, a.value; 'key','value' '4','val_2' @@ -1055,13 +1038,12 @@ No rows selected ' (TOK_QUERY (TOK_FROM (TOK_LEFTSEMIJOIN (TOK_TABREF (TOK_TABNAME t3) a) (TOK_TABREF (TOK_TABNAME t1) b) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL b) key)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_MAPJOIN (TOK_HINTARGLIST b))) (TOK_SELEXPR (. (TOK_TABLE_OR_COL a) key))) (TOK_SORTBY (TOK_TABSORTCOLNAMEASC (. (TOK_TABLE_OR_COL a) key)))))' '' 'STAGE DEPENDENCIES:' -' Stage-4 is a root stage' -' Stage-1 depends on stages: Stage-4' -' Stage-2 depends on stages: Stage-1' +' Stage-3 is a root stage' +' Stage-1 depends on stages: Stage-3' ' Stage-0 is a root stage' '' 'STAGE PLANS:' -' Stage: Stage-4' +' Stage: Stage-3' ' Map Reduce Local Work' ' Alias -> Map Local Tables:' ' b ' @@ -1111,38 +1093,22 @@ No rows selected ' 1 [Column[_col0]]' ' outputColumnNames: _col0' ' Position of Big Table: 0' -' File Output Operator' -' compressed: false' -' GlobalTableId: 0' -' table:' -' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' -' Local Work:' -' Map Reduce Local Work' -'' -' Stage: Stage-2' -' Map Reduce' -' Alias -> Map Operator Tree:' -' file:!!{hive.exec.local.scratchdir}!! ' -' Select Operator' -' expressions:' -' expr: _col0' -' type: int' -' outputColumnNames: _col0' -' Select Operator' -' expressions:' -' expr: _col0' -' type: int' -' outputColumnNames: _col0' -' Reduce Output Operator' -' key expressions:' -' expr: _col0' -' type: int' -' sort order: +' -' tag: -1' -' value expressions:' +' Select Operator' +' expressions:' ' expr: _col0' ' type: int' +' outputColumnNames: _col0' +' Reduce Output Operator' +' key expressions:' +' expr: _col0' +' type: int' +' sort order: +' +' tag: -1' +' value expressions:' +' expr: _col0' +' type: int' +' Local Work:' +' Map Reduce Local Work' ' Reduce Operator Tree:' ' Extract' ' File Output Operator' @@ -1157,7 +1123,7 @@ No rows selected ' limit: -1' '' '' -106 rows selected +89 rows selected >>> select /*+ mapjoin(b) */ a.key from t3 a left semi join t1 b on a.key = b.key sort by a.key; 'key' '0' @@ -1461,7 +1427,6 @@ No rows selected '' 'STAGE DEPENDENCIES:' ' Stage-1 is a root stage' -' Stage-2 depends on stages: Stage-1' ' Stage-0 is a root stage' '' 'STAGE PLANS:' @@ -1541,41 +1506,15 @@ No rows selected ' compressed: false' ' GlobalTableId: 0' ' table:' -' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' -'' -' Stage: Stage-2' -' Map Reduce' -' Alias -> Map Operator Tree:' -' file:!!{hive.exec.local.scratchdir}!! ' -' Reduce Output Operator' -' key expressions:' -' expr: _col0' -' type: int' -' expr: _col1' -' type: string' -' sort order: ++' -' tag: -1' -' value expressions:' -' expr: _col0' -' type: int' -' expr: _col1' -' type: string' -' Reduce Operator Tree:' -' Extract' -' File Output Operator' -' compressed: false' -' GlobalTableId: 0' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' '' ' Stage: Stage-0' ' Fetch Operator' ' limit: -1' '' '' -119 rows selected +92 rows selected >>> select * from t3 a left semi join t1 b on a.key = b.key and a.value=b.value sort by a.key, a.value; 'key','value' '0','val_0' @@ -1600,13 +1539,12 @@ No rows selected ' (TOK_QUERY (TOK_FROM (TOK_LEFTSEMIJOIN (TOK_LEFTSEMIJOIN (TOK_TABREF (TOK_TABNAME t3) a) (TOK_TABREF (TOK_TABNAME t1) b) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL b) key))) (TOK_TABREF (TOK_TABNAME t2) c) (= (. (TOK_TABLE_OR_COL a) key) (. (TOK_TABLE_OR_COL c) key)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_MAPJOIN (TOK_HINTARGLIST b c))) (TOK_SELEXPR (. (TOK_TABLE_OR_COL a) key))) (TOK_SORTBY (TOK_TABSORTCOLNAMEASC (. (TOK_TABLE_OR_COL a) key)))))' '' 'STAGE DEPENDENCIES:' -' Stage-5 is a root stage' -' Stage-1 depends on stages: Stage-5' -' Stage-2 depends on stages: Stage-1' +' Stage-4 is a root stage' +' Stage-1 depends on stages: Stage-4' ' Stage-0 is a root stage' '' 'STAGE PLANS:' -' Stage: Stage-5' +' Stage: Stage-4' ' Map Reduce Local Work' ' Alias -> Map Local Tables:' ' b ' @@ -1690,38 +1628,22 @@ No rows selected ' 2 [Column[_col0]]' ' outputColumnNames: _col0' ' Position of Big Table: 0' -' File Output Operator' -' compressed: false' -' GlobalTableId: 0' -' table:' -' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' -' Local Work:' -' Map Reduce Local Work' -'' -' Stage: Stage-2' -' Map Reduce' -' Alias -> Map Operator Tree:' -' file:!!{hive.exec.local.scratchdir}!! ' -' Select Operator' -' expressions:' -' expr: _col0' -' type: int' -' outputColumnNames: _col0' -' Select Operator' -' expressions:' -' expr: _col0' -' type: int' -' outputColumnNames: _col0' -' Reduce Output Operator' -' key expressions:' -' expr: _col0' -' type: int' -' sort order: +' -' tag: -1' -' value expressions:' +' Select Operator' +' expressions:' ' expr: _col0' ' type: int' +' outputColumnNames: _col0' +' Reduce Output Operator' +' key expressions:' +' expr: _col0' +' type: int' +' sort order: +' +' tag: -1' +' value expressions:' +' expr: _col0' +' type: int' +' Local Work:' +' Map Reduce Local Work' ' Reduce Operator Tree:' ' Extract' ' File Output Operator' @@ -1736,7 +1658,7 @@ No rows selected ' limit: -1' '' '' -140 rows selected +123 rows selected >>> select /*+ mapjoin(b, c) */ a.key from t3 a left semi join t1 b on a.key = b.key left semi join t2 c on a.key = c.key sort by a.key; 'key' '0' @@ -1762,7 +1684,6 @@ No rows selected '' 'STAGE DEPENDENCIES:' ' Stage-1 is a root stage' -' Stage-2 depends on stages: Stage-1' ' Stage-0 is a root stage' '' 'STAGE PLANS:' @@ -1840,37 +1761,15 @@ No rows selected ' compressed: false' ' GlobalTableId: 0' ' table:' -' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' -'' -' Stage: Stage-2' -' Map Reduce' -' Alias -> Map Operator Tree:' -' file:!!{hive.exec.local.scratchdir}!! ' -' Reduce Output Operator' -' key expressions:' -' expr: _col0' -' type: int' -' sort order: +' -' tag: -1' -' value expressions:' -' expr: _col0' -' type: int' -' Reduce Operator Tree:' -' Extract' -' File Output Operator' -' compressed: false' -' GlobalTableId: 0' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' '' ' Stage: Stage-0' ' Fetch Operator' ' limit: -1' '' '' -113 rows selected +90 rows selected >>> select a.key from t3 a left outer join t1 b on a.key = b.key left semi join t2 c on b.key = c.key sort by a.key; 'key' '0' @@ -2206,7 +2105,6 @@ No rows selected '' 'STAGE DEPENDENCIES:' ' Stage-1 is a root stage' -' Stage-2 depends on stages: Stage-1' ' Stage-0 is a root stage' '' 'STAGE PLANS:' @@ -2284,37 +2182,15 @@ No rows selected ' compressed: false' ' GlobalTableId: 0' ' table:' -' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' -'' -' Stage: Stage-2' -' Map Reduce' -' Alias -> Map Operator Tree:' -' file:!!{hive.exec.local.scratchdir}!! ' -' Reduce Output Operator' -' key expressions:' -' expr: _col0' -' type: int' -' sort order: +' -' tag: -1' -' value expressions:' -' expr: _col0' -' type: int' -' Reduce Operator Tree:' -' Extract' -' File Output Operator' -' compressed: false' -' GlobalTableId: 0' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' '' ' Stage: Stage-0' ' Fetch Operator' ' limit: -1' '' '' -113 rows selected +90 rows selected >>> select a.key from t3 a left semi join t2 b on a.key = b.key left outer join t1 c on a.key = c.key sort by a.key; 'key' '0' diff --git ql/src/test/results/beelinepositive/show_functions.q.out ql/src/test/results/beelinepositive/show_functions.q.out index 93cde28..1236889 100644 --- ql/src/test/results/beelinepositive/show_functions.q.out +++ ql/src/test/results/beelinepositive/show_functions.q.out @@ -47,12 +47,14 @@ Saving all output to "!!{outputDirectory}!!/show_functions.q.raw". Enter "record 'covar_pop' 'covar_samp' 'create_union' +'cume_dist' 'date_add' 'date_sub' 'datediff' 'day' 'dayofmonth' 'degrees' +'dense_rank' 'div' 'e' 'elt' @@ -64,6 +66,7 @@ Saving all output to "!!{outputDirectory}!!/show_functions.q.raw". Enter "record 'explode' 'field' 'find_in_set' +'first_value' 'floor' 'format_number' 'from_unixtime' @@ -83,7 +86,10 @@ Saving all output to "!!{outputDirectory}!!/show_functions.q.raw". Enter "record 'isnull' 'java_method' 'json_tuple' +'lag' +'last_value' 'lcase' +'lead' 'length' 'like' 'ln' @@ -105,9 +111,12 @@ Saving all output to "!!{outputDirectory}!!/show_functions.q.raw". Enter "record 'negative' 'ngrams' 'not' +'ntile' +'nvl' 'or' 'parse_url' 'parse_url_tuple' +'percent_rank' 'percentile' 'percentile_approx' 'pi' @@ -118,7 +127,9 @@ Saving all output to "!!{outputDirectory}!!/show_functions.q.raw". Enter "record 'printf' 'radians' 'rand' +'rank' 'reflect' +'reflect2' 'regexp' 'regexp_extract' 'regexp_replace' @@ -126,6 +137,7 @@ Saving all output to "!!{outputDirectory}!!/show_functions.q.raw". Enter "record 'reverse' 'rlike' 'round' +'row_number' 'rpad' 'rtrim' 'second' @@ -149,6 +161,7 @@ Saving all output to "!!{outputDirectory}!!/show_functions.q.raw". Enter "record 'sum' 'tan' 'to_date' +'to_unix_timestamp' 'to_utc_timestamp' 'translate' 'trim' @@ -174,7 +187,7 @@ Saving all output to "!!{outputDirectory}!!/show_functions.q.raw". Enter "record 'year' '|' '~' -172 rows selected +185 rows selected >>> >>> SHOW FUNCTIONS '^c.*'; 'tab_name' @@ -194,7 +207,8 @@ Saving all output to "!!{outputDirectory}!!/show_functions.q.raw". Enter "record 'covar_pop' 'covar_samp' 'create_union' -16 rows selected +'cume_dist' +17 rows selected >>> >>> SHOW FUNCTIONS '.*e$'; 'tab_name' @@ -203,15 +217,18 @@ Saving all output to "!!{outputDirectory}!!/show_functions.q.raw". Enter "record 'coalesce' 'e' 'explode' +'first_value' 'from_unixtime' 'in_file' 'inline' 'json_tuple' +'last_value' 'lcase' 'like' 'locate' 'minute' 'negative' +'ntile' 'parse_url_tuple' 'percentile' 'positive' @@ -225,7 +242,7 @@ Saving all output to "!!{outputDirectory}!!/show_functions.q.raw". Enter "record 'ucase' 'variance' 'xpath_double' -27 rows selected +30 rows selected >>> >>> SHOW FUNCTIONS 'log.*'; 'tab_name' diff --git ql/src/test/results/beelinepositive/skewjoin.q.out ql/src/test/results/beelinepositive/skewjoin.q.out index 222518a..ee3c8b6 100644 --- ql/src/test/results/beelinepositive/skewjoin.q.out +++ ql/src/test/results/beelinepositive/skewjoin.q.out @@ -92,26 +92,19 @@ INSERT OVERWRITE TABLE dest_j1 SELECT src1.key, src2.value; ' outputColumnNames: _col0, _col5' ' Select Operator' ' expressions:' -' expr: _col0' -' type: string' +' expr: UDFToInteger(_col0)' +' type: int' ' expr: _col5' ' type: string' ' outputColumnNames: _col0, _col1' -' Select Operator' -' expressions:' -' expr: UDFToInteger(_col0)' -' type: int' -' expr: _col1' -' type: string' -' outputColumnNames: _col0, _col1' -' File Output Operator' -' compressed: false' -' GlobalTableId: 1' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: skewjoin.dest_j1' +' File Output Operator' +' compressed: false' +' GlobalTableId: 1' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' name: skewjoin.dest_j1' '' ' Stage: Stage-5' ' Conditional Operator' @@ -152,26 +145,19 @@ INSERT OVERWRITE TABLE dest_j1 SELECT src1.key, src2.value; ' Position of Big Table: 0' ' Select Operator' ' expressions:' -' expr: _col0' -' type: string' +' expr: UDFToInteger(_col0)' +' type: int' ' expr: _col5' ' type: string' ' outputColumnNames: _col0, _col1' -' Select Operator' -' expressions:' -' expr: UDFToInteger(_col0)' -' type: int' -' expr: _col1' -' type: string' -' outputColumnNames: _col0, _col1' -' File Output Operator' -' compressed: false' -' GlobalTableId: 1' -' table:' -' input format: org.apache.hadoop.mapred.TextInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' -' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' -' name: skewjoin.dest_j1' +' File Output Operator' +' compressed: false' +' GlobalTableId: 1' +' table:' +' input format: org.apache.hadoop.mapred.TextInputFormat' +' output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' +' serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe' +' name: skewjoin.dest_j1' ' Local Work:' ' Map Reduce Local Work' '' @@ -189,7 +175,7 @@ INSERT OVERWRITE TABLE dest_j1 SELECT src1.key, src2.value; ' Stats-Aggr Operator' '' '' -153 rows selected +139 rows selected >>> >>> FROM src src1 JOIN src src2 ON (src1.key = src2.key) INSERT OVERWRITE TABLE dest_j1 SELECT src1.key, src2.value; @@ -1063,7 +1049,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80; ' alias: src' ' Filter Operator' ' predicate:' -' expr: ((key < 100.0) and (key < 80.0))' +' expr: ((key < 80.0) and (key < 100.0))' ' type: boolean' ' Select Operator' ' expressions:' @@ -1113,7 +1099,7 @@ ON src1.c1 = src3.c5 AND src3.c5 < 80; ' alias: src' ' Filter Operator' ' predicate:' -' expr: ((key < 80.0) and (key < 100.0))' +' expr: ((key < 100.0) and (key < 80.0))' ' type: boolean' ' Select Operator' ' expressions:' @@ -1379,13 +1365,12 @@ SELECT /*+ mapjoin(v)*/ sum(hash(k.key)), sum(hash(v.val)) FROM T1 k LEFT OUTER ' (TOK_QUERY (TOK_FROM (TOK_LEFTOUTERJOIN (TOK_TABREF (TOK_TABNAME T1) k) (TOK_TABREF (TOK_TABNAME T1) v) (= (+ (. (TOK_TABLE_OR_COL k) key) 1) (. (TOK_TABLE_OR_COL v) key)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_HINTLIST (TOK_HINT TOK_MAPJOIN (TOK_HINTARGLIST v))) (TOK_SELEXPR (TOK_FUNCTION sum (TOK_FUNCTION hash (. (TOK_TABLE_OR_COL k) key)))) (TOK_SELEXPR (TOK_FUNCTION sum (TOK_FUNCTION hash (. (TOK_TABLE_OR_COL v) val)))))))' '' 'STAGE DEPENDENCIES:' -' Stage-4 is a root stage' -' Stage-1 depends on stages: Stage-4' -' Stage-2 depends on stages: Stage-1' +' Stage-3 is a root stage' +' Stage-1 depends on stages: Stage-3' ' Stage-0 is a root stage' '' 'STAGE PLANS:' -' Stage: Stage-4' +' Stage: Stage-3' ' Map Reduce Local Work' ' Alias -> Map Local Tables:' ' v ' @@ -1423,48 +1408,30 @@ SELECT /*+ mapjoin(v)*/ sum(hash(k.key)), sum(hash(v.val)) FROM T1 k LEFT OUTER ' 1 [class org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge(Column[key]()]' ' outputColumnNames: _col0, _col5' ' Position of Big Table: 0' -' File Output Operator' -' compressed: false' -' GlobalTableId: 0' -' table:' -' input format: org.apache.hadoop.mapred.SequenceFileInputFormat' -' output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' +' Select Operator' +' expressions:' +' expr: _col0' +' type: string' +' expr: _col5' +' type: string' +' outputColumnNames: _col0, _col5' +' Group By Operator' +' aggregations:' +' expr: sum(hash(_col0))' +' expr: sum(hash(_col5))' +' bucketGroup: false' +' mode: hash' +' outputColumnNames: _col0, _col1' +' Reduce Output Operator' +' sort order: ' +' tag: -1' +' value expressions:' +' expr: _col0' +' type: bigint' +' expr: _col1' +' type: bigint' ' Local Work:' ' Map Reduce Local Work' -'' -' Stage: Stage-2' -' Map Reduce' -' Alias -> Map Operator Tree:' -' file:!!{hive.exec.local.scratchdir}!! ' -' Select Operator' -' expressions:' -' expr: _col0' -' type: string' -' expr: _col5' -' type: string' -' outputColumnNames: _col0, _col5' -' Select Operator' -' expressions:' -' expr: _col0' -' type: string' -' expr: _col5' -' type: string' -' outputColumnNames: _col0, _col5' -' Group By Operator' -' aggregations:' -' expr: sum(hash(_col0))' -' expr: sum(hash(_col5))' -' bucketGroup: false' -' mode: hash' -' outputColumnNames: _col0, _col1' -' Reduce Output Operator' -' sort order: ' -' tag: -1' -' value expressions:' -' expr: _col0' -' type: bigint' -' expr: _col1' -' type: bigint' ' Reduce Operator Tree:' ' Group By Operator' ' aggregations:' @@ -1492,7 +1459,7 @@ SELECT /*+ mapjoin(v)*/ sum(hash(k.key)), sum(hash(v.val)) FROM T1 k LEFT OUTER ' limit: -1' '' '' -117 rows selected +98 rows selected >>> SELECT /*+ mapjoin(v)*/ sum(hash(k.key)), sum(hash(v.val)) FROM T1 k LEFT OUTER JOIN T1 v ON k.key+1=v.key; '_c1','_c2' '372','6320' diff --git ql/src/test/results/beelinepositive/timestamp_3.q.out ql/src/test/results/beelinepositive/timestamp_3.q.out index d2ec1c6..4ce6e95 100644 --- ql/src/test/results/beelinepositive/timestamp_3.q.out +++ ql/src/test/results/beelinepositive/timestamp_3.q.out @@ -13,35 +13,35 @@ select cast(cast('1.3041352164485E9' as double) as timestamp) from src limit 1; '_c0' No rows selected >>> select cast(t as boolean) from timestamp_3 limit 1; -'_c0' +'t' 'true' 1 row selected >>> select cast(t as tinyint) from timestamp_3 limit 1; -'_c0' +'t' '48' 1 row selected >>> select cast(t as smallint) from timestamp_3 limit 1; -'_c0' +'t' '-31184' 1 row selected >>> select cast(t as int) from timestamp_3 limit 1; -'_c0' +'t' '1304135216' 1 row selected >>> select cast(t as bigint) from timestamp_3 limit 1; -'_c0' +'t' '1304135216' 1 row selected >>> select cast(t as float) from timestamp_3 limit 1; -'_c0' +'t' '1.304135168E9' 1 row selected >>> select cast(t as double) from timestamp_3 limit 1; -'_c0' +'t' '1.3041352164485E9' 1 row selected >>> select cast(t as string) from timestamp_3 limit 1; -'_c0' +'t' '2011-04-29 20:46:56.4485' 1 row selected >>> diff --git ql/src/test/results/beelinepositive/type_widening.q.out ql/src/test/results/beelinepositive/type_widening.q.out index 0b50828..a81c0ec 100644 --- ql/src/test/results/beelinepositive/type_widening.q.out +++ ql/src/test/results/beelinepositive/type_widening.q.out @@ -59,29 +59,24 @@ Saving all output to "!!{outputDirectory}!!/type_widening.q.raw". Enter "record" ' alias: src' ' Select Operator' ' expressions:' -' expr: 0' -' type: int' +' expr: UDFToLong(0)' +' type: bigint' ' outputColumnNames: _col0' -' Select Operator' -' expressions:' -' expr: UDFToLong(_col0)' -' type: bigint' -' outputColumnNames: _col0' -' Union' -' Select Operator' -' expressions:' +' Union' +' Select Operator' +' expressions:' +' expr: _col0' +' type: bigint' +' outputColumnNames: _col0' +' Reduce Output Operator' +' key expressions:' +' expr: _col0' +' type: bigint' +' sort order: +' +' tag: -1' +' value expressions:' ' expr: _col0' ' type: bigint' -' outputColumnNames: _col0' -' Reduce Output Operator' -' key expressions:' -' expr: _col0' -' type: bigint' -' sort order: +' -' tag: -1' -' value expressions:' -' expr: _col0' -' type: bigint' ' null-subquery2:a-subquery2:src ' ' TableScan' ' alias: src' @@ -119,7 +114,7 @@ Saving all output to "!!{outputDirectory}!!/type_widening.q.raw". Enter "record" ' limit: -1' '' '' -76 rows selected +71 rows selected >>> SELECT * FROM (SELECT 0 AS numcol FROM src UNION ALL SELECT 9223372036854775807 AS numcol FROM src) a ORDER BY numcol; 'numcol' '0' diff --git ql/src/test/results/clientnegative/desc_failure2.q.out ql/src/test/results/clientnegative/desc_failure2.q.out index 333a26a..f0f6668 100644 --- ql/src/test/results/clientnegative/desc_failure2.q.out +++ ql/src/test/results/clientnegative/desc_failure2.q.out @@ -2,16 +2,8 @@ PREHOOK: query: DESC srcpart PREHOOK: type: DESCTABLE POSTHOOK: query: DESC srcpart POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string FAILED: SemanticException [Error 10006]: Partition not found {ds=2012-04-08, hr=15} diff --git ql/src/test/results/clientnegative/drop_partition_filter_failure2.q.out ql/src/test/results/clientnegative/drop_partition_filter_failure2.q.out index ea8f147..405015c 100644 --- ql/src/test/results/clientnegative/drop_partition_filter_failure2.q.out +++ ql/src/test/results/clientnegative/drop_partition_filter_failure2.q.out @@ -7,18 +7,10 @@ PREHOOK: query: describe ptestfilter PREHOOK: type: DESCTABLE POSTHOOK: query: describe ptestfilter POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b int None -c string None -d int None - -# Partition Information -# col_name data_type comment - -c string None -d int None +a string +b int +c string +d int PREHOOK: query: alter table ptestfilter add partition (c='US', d=1) PREHOOK: type: ALTERTABLE_ADDPARTS PREHOOK: Input: default@ptestfilter diff --git ql/src/test/results/clientnegative/protectmode_part_no_drop.q.out ql/src/test/results/clientnegative/protectmode_part_no_drop.q.out index 035fb36..f15cdbf 100644 --- ql/src/test/results/clientnegative/protectmode_part_no_drop.q.out +++ ql/src/test/results/clientnegative/protectmode_part_no_drop.q.out @@ -31,16 +31,9 @@ PREHOOK: query: desc extended tbl_protectmode_no_drop partition (p='p1') PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl_protectmode_no_drop partition (p='p1') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -c1 string None -c2 string None -p string None - -# Partition Information -# col_name data_type comment - -p string None +c1 string +c2 string +p string #### A masked pattern was here #### FAILED: SemanticException [Error 30011]: Partition protected from being dropped default@tbl_protectmode_no_drop@p=p1 diff --git ql/src/test/results/clientnegative/protectmode_tbl2.q.out ql/src/test/results/clientnegative/protectmode_tbl2.q.out index 480ca5c..844f055 100644 --- ql/src/test/results/clientnegative/protectmode_tbl2.q.out +++ ql/src/test/results/clientnegative/protectmode_tbl2.q.out @@ -46,15 +46,8 @@ PREHOOK: query: desc extended tbl_protectmode2 PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl_protectmode2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col string None -p string None - -# Partition Information -# col_name data_type comment - -p string None +col string +p string #### A masked pattern was here #### FAILED: SemanticException [Error 10113]: Query against an offline table or partition Table tbl_protectmode2 diff --git ql/src/test/results/clientnegative/protectmode_tbl3.q.out ql/src/test/results/clientnegative/protectmode_tbl3.q.out index 6dc70ef..c625cc0 100644 --- ql/src/test/results/clientnegative/protectmode_tbl3.q.out +++ ql/src/test/results/clientnegative/protectmode_tbl3.q.out @@ -31,9 +31,7 @@ PREHOOK: query: desc extended tbl_protectmode_4 PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl_protectmode_4 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col string None +col string #### A masked pattern was here #### FAILED: SemanticException [Error 10113]: Query against an offline table or partition Table tbl_protectmode_4 diff --git ql/src/test/results/clientnegative/protectmode_tbl4.q.out ql/src/test/results/clientnegative/protectmode_tbl4.q.out index 7a5058f..3c8902d 100644 --- ql/src/test/results/clientnegative/protectmode_tbl4.q.out +++ ql/src/test/results/clientnegative/protectmode_tbl4.q.out @@ -55,15 +55,8 @@ PREHOOK: query: desc extended tbl_protectmode_tbl4 PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl_protectmode_tbl4 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col string None -p string None - -# Partition Information -# col_name data_type comment - -p string None +col string +p string #### A masked pattern was here #### FAILED: SemanticException [Error 10113]: Query against an offline table or partition Table tbl_protectmode_tbl4 diff --git ql/src/test/results/clientnegative/protectmode_tbl5.q.out ql/src/test/results/clientnegative/protectmode_tbl5.q.out index 4bb7b87..9825f84 100644 --- ql/src/test/results/clientnegative/protectmode_tbl5.q.out +++ ql/src/test/results/clientnegative/protectmode_tbl5.q.out @@ -55,15 +55,8 @@ PREHOOK: query: desc extended tbl_protectmode_tbl5 PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl_protectmode_tbl5 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col string None -p string None - -# Partition Information -# col_name data_type comment - -p string None +col string +p string #### A masked pattern was here #### FAILED: SemanticException [Error 10113]: Query against an offline table or partition Table tbl_protectmode_tbl5 diff --git ql/src/test/results/clientnegative/protectmode_tbl_no_drop.q.out ql/src/test/results/clientnegative/protectmode_tbl_no_drop.q.out index 7ef9334..2bcc424 100644 --- ql/src/test/results/clientnegative/protectmode_tbl_no_drop.q.out +++ ql/src/test/results/clientnegative/protectmode_tbl_no_drop.q.out @@ -31,9 +31,7 @@ PREHOOK: query: desc extended tbl_protectmode__no_drop PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl_protectmode__no_drop POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col string None +col string #### A masked pattern was here #### PREHOOK: query: drop table tbl_protectmode__no_drop diff --git ql/src/test/results/clientnegative/set_hiveconf_validation0.q.out ql/src/test/results/clientnegative/set_hiveconf_validation0.q.out index 2cd2f58..9ecff1d 100644 --- ql/src/test/results/clientnegative/set_hiveconf_validation0.q.out +++ ql/src/test/results/clientnegative/set_hiveconf_validation0.q.out @@ -4,8 +4,6 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: -- should fail: hive.join.cache.size accepts int type desc src POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default +key string default +value string default Query returned non-zero code: 1, cause: 'SET hive.join.cache.size=test' FAILED because hive.join.cache.size expects an INT value. diff --git ql/src/test/results/clientnegative/set_hiveconf_validation1.q.out ql/src/test/results/clientnegative/set_hiveconf_validation1.q.out index 054cbe8..893d012 100644 --- ql/src/test/results/clientnegative/set_hiveconf_validation1.q.out +++ ql/src/test/results/clientnegative/set_hiveconf_validation1.q.out @@ -4,8 +4,6 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: -- should fail: hive.map.aggr.hash.min.reduction accepts float type desc src POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default +key string default +value string default Query returned non-zero code: 1, cause: 'SET hive.map.aggr.hash.min.reduction=false' FAILED because hive.map.aggr.hash.min.reduction expects an FLOAT value. diff --git ql/src/test/results/clientpositive/alter1.q.out ql/src/test/results/clientpositive/alter1.q.out index 4e5bbdc..8c8f77a 100644 --- ql/src/test/results/clientpositive/alter1.q.out +++ ql/src/test/results/clientpositive/alter1.q.out @@ -7,10 +7,8 @@ PREHOOK: query: describe extended alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None +a int +b int #### A masked pattern was here #### PREHOOK: query: alter table alter1 set tblproperties ('a'='1', 'c'='3') @@ -25,10 +23,8 @@ PREHOOK: query: describe extended alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None +a int +b int #### A masked pattern was here #### PREHOOK: query: alter table alter1 set tblproperties ('a'='1', 'c'='4', 'd'='3') @@ -43,10 +39,8 @@ PREHOOK: query: describe extended alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None +a int +b int #### A masked pattern was here #### PREHOOK: query: alter table alter1 set tblproperties ('EXTERNAL'='TRUE') @@ -61,10 +55,8 @@ PREHOOK: query: describe extended alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None +a int +b int #### A masked pattern was here #### PREHOOK: query: alter table alter1 set tblproperties ('EXTERNAL'='FALSE') @@ -79,10 +71,8 @@ PREHOOK: query: describe extended alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None +a int +b int #### A masked pattern was here #### PREHOOK: query: alter table alter1 set serdeproperties('s1'='9') @@ -97,10 +87,8 @@ PREHOOK: query: describe extended alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None +a int +b int #### A masked pattern was here #### PREHOOK: query: alter table alter1 set serdeproperties('s1'='10', 's2' ='20') @@ -115,10 +103,8 @@ PREHOOK: query: describe extended alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None +a int +b int #### A masked pattern was here #### PREHOOK: query: alter table alter1 set serde 'org.apache.hadoop.hive.serde2.TestSerDe' with serdeproperties('s1'='9') @@ -133,10 +119,8 @@ PREHOOK: query: describe extended alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string from deserializer -b string from deserializer +a string from deserializer +b string from deserializer #### A masked pattern was here #### PREHOOK: query: alter table alter1 set serde 'org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe' @@ -151,10 +135,8 @@ PREHOOK: query: describe extended alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string from deserializer -b string from deserializer +a string from deserializer +b string from deserializer #### A masked pattern was here #### PREHOOK: query: alter table alter1 replace columns (a int, b int, c string) @@ -169,11 +151,9 @@ PREHOOK: query: describe alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: describe alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None -c string None +a int +b int +c string PREHOOK: query: -- Cleanup DROP TABLE alter1 PREHOOK: type: DROPTABLE @@ -221,10 +201,8 @@ PREHOOK: query: DESCRIBE EXTENDED alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None +a int +b int #### A masked pattern was here #### PREHOOK: query: ALTER TABLE alter1 SET TBLPROPERTIES ('a'='1', 'c'='3') @@ -239,10 +217,8 @@ PREHOOK: query: DESCRIBE EXTENDED alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None +a int +b int #### A masked pattern was here #### PREHOOK: query: ALTER TABLE alter1 SET TBLPROPERTIES ('a'='1', 'c'='4', 'd'='3') @@ -257,10 +233,8 @@ PREHOOK: query: DESCRIBE EXTENDED alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None +a int +b int #### A masked pattern was here #### PREHOOK: query: ALTER TABLE alter1 SET TBLPROPERTIES ('EXTERNAL'='TRUE') @@ -275,10 +249,8 @@ PREHOOK: query: DESCRIBE EXTENDED alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None +a int +b int #### A masked pattern was here #### PREHOOK: query: ALTER TABLE alter1 SET TBLPROPERTIES ('EXTERNAL'='FALSE') @@ -293,10 +265,8 @@ PREHOOK: query: DESCRIBE EXTENDED alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None +a int +b int #### A masked pattern was here #### PREHOOK: query: ALTER TABLE alter1 SET SERDEPROPERTIES('s1'='9') @@ -311,10 +281,8 @@ PREHOOK: query: DESCRIBE EXTENDED alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None +a int +b int #### A masked pattern was here #### PREHOOK: query: ALTER TABLE alter1 SET SERDEPROPERTIES('s1'='10', 's2' ='20') @@ -329,10 +297,8 @@ PREHOOK: query: DESCRIBE EXTENDED alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None +a int +b int #### A masked pattern was here #### PREHOOK: query: ALTER TABLE alter1 SET SERDE 'org.apache.hadoop.hive.serde2.TestSerDe' WITH SERDEPROPERTIES ('s1'='9') @@ -347,10 +313,8 @@ PREHOOK: query: DESCRIBE EXTENDED alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string from deserializer -b string from deserializer +a string from deserializer +b string from deserializer #### A masked pattern was here #### PREHOOK: query: ALTER TABLE alter1 SET SERDE 'org.apache.hadoop.hive.serde2.MetadataTypedColumnsetSerDe' @@ -365,10 +329,8 @@ PREHOOK: query: DESCRIBE EXTENDED alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string from deserializer -b string from deserializer +a string from deserializer +b string from deserializer #### A masked pattern was here #### PREHOOK: query: ALTER TABLE alter1 REPLACE COLUMNS (a int, b int, c string) @@ -383,11 +345,9 @@ PREHOOK: query: DESCRIBE alter1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE alter1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None -c string None +a int +b int +c string PREHOOK: query: DROP TABLE alter1 PREHOOK: type: DROPTABLE PREHOOK: Input: alter1_db@alter1 diff --git ql/src/test/results/clientpositive/alter2.q.out ql/src/test/results/clientpositive/alter2.q.out index 9923ab8..fd57426 100644 --- ql/src/test/results/clientpositive/alter2.q.out +++ ql/src/test/results/clientpositive/alter2.q.out @@ -7,16 +7,9 @@ PREHOOK: query: describe extended alter2 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None -insertdate string None - -# Partition Information -# col_name data_type comment - -insertdate string None +a int +b int +insertdate string #### A masked pattern was here #### PREHOOK: query: show partitions alter2 @@ -34,16 +27,9 @@ PREHOOK: query: describe extended alter2 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None -insertdate string None - -# Partition Information -# col_name data_type comment - -insertdate string None +a int +b int +insertdate string #### A masked pattern was here #### PREHOOK: query: show partitions alter2 @@ -62,16 +48,9 @@ PREHOOK: query: describe extended alter2 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None -insertdate string None - -# Partition Information -# col_name data_type comment - -insertdate string None +a int +b int +insertdate string #### A masked pattern was here #### PREHOOK: query: show partitions alter2 @@ -97,16 +76,9 @@ PREHOOK: query: describe extended alter2 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None -insertdate string None - -# Partition Information -# col_name data_type comment - -insertdate string None +a int +b int +insertdate string #### A masked pattern was here #### PREHOOK: query: show partitions alter2 @@ -124,16 +96,9 @@ PREHOOK: query: describe extended alter2 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None -insertdate string None - -# Partition Information -# col_name data_type comment - -insertdate string None +a int +b int +insertdate string #### A masked pattern was here #### PREHOOK: query: show partitions alter2 @@ -152,16 +117,9 @@ PREHOOK: query: describe extended alter2 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None -insertdate string None - -# Partition Information -# col_name data_type comment - -insertdate string None +a int +b int +insertdate string #### A masked pattern was here #### PREHOOK: query: show partitions alter2 @@ -217,16 +175,9 @@ PREHOOK: query: DESCRIBE EXTENDED alter2 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED alter2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None -insertdate string None - -# Partition Information -# col_name data_type comment - -insertdate string None +a int +b int +insertdate string #### A masked pattern was here #### PREHOOK: query: SHOW PARTITIONS alter2 @@ -244,16 +195,9 @@ PREHOOK: query: DESCRIBE EXTENDED alter2 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED alter2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None -insertdate string None - -# Partition Information -# col_name data_type comment - -insertdate string None +a int +b int +insertdate string #### A masked pattern was here #### PREHOOK: query: SHOW PARTITIONS alter2 @@ -272,16 +216,9 @@ PREHOOK: query: DESCRIBE EXTENDED alter2 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED alter2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None -insertdate string None - -# Partition Information -# col_name data_type comment - -insertdate string None +a int +b int +insertdate string #### A masked pattern was here #### PREHOOK: query: SHOW PARTITIONS alter2 @@ -307,16 +244,9 @@ PREHOOK: query: DESCRIBE EXTENDED alter2 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED alter2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None -insertdate string None - -# Partition Information -# col_name data_type comment - -insertdate string None +a int +b int +insertdate string #### A masked pattern was here #### PREHOOK: query: SHOW PARTITIONS alter2 @@ -334,16 +264,9 @@ PREHOOK: query: DESCRIBE EXTENDED alter2 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED alter2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None -insertdate string None - -# Partition Information -# col_name data_type comment - -insertdate string None +a int +b int +insertdate string #### A masked pattern was here #### PREHOOK: query: SHOW PARTITIONS alter2 @@ -362,16 +285,9 @@ PREHOOK: query: DESCRIBE EXTENDED alter2 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED alter2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None -insertdate string None - -# Partition Information -# col_name data_type comment - -insertdate string None +a int +b int +insertdate string #### A masked pattern was here #### PREHOOK: query: SHOW PARTITIONS alter2 diff --git ql/src/test/results/clientpositive/alter3.q.out ql/src/test/results/clientpositive/alter3.q.out index c8716ba..5f3b0b4 100644 --- ql/src/test/results/clientpositive/alter3.q.out +++ ql/src/test/results/clientpositive/alter3.q.out @@ -58,17 +58,9 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter3_renamed POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: alter3 PARTITION(pcol1=test_part:,pcol2=test_part:).col1 SIMPLE [(alter3_src)alter3_src.FieldSchema(name:col1, type:string, comment:null), ] -# col_name data_type comment - -col1 string None -pcol1 string None -pcol2 string None - -# Partition Information -# col_name data_type comment - -pcol1 string None -pcol2 string None +col1 string +pcol1 string +pcol2 string #### A masked pattern was here #### PREHOOK: query: describe extended alter3_renamed partition (pCol1='test_part:', pcol2='test_part:') @@ -76,17 +68,9 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter3_renamed partition (pCol1='test_part:', pcol2='test_part:') POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: alter3 PARTITION(pcol1=test_part:,pcol2=test_part:).col1 SIMPLE [(alter3_src)alter3_src.FieldSchema(name:col1, type:string, comment:null), ] -# col_name data_type comment - -col1 string None -pcol1 string None -pcol2 string None - -# Partition Information -# col_name data_type comment - -pcol1 string None -pcol2 string None +col1 string +pcol1 string +pcol2 string #### A masked pattern was here #### PREHOOK: query: select * from alter3_renamed where pcol1='test_part:' and pcol2='test_part:' @@ -135,17 +119,9 @@ POSTHOOK: query: describe extended alter3_like_renamed POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: alter3 PARTITION(pcol1=test_part:,pcol2=test_part:).col1 SIMPLE [(alter3_src)alter3_src.FieldSchema(name:col1, type:string, comment:null), ] POSTHOOK: Lineage: alter3_like PARTITION(pcol1=test_part:,pcol2=test_part:).col1 SIMPLE [(alter3_src)alter3_src.FieldSchema(name:col1, type:string, comment:null), ] -# col_name data_type comment - -col1 string None -pcol1 string None -pcol2 string None - -# Partition Information -# col_name data_type comment - -pcol1 string None -pcol2 string None +col1 string +pcol1 string +pcol2 string #### A masked pattern was here #### PREHOOK: query: -- Cleanup @@ -292,17 +268,9 @@ POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: alter3 PARTITION(pcol1=test_part:,pcol2=test_part:).col1 SIMPLE [(alter3_src)alter3_src.FieldSchema(name:col1, type:string, comment:null), ] POSTHOOK: Lineage: alter3 PARTITION(pcol1=test_part:,pcol2=test_part:).col1 SIMPLE [(alter3_src)alter3_src.FieldSchema(name:col1, type:string, comment:null), ] POSTHOOK: Lineage: alter3_like PARTITION(pcol1=test_part:,pcol2=test_part:).col1 SIMPLE [(alter3_src)alter3_src.FieldSchema(name:col1, type:string, comment:null), ] -# col_name data_type comment - -col1 string None -pcol1 string None -pcol2 string None - -# Partition Information -# col_name data_type comment - -pcol1 string None -pcol2 string None +col1 string +pcol1 string +pcol2 string #### A masked pattern was here #### PREHOOK: query: DESCRIBE EXTENDED alter3_renamed PARTITION (pCol1='test_part:', pcol2='test_part:') @@ -312,17 +280,9 @@ POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: alter3 PARTITION(pcol1=test_part:,pcol2=test_part:).col1 SIMPLE [(alter3_src)alter3_src.FieldSchema(name:col1, type:string, comment:null), ] POSTHOOK: Lineage: alter3 PARTITION(pcol1=test_part:,pcol2=test_part:).col1 SIMPLE [(alter3_src)alter3_src.FieldSchema(name:col1, type:string, comment:null), ] POSTHOOK: Lineage: alter3_like PARTITION(pcol1=test_part:,pcol2=test_part:).col1 SIMPLE [(alter3_src)alter3_src.FieldSchema(name:col1, type:string, comment:null), ] -# col_name data_type comment - -col1 string None -pcol1 string None -pcol2 string None - -# Partition Information -# col_name data_type comment - -pcol1 string None -pcol2 string None +col1 string +pcol1 string +pcol2 string #### A masked pattern was here #### PREHOOK: query: SELECT * FROM alter3_renamed WHERE pcol1='test_part:' AND pcol2='test_part:' @@ -379,16 +339,8 @@ POSTHOOK: Lineage: alter3 PARTITION(pcol1=test_part:,pcol2=test_part:).col1 SIMP POSTHOOK: Lineage: alter3 PARTITION(pcol1=test_part:,pcol2=test_part:).col1 SIMPLE [(alter3_src)alter3_src.FieldSchema(name:col1, type:string, comment:null), ] POSTHOOK: Lineage: alter3_like PARTITION(pcol1=test_part:,pcol2=test_part:).col1 SIMPLE [(alter3_src)alter3_src.FieldSchema(name:col1, type:string, comment:null), ] POSTHOOK: Lineage: alter3_like PARTITION(pcol1=test_part:,pcol2=test_part:).col1 SIMPLE [(alter3_src)alter3_src.FieldSchema(name:col1, type:string, comment:null), ] -# col_name data_type comment - -col1 string None -pcol1 string None -pcol2 string None - -# Partition Information -# col_name data_type comment - -pcol1 string None -pcol2 string None +col1 string +pcol1 string +pcol2 string #### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/alter4.q.out ql/src/test/results/clientpositive/alter4.q.out index d618fa6..e471131 100644 --- ql/src/test/results/clientpositive/alter4.q.out +++ ql/src/test/results/clientpositive/alter4.q.out @@ -7,10 +7,8 @@ PREHOOK: query: DESCRIBE EXTENDED set_bucketing_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED set_bucketing_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None +key int +value string #### A masked pattern was here #### PREHOOK: query: ALTER TABLE set_bucketing_test NOT CLUSTERED @@ -25,10 +23,8 @@ PREHOOK: query: DESCRIBE EXTENDED set_bucketing_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED set_bucketing_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None +key int +value string #### A masked pattern was here #### PREHOOK: query: -- Cleanup @@ -78,10 +74,8 @@ PREHOOK: query: DESCRIBE EXTENDED set_bucketing_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED set_bucketing_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None +key int +value string #### A masked pattern was here #### PREHOOK: query: ALTER TABLE set_bucketing_test NOT CLUSTERED @@ -96,10 +90,8 @@ PREHOOK: query: DESCRIBE EXTENDED set_bucketing_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED set_bucketing_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None +key int +value string #### A masked pattern was here #### PREHOOK: query: DROP TABLE set_bucketing_test diff --git ql/src/test/results/clientpositive/alter5.q.out ql/src/test/results/clientpositive/alter5.q.out index 1db890b..e60e7ca 100644 --- ql/src/test/results/clientpositive/alter5.q.out +++ ql/src/test/results/clientpositive/alter5.q.out @@ -41,15 +41,8 @@ PREHOOK: query: describe extended alter5 partition (dt='a') PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter5 partition (dt='a') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col1 string None -dt string None - -# Partition Information -# col_name data_type comment - -dt string None +col1 string +dt string #### A masked pattern was here #### PREHOOK: query: insert overwrite table alter5 partition (dt='a') select col1 from alter5_src @@ -81,15 +74,8 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter5 partition (dt='a') POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: alter5 PARTITION(dt=a).col1 SIMPLE [(alter5_src)alter5_src.FieldSchema(name:col1, type:string, comment:null), ] -# col_name data_type comment - -col1 string None -dt string None - -# Partition Information -# col_name data_type comment - -dt string None +col1 string +dt string #### A masked pattern was here #### PREHOOK: query: -- Cleanup @@ -176,15 +162,8 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended alter5 partition (dt='a') POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: alter5 PARTITION(dt=a).col1 SIMPLE [(alter5_src)alter5_src.FieldSchema(name:col1, type:string, comment:null), ] -# col_name data_type comment - -col1 string None -dt string None - -# Partition Information -# col_name data_type comment - -dt string None +col1 string +dt string #### A masked pattern was here #### PREHOOK: query: insert overwrite table alter5 partition (dt='a') select col1 from alter5_src @@ -219,14 +198,7 @@ POSTHOOK: query: describe extended alter5 partition (dt='a') POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: alter5 PARTITION(dt=a).col1 SIMPLE [(alter5_src)alter5_src.FieldSchema(name:col1, type:string, comment:null), ] POSTHOOK: Lineage: alter5 PARTITION(dt=a).col1 SIMPLE [(alter5_src)alter5_src.FieldSchema(name:col1, type:string, comment:null), ] -# col_name data_type comment - -col1 string None -dt string None - -# Partition Information -# col_name data_type comment - -dt string None +col1 string +dt string #### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/alter_index.q.out ql/src/test/results/clientpositive/alter_index.q.out index d34bd67..5218c8f 100644 --- ql/src/test/results/clientpositive/alter_index.q.out +++ ql/src/test/results/clientpositive/alter_index.q.out @@ -11,11 +11,9 @@ PREHOOK: query: desc extended default__src_src_index_8__ PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended default__src_src_index_8__ POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -_bucketname string -_offsets array +key string default +_bucketname string +_offsets array #### A masked pattern was here #### PREHOOK: query: alter index src_index_8 on src set IDXPROPERTIES ("prop1"="val1_new", "prop3"="val3") @@ -26,11 +24,9 @@ PREHOOK: query: desc extended default__src_src_index_8__ PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended default__src_src_index_8__ POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -_bucketname string -_offsets array +key string default +_bucketname string +_offsets array #### A masked pattern was here #### PREHOOK: query: drop index src_index_8 on src diff --git ql/src/test/results/clientpositive/alter_merge_2.q.out ql/src/test/results/clientpositive/alter_merge_2.q.out index df892a3..d8a527e 100644 --- ql/src/test/results/clientpositive/alter_merge_2.q.out +++ ql/src/test/results/clientpositive/alter_merge_2.q.out @@ -14,18 +14,10 @@ PREHOOK: query: desc extended src_rc_merge_test_part partition (ds='2012-01-03', PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended src_rc_merge_test_part partition (ds='2012-01-03', ts='2012-01-03+14:46:31') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None -ds string None -ts string None - -# Partition Information -# col_name data_type comment - -ds string None -ts string None +key int +value string +ds string +ts string #### A masked pattern was here #### PREHOOK: query: load data local inpath '../data/files/smbbucket_1.rc' into table src_rc_merge_test_part partition (ds='2012-01-03', ts='2012-01-03+14:46:31') diff --git ql/src/test/results/clientpositive/alter_merge_stats.q.out ql/src/test/results/clientpositive/alter_merge_stats.q.out index 19e8942..bbfac8d 100644 --- ql/src/test/results/clientpositive/alter_merge_stats.q.out +++ ql/src/test/results/clientpositive/alter_merge_stats.q.out @@ -42,10 +42,8 @@ PREHOOK: query: desc extended src_rc_merge_test_stat PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended src_rc_merge_test_stat POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None +key int +value string #### A masked pattern was here #### PREHOOK: query: analyze table src_rc_merge_test_stat compute statistics @@ -60,10 +58,8 @@ PREHOOK: query: desc extended src_rc_merge_test_stat PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended src_rc_merge_test_stat POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None +key int +value string #### A masked pattern was here #### PREHOOK: query: alter table src_rc_merge_test_stat concatenate @@ -95,10 +91,8 @@ PREHOOK: query: desc extended src_rc_merge_test_stat PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended src_rc_merge_test_stat POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None +key int +value string #### A masked pattern was here #### PREHOOK: query: create table src_rc_merge_test_part_stat(key int, value string) partitioned by (ds string) stored as rcfile @@ -152,16 +146,9 @@ PREHOOK: query: desc extended src_rc_merge_test_part_stat PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended src_rc_merge_test_part_stat POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key int +value string +ds string #### A masked pattern was here #### PREHOOK: query: analyze table src_rc_merge_test_part_stat partition(ds='2011') compute statistics @@ -180,16 +167,9 @@ PREHOOK: query: desc extended src_rc_merge_test_part_stat PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended src_rc_merge_test_part_stat POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key int +value string +ds string #### A masked pattern was here #### PREHOOK: query: alter table src_rc_merge_test_part_stat partition (ds='2011') concatenate @@ -221,16 +201,9 @@ PREHOOK: query: desc extended src_rc_merge_test_part_stat PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended src_rc_merge_test_part_stat POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key int +value string +ds string #### A masked pattern was here #### PREHOOK: query: drop table src_rc_merge_test_stat diff --git ql/src/test/results/clientpositive/alter_partition_coltype.q.out ql/src/test/results/clientpositive/alter_partition_coltype.q.out index 2a8d185..59a83b9 100644 --- ql/src/test/results/clientpositive/alter_partition_coltype.q.out +++ ql/src/test/results/clientpositive/alter_partition_coltype.q.out @@ -23,18 +23,10 @@ POSTHOOK: query: desc alter_coltype POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: alter_coltype PARTITION(dt=100x,ts=6:30pm).key SIMPLE [(src1)src1.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: alter_coltype PARTITION(dt=100x,ts=6:30pm).value SIMPLE [(src1)src1.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None -dt string None -ts string None - -# Partition Information -# col_name data_type comment - -dt string None -ts string None +key string +value string +dt string +ts string PREHOOK: query: -- select with paritition predicate. select count(*) from alter_coltype where dt = '100x' PREHOOK: type: QUERY @@ -1227,18 +1219,10 @@ POSTHOOK: Lineage: alter_coltype PARTITION(dt=100x,ts=3.0).key SIMPLE [(src1)src POSTHOOK: Lineage: alter_coltype PARTITION(dt=100x,ts=3.0).value SIMPLE [(src1)src1.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: alter_coltype PARTITION(dt=100x,ts=6:30pm).key SIMPLE [(src1)src1.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: alter_coltype PARTITION(dt=100x,ts=6:30pm).value SIMPLE [(src1)src1.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None -dt int None -ts double None - -# Partition Information -# col_name data_type comment - -dt int None -ts double None +key string +value string +dt int +ts double PREHOOK: query: desc alter_coltype partition (dt='100x', ts='6:30pm') PREHOOK: type: DESCTABLE POSTHOOK: query: desc alter_coltype partition (dt='100x', ts='6:30pm') @@ -1249,18 +1233,10 @@ POSTHOOK: Lineage: alter_coltype PARTITION(dt=100x,ts=3.0).key SIMPLE [(src1)src POSTHOOK: Lineage: alter_coltype PARTITION(dt=100x,ts=3.0).value SIMPLE [(src1)src1.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: alter_coltype PARTITION(dt=100x,ts=6:30pm).key SIMPLE [(src1)src1.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: alter_coltype PARTITION(dt=100x,ts=6:30pm).value SIMPLE [(src1)src1.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None -dt int None -ts double None - -# Partition Information -# col_name data_type comment - -dt int None -ts double None +key string +value string +dt int +ts double PREHOOK: query: desc alter_coltype partition (dt='100x', ts=3.0) PREHOOK: type: DESCTABLE POSTHOOK: query: desc alter_coltype partition (dt='100x', ts=3.0) @@ -1271,18 +1247,10 @@ POSTHOOK: Lineage: alter_coltype PARTITION(dt=100x,ts=3.0).key SIMPLE [(src1)src POSTHOOK: Lineage: alter_coltype PARTITION(dt=100x,ts=3.0).value SIMPLE [(src1)src1.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: alter_coltype PARTITION(dt=100x,ts=6:30pm).key SIMPLE [(src1)src1.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: alter_coltype PARTITION(dt=100x,ts=6:30pm).value SIMPLE [(src1)src1.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None -dt int None -ts double None - -# Partition Information -# col_name data_type comment - -dt int None -ts double None +key string +value string +dt int +ts double PREHOOK: query: desc alter_coltype partition (dt=10, ts=3.0) PREHOOK: type: DESCTABLE POSTHOOK: query: desc alter_coltype partition (dt=10, ts=3.0) @@ -1293,18 +1261,10 @@ POSTHOOK: Lineage: alter_coltype PARTITION(dt=100x,ts=3.0).key SIMPLE [(src1)src POSTHOOK: Lineage: alter_coltype PARTITION(dt=100x,ts=3.0).value SIMPLE [(src1)src1.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: alter_coltype PARTITION(dt=100x,ts=6:30pm).key SIMPLE [(src1)src1.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: alter_coltype PARTITION(dt=100x,ts=6:30pm).value SIMPLE [(src1)src1.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None -dt int None -ts double None - -# Partition Information -# col_name data_type comment - -dt int None -ts double None +key string +value string +dt int +ts double PREHOOK: query: drop table alter_coltype PREHOOK: type: DROPTABLE PREHOOK: Input: default@alter_coltype diff --git ql/src/test/results/clientpositive/alter_partition_format_loc.q.out ql/src/test/results/clientpositive/alter_partition_format_loc.q.out index 22f3467..6d3f631 100644 --- ql/src/test/results/clientpositive/alter_partition_format_loc.q.out +++ ql/src/test/results/clientpositive/alter_partition_format_loc.q.out @@ -7,10 +7,8 @@ PREHOOK: query: desc extended alter_partition_format_test PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended alter_partition_format_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None +key int +value string #### A masked pattern was here #### PREHOOK: query: alter table alter_partition_format_test set fileformat rcfile @@ -25,10 +23,8 @@ PREHOOK: query: desc extended alter_partition_format_test PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended alter_partition_format_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None +key int +value string #### A masked pattern was here #### PREHOOK: type: ALTERTABLE_LOCATION @@ -42,10 +38,8 @@ PREHOOK: query: desc extended alter_partition_format_test PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended alter_partition_format_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None +key int +value string #### A masked pattern was here #### PREHOOK: query: drop table alter_partition_format_test @@ -74,16 +68,9 @@ PREHOOK: query: desc extended alter_partition_format_test partition(ds='2010') PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended alter_partition_format_test partition(ds='2010') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key int +value string +ds string #### A masked pattern was here #### PREHOOK: query: alter table alter_partition_format_test partition(ds='2010') set fileformat rcfile @@ -99,16 +86,9 @@ PREHOOK: query: desc extended alter_partition_format_test partition(ds='2010') PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended alter_partition_format_test partition(ds='2010') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key int +value string +ds string #### A masked pattern was here #### PREHOOK: type: ALTERPARTITION_LOCATION @@ -123,32 +103,18 @@ PREHOOK: query: desc extended alter_partition_format_test partition(ds='2010') PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended alter_partition_format_test partition(ds='2010') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key int +value string +ds string #### A masked pattern was here #### PREHOOK: query: desc extended alter_partition_format_test PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended alter_partition_format_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key int +value string +ds string #### A masked pattern was here #### PREHOOK: query: alter table alter_partition_format_test set fileformat rcfile @@ -163,16 +129,9 @@ PREHOOK: query: desc extended alter_partition_format_test PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended alter_partition_format_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key int +value string +ds string #### A masked pattern was here #### PREHOOK: type: ALTERTABLE_LOCATION @@ -186,16 +145,9 @@ PREHOOK: query: desc extended alter_partition_format_test PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended alter_partition_format_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key int +value string +ds string #### A masked pattern was here #### PREHOOK: query: drop table alter_partition_format_test diff --git ql/src/test/results/clientpositive/alter_table_serde.q.out ql/src/test/results/clientpositive/alter_table_serde.q.out index 0c80720..c88a985 100644 --- ql/src/test/results/clientpositive/alter_table_serde.q.out +++ ql/src/test/results/clientpositive/alter_table_serde.q.out @@ -9,11 +9,9 @@ PREHOOK: query: describe extended test_table PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended test_table POSTHOOK: type: DESCTABLE -# col_name data_type comment - -id int None -query string None -name string None +id int +query string +name string #### A masked pattern was here #### PREHOOK: query: alter table test_table set serde 'org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe' @@ -28,11 +26,9 @@ PREHOOK: query: describe extended test_table PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended test_table POSTHOOK: type: DESCTABLE -# col_name data_type comment - -id int from deserializer -query string from deserializer -name string from deserializer +id int from deserializer +query string from deserializer +name string from deserializer #### A masked pattern was here #### PREHOOK: query: alter table test_table set serdeproperties ('field.delim' = ',') @@ -47,11 +43,9 @@ PREHOOK: query: describe extended test_table PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended test_table POSTHOOK: type: DESCTABLE -# col_name data_type comment - -id int from deserializer -query string from deserializer -name string from deserializer +id int from deserializer +query string from deserializer +name string from deserializer #### A masked pattern was here #### PREHOOK: query: drop table test_table @@ -80,17 +74,10 @@ PREHOOK: query: describe extended test_table partition (dt='2011') PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended test_table partition (dt='2011') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -id int None -query string None -name string None -dt string None - -# Partition Information -# col_name data_type comment - -dt string None +id int +query string +name string +dt string #### A masked pattern was here #### PREHOOK: query: alter table test_table set serde 'org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe' @@ -105,17 +92,10 @@ PREHOOK: query: describe extended test_table partition (dt='2011') PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended test_table partition (dt='2011') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -id int None -query string None -name string None -dt string None - -# Partition Information -# col_name data_type comment - -dt string None +id int +query string +name string +dt string #### A masked pattern was here #### PREHOOK: query: alter table test_table set serdeproperties ('field.delim' = ',') @@ -130,17 +110,10 @@ PREHOOK: query: describe extended test_table partition (dt='2011') PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended test_table partition (dt='2011') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -id int None -query string None -name string None -dt string None - -# Partition Information -# col_name data_type comment - -dt string None +id int +query string +name string +dt string #### A masked pattern was here #### PREHOOK: query: -- test partitions @@ -160,17 +133,10 @@ PREHOOK: query: describe extended test_table partition (dt='2011') PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended test_table partition (dt='2011') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -id int None -query string None -name string None -dt string None - -# Partition Information -# col_name data_type comment - -dt string None +id int +query string +name string +dt string #### A masked pattern was here #### PREHOOK: query: alter table test_table partition(dt='2011') set serdeproperties ('field.delim' = ',') @@ -186,17 +152,10 @@ PREHOOK: query: describe extended test_table partition (dt='2011') PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended test_table partition (dt='2011') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -id int None -query string None -name string None -dt string None - -# Partition Information -# col_name data_type comment - -dt string None +id int +query string +name string +dt string #### A masked pattern was here #### PREHOOK: query: drop table test_table diff --git ql/src/test/results/clientpositive/alter_view_rename.q.out ql/src/test/results/clientpositive/alter_view_rename.q.out index db21e01..56db724 100644 --- ql/src/test/results/clientpositive/alter_view_rename.q.out +++ ql/src/test/results/clientpositive/alter_view_rename.q.out @@ -14,11 +14,9 @@ PREHOOK: query: DESCRIBE EXTENDED view1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED view1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -foo int None -bar string None -ds string None +foo int +bar string +ds string #### A masked pattern was here #### PREHOOK: query: ALTER VIEW view1 RENAME TO view2 @@ -34,11 +32,9 @@ PREHOOK: query: DESCRIBE EXTENDED view2 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED view2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -foo int None -bar string None -ds string None +foo int +bar string +ds string #### A masked pattern was here #### PREHOOK: query: SELECT * FROM view2 diff --git ql/src/test/results/clientpositive/archive_corrupt.q.out ql/src/test/results/clientpositive/archive_corrupt.q.out index 0a9d722..c4cff82 100644 --- ql/src/test/results/clientpositive/archive_corrupt.q.out +++ ql/src/test/results/clientpositive/archive_corrupt.q.out @@ -90,18 +90,10 @@ POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=11).key SIMPLE [(srcpar POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: alter table tstsrcpart archive partition (ds='2008-04-08', hr='11') @@ -128,18 +120,10 @@ POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=11).key SIMPLE [(srcpar POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: alter table tstsrcpart unarchive partition (ds='2008-04-08', hr='11') diff --git ql/src/test/results/clientpositive/authorization_7.q.out ql/src/test/results/clientpositive/authorization_7.q.out index ee94e92..285a1ac 100644 --- ql/src/test/results/clientpositive/authorization_7.q.out +++ ql/src/test/results/clientpositive/authorization_7.q.out @@ -11,10 +11,8 @@ PREHOOK: query: DESCRIBE src_authorization_7 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE src_authorization_7 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None +key int +value string PREHOOK: query: DROP TABLE src_authorization_7 PREHOOK: type: DROPTABLE PREHOOK: Input: default@src_authorization_7 @@ -40,10 +38,8 @@ PREHOOK: query: DESCRIBE src_authorization_7 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE src_authorization_7 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None +key int +value string PREHOOK: query: DROP TABLE src_authorization_7 PREHOOK: type: DROPTABLE PREHOOK: Input: default@src_authorization_7 diff --git ql/src/test/results/clientpositive/autogen_colalias.q.out ql/src/test/results/clientpositive/autogen_colalias.q.out index 889494d..3964079 100644 --- ql/src/test/results/clientpositive/autogen_colalias.q.out +++ ql/src/test/results/clientpositive/autogen_colalias.q.out @@ -19,18 +19,16 @@ PREHOOK: query: describe dest_grouped_old1 PREHOOK: type: DESCTABLE POSTHOOK: query: describe dest_grouped_old1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -_c0 int None -zz int None -key string None -_c3 int None -_c4 bigint None -_c5 double None -_c6 bigint None -_c7 bigint None -_c8 int None -_c9 int None +_c0 int +zz int +key string +_c3 int +_c4 bigint +_c5 double +_c6 bigint +_c7 bigint +_c8 int +_c9 int PREHOOK: query: create table dest_grouped_old2 as select distinct src.key from src PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src @@ -42,9 +40,7 @@ PREHOOK: query: describe dest_grouped_old2 PREHOOK: type: DESCTABLE POSTHOOK: query: describe dest_grouped_old2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string None +key string PREHOOK: query: create table dest_grouped_new1 as select 1+1, 2+2 as zz, ((src.key % 2)+2)/2, test_max(length(src.value)), count(src.value), sin(count(src.value)), count(sin(src.value)), unix_timestamp(), CAST(SUM(IF(value > 10, value, 1)) AS INT), if(src.key > 10, (src.key +5) % 2, 0) @@ -62,18 +58,16 @@ PREHOOK: query: describe dest_grouped_new1 PREHOOK: type: DESCTABLE POSTHOOK: query: describe dest_grouped_new1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -column_0 int None -zz int None -column_2 double None -test_max_length_src__3 int None -count_src_value_4 bigint None -sin_count_src_value_5 double None -count_sin_src_value_6 bigint None -unix_timestamp_7 bigint None -sum_if_value_10_valu_8 int None -if_src_key_10_src_ke_9 double None +column_0 int +zz int +column_2 double +test_max_length_src__3 int +count_src_value_4 bigint +sin_count_src_value_5 double +count_sin_src_value_6 bigint +unix_timestamp_7 bigint +sum_if_value_10_valu_8 int +if_src_key_10_src_ke_9 double PREHOOK: query: create table dest_grouped_new2 as select distinct src.key from src PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src @@ -85,9 +79,7 @@ PREHOOK: query: describe dest_grouped_new2 PREHOOK: type: DESCTABLE POSTHOOK: query: describe dest_grouped_new2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string None +key string PREHOOK: query: -- Drop the temporary function at the end till HIVE-3160 gets fixed DROP TEMPORARY FUNCTION test_max PREHOOK: type: DROPFUNCTION diff --git ql/src/test/results/clientpositive/avro_change_schema.q.out ql/src/test/results/clientpositive/avro_change_schema.q.out index 87edc83..ad56a65 100644 --- ql/src/test/results/clientpositive/avro_change_schema.q.out +++ ql/src/test/results/clientpositive/avro_change_schema.q.out @@ -33,10 +33,8 @@ PREHOOK: query: DESCRIBE avro2 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE avro2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -string1 string from deserializer -string2 string from deserializer +string1 string from deserializer +string2 string from deserializer PREHOOK: query: ALTER TABLE avro2 SET TBLPROPERTIES ('avro.schema.literal'='{ "namespace": "org.apache.hive", "name": "second_schema", "type": "record", @@ -63,8 +61,6 @@ PREHOOK: query: DESCRIBE avro2 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE avro2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -int1 int from deserializer -float1 float from deserializer -double1 double from deserializer +int1 int from deserializer +float1 float from deserializer +double1 double from deserializer diff --git ql/src/test/results/clientpositive/avro_evolved_schemas.q.out ql/src/test/results/clientpositive/avro_evolved_schemas.q.out index 40d700d..b2fd75c 100644 --- ql/src/test/results/clientpositive/avro_evolved_schemas.q.out +++ ql/src/test/results/clientpositive/avro_evolved_schemas.q.out @@ -75,12 +75,10 @@ PREHOOK: query: DESCRIBE doctors_with_new_field PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE doctors_with_new_field POSTHOOK: type: DESCTABLE -# col_name data_type comment - -number int from deserializer -first_name string from deserializer -last_name string from deserializer -extra_field string from deserializer +number int from deserializer +first_name string from deserializer +last_name string from deserializer +extra_field string from deserializer PREHOOK: query: LOAD DATA LOCAL INPATH '../data/files/doctors.avro' INTO TABLE doctors_with_new_field PREHOOK: type: LOAD PREHOOK: Output: default@doctors_with_new_field diff --git ql/src/test/results/clientpositive/avro_joins.q.out ql/src/test/results/clientpositive/avro_joins.q.out index db98485..208c09d 100644 --- ql/src/test/results/clientpositive/avro_joins.q.out +++ ql/src/test/results/clientpositive/avro_joins.q.out @@ -77,12 +77,10 @@ PREHOOK: query: DESCRIBE doctors4 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE doctors4 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -number int from deserializer -first_name string from deserializer -last_name string from deserializer -extra_field string from deserializer +number int from deserializer +first_name string from deserializer +last_name string from deserializer +extra_field string from deserializer PREHOOK: query: LOAD DATA LOCAL INPATH '../data/files/doctors.avro' INTO TABLE doctors4 PREHOOK: type: LOAD PREHOOK: Output: default@doctors4 @@ -152,11 +150,9 @@ PREHOOK: query: DESCRIBE episodes PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE episodes POSTHOOK: type: DESCTABLE -# col_name data_type comment - -title string from deserializer -air_date string from deserializer -doctor int from deserializer +title string from deserializer +air_date string from deserializer +doctor int from deserializer PREHOOK: query: LOAD DATA LOCAL INPATH '../data/files/episodes.avro' INTO TABLE episodes PREHOOK: type: LOAD PREHOOK: Output: default@episodes diff --git ql/src/test/results/clientpositive/avro_sanity_test.q.out ql/src/test/results/clientpositive/avro_sanity_test.q.out index e625a4c..40e4e42 100644 --- ql/src/test/results/clientpositive/avro_sanity_test.q.out +++ ql/src/test/results/clientpositive/avro_sanity_test.q.out @@ -63,11 +63,9 @@ PREHOOK: query: DESCRIBE doctors PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE doctors POSTHOOK: type: DESCTABLE -# col_name data_type comment - -number int from deserializer -first_name string from deserializer -last_name string from deserializer +number int from deserializer +first_name string from deserializer +last_name string from deserializer PREHOOK: query: LOAD DATA LOCAL INPATH '../data/files/doctors.avro' INTO TABLE doctors PREHOOK: type: LOAD PREHOOK: Output: default@doctors diff --git ql/src/test/results/clientpositive/avro_schema_error_message.q.out ql/src/test/results/clientpositive/avro_schema_error_message.q.out index b699181..4fe9581 100644 --- ql/src/test/results/clientpositive/avro_schema_error_message.q.out +++ ql/src/test/results/clientpositive/avro_schema_error_message.q.out @@ -21,12 +21,10 @@ PREHOOK: query: DESCRIBE avro_with_no_schema PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE avro_with_no_schema POSTHOOK: type: DESCTABLE -# col_name data_type comment - -error_error_error_error_error_error_error string from deserializer -cannot_determine_schema string from deserializer -check string from deserializer -schema string from deserializer -url string from deserializer -and string from deserializer -literal string from deserializer +error_error_error_error_error_error_error string from deserializer +cannot_determine_schema string from deserializer +check string from deserializer +schema string from deserializer +url string from deserializer +and string from deserializer +literal string from deserializer diff --git ql/src/test/results/clientpositive/avro_schema_literal.q.out ql/src/test/results/clientpositive/avro_schema_literal.q.out index 87de151..4f821d1 100644 --- ql/src/test/results/clientpositive/avro_schema_literal.q.out +++ ql/src/test/results/clientpositive/avro_schema_literal.q.out @@ -63,21 +63,19 @@ PREHOOK: query: DESCRIBE avro1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE avro1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -string1 string from deserializer -int1 int from deserializer -tinyint1 int from deserializer -smallint1 int from deserializer -bigint1 bigint from deserializer -boolean1 boolean from deserializer -float1 float from deserializer -double1 double from deserializer -list1 array from deserializer -map1 map from deserializer -struct1 struct from deserializer -union1 uniontype from deserializer -enum1 string from deserializer -nullableint int from deserializer -bytes1 array from deserializer -fixed1 array from deserializer +string1 string from deserializer +int1 int from deserializer +tinyint1 int from deserializer +smallint1 int from deserializer +bigint1 bigint from deserializer +boolean1 boolean from deserializer +float1 float from deserializer +double1 double from deserializer +list1 array from deserializer +map1 map from deserializer +struct1 struct from deserializer +union1 uniontype from deserializer +enum1 string from deserializer +nullableint int from deserializer +bytes1 array from deserializer +fixed1 array from deserializer diff --git ql/src/test/results/clientpositive/ba_table1.q.out ql/src/test/results/clientpositive/ba_table1.q.out index 2b98c59..3da85ea 100644 --- ql/src/test/results/clientpositive/ba_table1.q.out +++ ql/src/test/results/clientpositive/ba_table1.q.out @@ -15,10 +15,8 @@ PREHOOK: query: describe extended ba_test PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended ba_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -ba_key binary None -ba_val binary None +ba_key binary +ba_val binary #### A masked pattern was here #### PREHOOK: query: from src insert overwrite table ba_test select cast (src.key as binary), cast (src.value as binary) diff --git ql/src/test/results/clientpositive/ba_table2.q.out ql/src/test/results/clientpositive/ba_table2.q.out index 8942cb9..9a00dc6 100644 --- ql/src/test/results/clientpositive/ba_table2.q.out +++ ql/src/test/results/clientpositive/ba_table2.q.out @@ -23,10 +23,8 @@ PREHOOK: query: describe extended ba_test PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended ba_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -ba_key binary from deserializer -ba_val binary from deserializer +ba_key binary from deserializer +ba_val binary from deserializer #### A masked pattern was here #### PREHOOK: query: from src insert overwrite table ba_test select cast (src.key as binary), cast (src.value as binary) diff --git ql/src/test/results/clientpositive/ba_table_union.q.out ql/src/test/results/clientpositive/ba_table_union.q.out index 85ea093..9e4bf82 100644 --- ql/src/test/results/clientpositive/ba_table_union.q.out +++ ql/src/test/results/clientpositive/ba_table_union.q.out @@ -13,10 +13,8 @@ PREHOOK: query: describe extended ba_test PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended ba_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -ba_key binary None -ba_val binary None +ba_key binary +ba_val binary #### A masked pattern was here #### PREHOOK: query: from src insert overwrite table ba_test select cast (src.key as binary), cast (src.value as binary) diff --git ql/src/test/results/clientpositive/binary_table_bincolserde.q.out ql/src/test/results/clientpositive/binary_table_bincolserde.q.out index 2b90acd..3f5e371 100644 --- ql/src/test/results/clientpositive/binary_table_bincolserde.q.out +++ ql/src/test/results/clientpositive/binary_table_bincolserde.q.out @@ -23,10 +23,8 @@ PREHOOK: query: describe extended ba_test PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended ba_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -ba_key binary from deserializer -ba_val binary from deserializer +ba_key binary from deserializer +ba_val binary from deserializer #### A masked pattern was here #### PREHOOK: query: from src insert overwrite table ba_test select cast (src.key as binary), cast (src.value as binary) diff --git ql/src/test/results/clientpositive/binary_table_colserde.q.out ql/src/test/results/clientpositive/binary_table_colserde.q.out index e747129..33ae7ea 100644 --- ql/src/test/results/clientpositive/binary_table_colserde.q.out +++ ql/src/test/results/clientpositive/binary_table_colserde.q.out @@ -15,10 +15,8 @@ PREHOOK: query: describe extended ba_test PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended ba_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -ba_key binary None -ba_val binary None +ba_key binary +ba_val binary #### A masked pattern was here #### PREHOOK: query: from src insert overwrite table ba_test select cast (src.key as binary), cast (src.value as binary) diff --git ql/src/test/results/clientpositive/bucket_groupby.q.out ql/src/test/results/clientpositive/bucket_groupby.q.out index 0ff90d1..365fa12 100644 --- ql/src/test/results/clientpositive/bucket_groupby.q.out +++ ql/src/test/results/clientpositive/bucket_groupby.q.out @@ -7,16 +7,9 @@ PREHOOK: query: describe extended clustergroupby PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended clustergroupby POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key string +value string +ds string #### A masked pattern was here #### PREHOOK: query: alter table clustergroupby clustered by (key) into 1 buckets @@ -143,16 +136,9 @@ POSTHOOK: query: describe extended clustergroupby POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key string +value string +ds string #### A masked pattern was here #### PREHOOK: query: insert overwrite table clustergroupby partition (ds='101') select key, value from src distribute by key @@ -1225,16 +1211,9 @@ POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).key SIMPLE [(src)src.FieldSc POSTHOOK: Lineage: clustergroupby PARTITION(ds=100).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: clustergroupby PARTITION(ds=101).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: clustergroupby PARTITION(ds=101).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key string +value string +ds string #### A masked pattern was here #### PREHOOK: query: insert overwrite table clustergroupby partition (ds='102') select key, value from src distribute by value sort by key, value @@ -1636,16 +1615,9 @@ POSTHOOK: Lineage: clustergroupby PARTITION(ds=101).key SIMPLE [(src)src.FieldSc POSTHOOK: Lineage: clustergroupby PARTITION(ds=101).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: clustergroupby PARTITION(ds=102).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: clustergroupby PARTITION(ds=102).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key string +value string +ds string #### A masked pattern was here #### PREHOOK: query: insert overwrite table clustergroupby partition (ds='103') select key, value from src distribute by value, key sort by key diff --git ql/src/test/results/clientpositive/columnarserde_create_shortcut.q.out ql/src/test/results/clientpositive/columnarserde_create_shortcut.q.out index f4b2830..70c3b0b 100644 --- ql/src/test/results/clientpositive/columnarserde_create_shortcut.q.out +++ ql/src/test/results/clientpositive/columnarserde_create_shortcut.q.out @@ -182,10 +182,8 @@ POSTHOOK: Lineage: columnarserde_create_shortcut.d SIMPLE [(src_thrift)src_thrif POSTHOOK: Lineage: columnarserde_create_shortcut.e SIMPLE [(src_thrift)src_thrift.FieldSchema(name:astring, type:string, comment:from deserializer), ] POSTHOOK: Lineage: columnshortcuttable.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: columnshortcuttable.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None +key string +value string PREHOOK: query: SELECT columnShortcutTable.* FROM columnShortcutTable ORDER BY key ASC, value ASC PREHOOK: type: QUERY PREHOOK: Input: default@columnshortcuttable diff --git ql/src/test/results/clientpositive/combine3.q.out ql/src/test/results/clientpositive/combine3.q.out index 331e8d6..00b4d1f 100644 --- ql/src/test/results/clientpositive/combine3.q.out +++ ql/src/test/results/clientpositive/combine3.q.out @@ -47,18 +47,10 @@ POSTHOOK: Lineage: combine_3_srcpart_seq_rc PARTITION(ds=2010-08-03,hr=001).key POSTHOOK: Lineage: combine_3_srcpart_seq_rc PARTITION(ds=2010-08-03,hr=001).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: combine_3_srcpart_seq_rc PARTITION(ds=2010-08-03,hr=00).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: combine_3_srcpart_seq_rc PARTITION(ds=2010-08-03,hr=00).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key int None -value string None -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key int +value string +ds string +hr string #### A masked pattern was here #### PREHOOK: query: desc extended combine_3_srcpart_seq_rc partition(ds="2010-08-03", hr="001") @@ -69,18 +61,10 @@ POSTHOOK: Lineage: combine_3_srcpart_seq_rc PARTITION(ds=2010-08-03,hr=001).key POSTHOOK: Lineage: combine_3_srcpart_seq_rc PARTITION(ds=2010-08-03,hr=001).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: combine_3_srcpart_seq_rc PARTITION(ds=2010-08-03,hr=00).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: combine_3_srcpart_seq_rc PARTITION(ds=2010-08-03,hr=00).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key int None -value string None -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key int +value string +ds string +hr string #### A masked pattern was here #### PREHOOK: query: select key, value, ds, hr from combine_3_srcpart_seq_rc where ds="2010-08-03" order by key, hr limit 30 diff --git ql/src/test/results/clientpositive/convert_enum_to_string.q.out ql/src/test/results/clientpositive/convert_enum_to_string.q.out index af981a7..e58ed44 100644 --- ql/src/test/results/clientpositive/convert_enum_to_string.q.out +++ ql/src/test/results/clientpositive/convert_enum_to_string.q.out @@ -21,32 +21,25 @@ PREHOOK: query: describe convert_enum_to_string PREHOOK: type: DESCTABLE POSTHOOK: query: describe convert_enum_to_string POSTHOOK: type: DESCTABLE -# col_name data_type comment - -my_bool boolean from deserializer -my_byte tinyint from deserializer -my_16bit_int smallint from deserializer -my_32bit_int int from deserializer -my_64bit_int bigint from deserializer -my_double double from deserializer -my_string string from deserializer -my_binary struct from deserializer -my_string_string_map map from deserializer -my_string_enum_map map from deserializer -my_enum_string_map map from deserializer -my_enum_struct_map map>> from deserializer -my_enum_stringlist_map map> from deserializer -my_enum_structlist_map map>>> from deserializer -my_stringlist array from deserializer -my_structlist array>> from deserializer -my_enumlist array from deserializer -my_stringset struct<> from deserializer -my_enumset struct<> from deserializer -my_structset struct<> from deserializer -optionals struct<> from deserializer -b string None - -# Partition Information -# col_name data_type comment - -b string None +my_bool boolean from deserializer +my_byte tinyint from deserializer +my_16bit_int smallint from deserializer +my_32bit_int int from deserializer +my_64bit_int bigint from deserializer +my_double double from deserializer +my_string string from deserializer +my_binary struct from deserializer +my_string_string_map map from deserializer +my_string_enum_map map from deserializer +my_enum_string_map map from deserializer +my_enum_struct_map map>> from deserializer +my_enum_stringlist_map map> from deserializer +my_enum_structlist_map map>>> from deserializer +my_stringlist array from deserializer +my_structlist array>> from deserializer +my_enumlist array from deserializer +my_stringset struct<> from deserializer +my_enumset struct<> from deserializer +my_structset struct<> from deserializer +optionals struct<> from deserializer +b string diff --git ql/src/test/results/clientpositive/create_1.q.out ql/src/test/results/clientpositive/create_1.q.out index ed65478..8668420 100644 --- ql/src/test/results/clientpositive/create_1.q.out +++ ql/src/test/results/clientpositive/create_1.q.out @@ -7,18 +7,14 @@ PREHOOK: query: DESCRIBE table1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE table1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b string None +a string +b string PREHOOK: query: DESCRIBE EXTENDED table1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED table1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b string None +a string +b string #### A masked pattern was here #### PREHOOK: query: CREATE TABLE IF NOT EXISTS table1 (a STRING, b STRING) STORED AS TEXTFILE @@ -34,18 +30,14 @@ PREHOOK: query: DESCRIBE table2 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE table2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b int None +a string +b int PREHOOK: query: DESCRIBE EXTENDED table2 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED table2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b int None +a string +b int #### A masked pattern was here #### PREHOOK: query: CREATE TABLE table3 (a STRING, b STRING) @@ -61,18 +53,14 @@ PREHOOK: query: DESCRIBE table3 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE table3 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b string None +a string +b string PREHOOK: query: DESCRIBE EXTENDED table3 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED table3 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b string None +a string +b string #### A masked pattern was here #### PREHOOK: query: CREATE TABLE table4 (a STRING, b STRING) @@ -88,18 +76,14 @@ PREHOOK: query: DESCRIBE table4 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE table4 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b string None +a string +b string PREHOOK: query: DESCRIBE EXTENDED table4 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED table4 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b string None +a string +b string #### A masked pattern was here #### PREHOOK: query: CREATE TABLE table5 (a STRING, b STRING) @@ -115,17 +99,13 @@ PREHOOK: query: DESCRIBE table5 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE table5 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b string None +a string +b string PREHOOK: query: DESCRIBE EXTENDED table5 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED table5 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b string None +a string +b string #### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/create_default_prop.q.out ql/src/test/results/clientpositive/create_default_prop.q.out index b8f9985..ccd7a7e 100644 --- ql/src/test/results/clientpositive/create_default_prop.q.out +++ ql/src/test/results/clientpositive/create_default_prop.q.out @@ -7,9 +7,7 @@ PREHOOK: query: DESC EXTENDED table_p1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESC EXTENDED table_p1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None +a string #### A masked pattern was here #### PREHOOK: query: CREATE TABLE table_p2 LIKE table_p1 @@ -21,9 +19,7 @@ PREHOOK: query: DESC EXTENDED table_p2 PREHOOK: type: DESCTABLE POSTHOOK: query: DESC EXTENDED table_p2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None +a string #### A masked pattern was here #### PREHOOK: query: CREATE TABLE table_p3 AS SELECT * FROM table_p1 @@ -37,8 +33,6 @@ PREHOOK: query: DESC EXTENDED table_p3 PREHOOK: type: DESCTABLE POSTHOOK: query: DESC EXTENDED table_p3 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None +a string #### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/create_escape.q.out ql/src/test/results/clientpositive/create_escape.q.out index f6b93e6..37f693d 100644 --- ql/src/test/results/clientpositive/create_escape.q.out +++ ql/src/test/results/clientpositive/create_escape.q.out @@ -11,18 +11,14 @@ PREHOOK: query: DESCRIBE table1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE table1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b string None +a string +b string PREHOOK: query: DESCRIBE EXTENDED table1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED table1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b string None +a string +b string #### A masked pattern was here #### PREHOOK: query: INSERT OVERWRITE TABLE table1 SELECT key, '\\\t\\' FROM src WHERE key = 86 diff --git ql/src/test/results/clientpositive/create_insert_outputformat.q.out ql/src/test/results/clientpositive/create_insert_outputformat.q.out index d6599a7..1ca2281 100644 --- ql/src/test/results/clientpositive/create_insert_outputformat.q.out +++ ql/src/test/results/clientpositive/create_insert_outputformat.q.out @@ -25,10 +25,8 @@ POSTHOOK: query: describe table_test_output_format POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: table_test_output_format.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: table_test_output_format.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key int None -value string None +key int +value string PREHOOK: query: CREATE TABLE table_test_output_format_sequencefile(key INT, value STRING) STORED AS INPUTFORMAT 'org.apache.hadoop.mapred.SequenceFileInputFormat' OUTPUTFORMAT 'org.apache.hadoop.mapred.SequenceFileOutputFormat' @@ -62,10 +60,8 @@ POSTHOOK: Lineage: table_test_output_format.key EXPRESSION [(src)src.FieldSchema POSTHOOK: Lineage: table_test_output_format.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: table_test_output_format_sequencefile.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: table_test_output_format_sequencefile.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key int None -value string None +key int +value string PREHOOK: query: CREATE TABLE table_test_output_format_hivesequencefile(key INT, value STRING) STORED AS INPUTFORMAT 'org.apache.hadoop.mapred.SequenceFileInputFormat' OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat' @@ -105,7 +101,5 @@ POSTHOOK: Lineage: table_test_output_format_hivesequencefile.key EXPRESSION [(sr POSTHOOK: Lineage: table_test_output_format_hivesequencefile.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: table_test_output_format_sequencefile.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: table_test_output_format_sequencefile.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key int None -value string None +key int +value string diff --git ql/src/test/results/clientpositive/create_like_view.q.out ql/src/test/results/clientpositive/create_like_view.q.out index cb43bb3..5abe38c 100644 --- ql/src/test/results/clientpositive/create_like_view.q.out +++ ql/src/test/results/clientpositive/create_like_view.q.out @@ -23,10 +23,8 @@ PREHOOK: query: DESCRIBE table1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE table1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b string None +a string +b string PREHOOK: query: DESCRIBE FORMATTED table1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE FORMATTED table1 @@ -72,10 +70,8 @@ PREHOOK: query: DESCRIBE table2 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE table2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b string None +a string +b string PREHOOK: query: DESCRIBE FORMATTED table2 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE FORMATTED table2 @@ -122,10 +118,8 @@ PREHOOK: query: DESCRIBE table3 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE table3 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b string None +a string +b string PREHOOK: query: DESCRIBE FORMATTED table3 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE FORMATTED table3 diff --git ql/src/test/results/clientpositive/create_nested_type.q.out ql/src/test/results/clientpositive/create_nested_type.q.out index 9c44708..d71178a 100644 --- ql/src/test/results/clientpositive/create_nested_type.q.out +++ ql/src/test/results/clientpositive/create_nested_type.q.out @@ -17,22 +17,18 @@ PREHOOK: query: DESCRIBE table1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE table1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b array None -c array> None -d map> None +a string +b array +c array> +d map> PREHOOK: query: DESCRIBE EXTENDED table1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED table1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b array None -c array> None -d map> None +a string +b array +c array> +d map> #### A masked pattern was here #### PREHOOK: query: LOAD DATA LOCAL INPATH '../data/files/create_nested_type.txt' OVERWRITE INTO TABLE table1 diff --git ql/src/test/results/clientpositive/create_view.q.out ql/src/test/results/clientpositive/create_view.q.out index e1e8cea..8ae4415 100644 --- ql/src/test/results/clientpositive/create_view.q.out +++ ql/src/test/results/clientpositive/create_view.q.out @@ -225,16 +225,12 @@ PREHOOK: query: DESCRIBE view1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE view1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -value string None +value string PREHOOK: query: DESCRIBE EXTENDED view1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED view1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -value string None +value string #### A masked pattern was here #### PREHOOK: query: DESCRIBE FORMATTED view1 @@ -270,18 +266,14 @@ PREHOOK: query: DESCRIBE view2 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE view2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string None -value string None +key string +value string PREHOOK: query: DESCRIBE EXTENDED view2 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED view2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string None -value string None +key string +value string #### A masked pattern was here #### PREHOOK: query: DESCRIBE FORMATTED view2 @@ -318,16 +310,12 @@ PREHOOK: query: DESCRIBE view3 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE view3 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -valoo string None +valoo string PREHOOK: query: DESCRIBE EXTENDED view3 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED view3 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -valoo string None +valoo string #### A masked pattern was here #### PREHOOK: query: DESCRIBE FORMATTED view3 @@ -372,9 +360,7 @@ PREHOOK: query: DESCRIBE EXTENDED view3 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED view3 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -valoo string None +valoo string #### A masked pattern was here #### PREHOOK: query: DESCRIBE FORMATTED view3 @@ -421,19 +407,15 @@ POSTHOOK: query: -- use DESCRIBE EXTENDED on a base table and an external table -- of comparison for view descriptions DESCRIBE EXTENDED table1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None +key int #### A masked pattern was here #### PREHOOK: query: DESCRIBE EXTENDED src1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED src1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default +key string default +value string default #### A masked pattern was here #### PREHOOK: query: -- use DESCRIBE EXTENDED on a base table as a point of comparison for @@ -444,9 +426,7 @@ POSTHOOK: query: -- use DESCRIBE EXTENDED on a base table as a point of comparis -- view descriptions DESCRIBE EXTENDED table1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None +key int #### A masked pattern was here #### PREHOOK: query: INSERT OVERWRITE TABLE table1 SELECT key FROM src WHERE key = 86 @@ -493,9 +473,7 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE view4 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -key int None +key int PREHOOK: query: ALTER TABLE table1 ADD COLUMNS (value STRING) PREHOOK: type: ALTERTABLE_ADDCOLS PREHOOK: Input: default@table1 @@ -532,18 +510,14 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE table1 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -key int None -value string None +key int +value string PREHOOK: query: DESCRIBE view4 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE view4 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -key int None +key int PREHOOK: query: CREATE VIEW view5 AS SELECT v1.key as key1, v2.key as key2 FROM view4 v1 join view4 v2 PREHOOK: type: CREATEVIEW @@ -575,10 +549,8 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE view5 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -key1 int None -key2 int None +key1 int +key2 int PREHOOK: query: -- verify that column name and comment in DDL portion -- overrides column alias in SELECT CREATE VIEW view6(valoo COMMENT 'I cannot spell') AS @@ -598,9 +570,7 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE view6 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -valoo string I cannot spell +valoo string I cannot spell PREHOOK: query: -- verify that ORDER BY and LIMIT are both supported in view def CREATE VIEW view7 AS SELECT * FROM src @@ -734,9 +704,7 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED view8 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -c string None +c string #### A masked pattern was here #### FROM table1, viewExpandedText:SELECT `_c0` AS `c` FROM (SELECT `test_translate`('abc', 'a', 'b') @@ -813,9 +781,7 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED view9 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -m int None +m int #### A masked pattern was here #### FROM src, viewExpandedText:SELECT `_c0` AS `m` FROM (SELECT `test_max`(length(`src`.`value`)) @@ -892,9 +858,7 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED view9 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -m int None +m int #### A masked pattern was here #### FROM src, viewExpandedText:SELECT `_c0` AS `m` FROM (SELECT `test_max`(length(`src`.`value`)) @@ -960,10 +924,8 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED view10 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None +key string +value string #### A masked pattern was here #### PREHOOK: query: DESCRIBE FORMATTED view10 @@ -1035,9 +997,7 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED view11 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -boom int None +boom int #### A masked pattern was here #### FROM table1, viewExpandedText:SELECT `test_explode`(array(1,2,3)) AS (`boom`) @@ -1105,11 +1065,9 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED view12 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None -mycol int None +key string +value string +mycol int #### A masked pattern was here #### PREHOOK: query: DESCRIBE FORMATTED view12 @@ -1193,9 +1151,7 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED view13 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -key int None +key int #### A masked pattern was here #### FROM srcbucket TABLESAMPLE (BUCKET 1 OUT OF 5 ON key) s, viewExpandedText:SELECT `s`.`key` @@ -1292,12 +1248,10 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED view14 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -k1 string None -v1 string None -k2 string None -v2 string None +k1 string +v1 string +k2 string +v2 string #### A masked pattern was here #### unionsrc2.key as k2, unionsrc2.value as v2 @@ -1426,10 +1380,8 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED view15 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -key string None -value_count bigint None +key string +value_count bigint #### A masked pattern was here #### FROM src @@ -1515,9 +1467,7 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED view16 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -value string None +value string #### A masked pattern was here #### FROM src, viewExpandedText:SELECT DISTINCT `src`.`value` @@ -1595,9 +1545,7 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE view16 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -value string None +value string PREHOOK: query: -- Likewise, DROP VIEW IF EXISTS should ignore a matching table name DROP VIEW IF EXISTS table1 PREHOOK: type: DROPVIEW @@ -1614,10 +1562,8 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE table1 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -key int None -value string None +key int +value string PREHOOK: query: -- this should work since currently we don't track view->table -- dependencies for implementing RESTRICT diff --git ql/src/test/results/clientpositive/create_view_partitioned.q.out ql/src/test/results/clientpositive/create_view_partitioned.q.out index d5d5dcb..ee9f0fd 100644 --- ql/src/test/results/clientpositive/create_view_partitioned.q.out +++ ql/src/test/results/clientpositive/create_view_partitioned.q.out @@ -35,15 +35,8 @@ PREHOOK: query: DESCRIBE EXTENDED vp1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED vp1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string None -value string None - -# Partition Information -# col_name data_type comment - -value string None +key string +value string #### A masked pattern was here #### FROM src diff --git ql/src/test/results/clientpositive/database.q.out ql/src/test/results/clientpositive/database.q.out index 8b92629..4264227 100644 --- ql/src/test/results/clientpositive/database.q.out +++ ql/src/test/results/clientpositive/database.q.out @@ -129,18 +129,14 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: -- DESCRIBE table in non-default DB DESCRIBE test_table POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col1 string None +col1 string PREHOOK: query: -- DESCRIBE EXTENDED in non-default DB DESCRIBE EXTENDED test_table PREHOOK: type: DESCTABLE POSTHOOK: query: -- DESCRIBE EXTENDED in non-default DB DESCRIBE EXTENDED test_table POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col1 string None +col1 string #### A masked pattern was here #### PREHOOK: query: -- CREATE LIKE in non-default DB @@ -160,9 +156,7 @@ PREHOOK: query: DESCRIBE EXTENDED test_table_like PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED test_table_like POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col1 string None +col1 string #### A masked pattern was here #### PREHOOK: query: -- LOAD and SELECT @@ -1344,10 +1338,8 @@ POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: bucketized_src.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: bucketized_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: temp_tbl2.c EXPRESSION [(temp_tbl)temp_tbl.null, ] -# col_name data_type comment - -key string default -value string default +key string default +value string default #### A masked pattern was here #### PREHOOK: query: -- character escaping diff --git ql/src/test/results/clientpositive/ddltime.q.out ql/src/test/results/clientpositive/ddltime.q.out index d919481..c110ccc 100644 --- ql/src/test/results/clientpositive/ddltime.q.out +++ ql/src/test/results/clientpositive/ddltime.q.out @@ -7,10 +7,8 @@ PREHOOK: query: desc extended T1 PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended T1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default +key string default +value string default #### A masked pattern was here #### PREHOOK: query: insert overwrite table T1 select * from src @@ -29,10 +27,8 @@ POSTHOOK: query: desc extended T1 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: t1.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: t1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default +key string default +value string default #### A masked pattern was here #### PREHOOK: query: insert overwrite table T1 select /*+ HOLD_DDLTIME*/ * from src @@ -55,10 +51,8 @@ POSTHOOK: Lineage: t1.key SIMPLE [(src)src.FieldSchema(name:key, type:string, co POSTHOOK: Lineage: t1.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: t1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: t1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default +key string default +value string default #### A masked pattern was here #### PREHOOK: query: insert overwrite table T1 select * from src @@ -85,10 +79,8 @@ POSTHOOK: Lineage: t1.key SIMPLE [(src)src.FieldSchema(name:key, type:string, co POSTHOOK: Lineage: t1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: t1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: t1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default +key string default +value string default #### A masked pattern was here #### PREHOOK: query: create table if not exists T2 like srcpart @@ -112,18 +104,10 @@ POSTHOOK: Lineage: t1.key SIMPLE [(src)src.FieldSchema(name:key, type:string, co POSTHOOK: Lineage: t1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: t1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: t1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: insert overwrite table T2 partition (ds = '2010-06-21', hr = '1') select key, value from src where key > 10 @@ -154,18 +138,10 @@ POSTHOOK: Lineage: t1.value SIMPLE [(src)src.FieldSchema(name:value, type:string POSTHOOK: Lineage: t1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: t2 PARTITION(ds=2010-06-21,hr=1).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: t2 PARTITION(ds=2010-06-21,hr=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: insert overwrite table T2 partition (ds = '2010-06-21', hr='1') select /*+ HOLD_DDLTIME */ key, value from src where key > 10 @@ -200,18 +176,10 @@ POSTHOOK: Lineage: t2 PARTITION(ds=2010-06-21,hr=1).key SIMPLE [(src)src.FieldSc POSTHOOK: Lineage: t2 PARTITION(ds=2010-06-21,hr=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: t2 PARTITION(ds=2010-06-21,hr=1).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: t2 PARTITION(ds=2010-06-21,hr=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: insert overwrite table T2 partition (ds='2010-06-01', hr='1') select key, value from src where key > 10 @@ -250,17 +218,9 @@ POSTHOOK: Lineage: t2 PARTITION(ds=2010-06-21,hr=1).key SIMPLE [(src)src.FieldSc POSTHOOK: Lineage: t2 PARTITION(ds=2010-06-21,hr=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: t2 PARTITION(ds=2010-06-21,hr=1).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: t2 PARTITION(ds=2010-06-21,hr=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/describe_pretty.q.out ql/src/test/results/clientpositive/describe_pretty.q.out index d94dfe5..3205c32 100644 --- ql/src/test/results/clientpositive/describe_pretty.q.out +++ ql/src/test/results/clientpositive/describe_pretty.q.out @@ -67,130 +67,66 @@ POSTHOOK: query: -- There will be an extra tab at the end of each comment line i DESCRIBE PRETTY test_table POSTHOOK: type: DESCTABLE -col_name data_type comment - -col1 int col1 one line comment -col2 string col2 - two lines comment -col3 string col3 - three lines - comment -col4 string col4 very long comment that is greater than 80 - chars and is likely to spill into multiple - lines -col5 string col5 very long multi-line comment where each - line is very long by itself and is likely to - spill - into multiple lines. Lorem ipsum dolor sit - amet, consectetur adipiscing elit. Proin in - dolor nisl, sodales - adipiscing tortor. Integer venenatis -col6 string This comment has a very long single word ABCDEF - GHIJKLMNOPQRSTUVXYZabcdefghijklmnopqrstuvzxyz12 - 3 which will not fit in a line by itself for - small column widths. -col7_nocomment string None +col1 int col1 one line comment +col2 string col2 +two lines comment +col3 string col3 +three lines +comment +col4 string col4 very long comment that is greater than 80 chars and is likely to spill into multiple lines +col5 string col5 very long multi-line comment where each line is very long by itself and is likely to spill +into multiple lines. Lorem ipsum dolor sit amet, consectetur adipiscing elit. Proin in dolor nisl, sodales +adipiscing tortor. Integer venenatis +col6 string This comment has a very long single word ABCDEFGHIJKLMNOPQRSTUVXYZabcdefghijklmnopqrstuvzxyz123 which will not fit in a line by itself for small column widths. +col7_nocomment string PREHOOK: query: DESCRIBE PRETTY test_table PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE PRETTY test_table POSTHOOK: type: DESCTABLE -col_name data_type comment - -col1 int col1 one line comment -col2 string col2 - two lines comment -col3 string col3 - three lines - comment -col4 string col4 very long comment that is greater than 80 chars and is likely to spill into multiple lines -col5 string col5 very long multi-line comment where each line is very long by itself and is likely to spill - into multiple lines. Lorem ipsum dolor sit amet, consectetur adipiscing elit. Proin in dolor nisl, sodales - adipiscing tortor. Integer venenatis -col6 string This comment has a very long single word ABCDEFGHIJKLMNOPQRSTUVXYZabcdefghijklmnopqrstuvzxyz123 which will not fit in a line by itself for small column widths. -col7_nocomment string None +col1 int col1 one line comment +col2 string col2 +two lines comment +col3 string col3 +three lines +comment +col4 string col4 very long comment that is greater than 80 chars and is likely to spill into multiple lines +col5 string col5 very long multi-line comment where each line is very long by itself and is likely to spill +into multiple lines. Lorem ipsum dolor sit amet, consectetur adipiscing elit. Proin in dolor nisl, sodales +adipiscing tortor. Integer venenatis +col6 string This comment has a very long single word ABCDEFGHIJKLMNOPQRSTUVXYZabcdefghijklmnopqrstuvzxyz123 which will not fit in a line by itself for small column widths. +col7_nocomment string PREHOOK: query: DESCRIBE PRETTY test_table PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE PRETTY test_table POSTHOOK: type: DESCTABLE -col_name data_type comment - -col1 int col1 one line - comment -col2 string col2 - two lines comment -col3 string col3 - three lines - comment -col4 string col4 very long - comment that is - greater than 80 - chars and is - likely to spill - into multiple - lines -col5 string col5 very long - multi-line - comment where - each line is very - long by itself - and is likely to - spill - into multiple - lines. Lorem - ipsum dolor sit - amet, consectetur - adipiscing elit. - Proin in dolor - nisl, sodales - adipiscing - tortor. Integer - venenatis -col6 string This comment has - a very long - single word ABCDE - FGHIJKLMNOPQRSTUV - XYZabcdefghijklmn - opqrstuvzxyz123 - which will not - fit in a line by - itself for small - column widths. -col7_nocomment string None +col1 int col1 one line comment +col2 string col2 +two lines comment +col3 string col3 +three lines +comment +col4 string col4 very long comment that is greater than 80 chars and is likely to spill into multiple lines +col5 string col5 very long multi-line comment where each line is very long by itself and is likely to spill +into multiple lines. Lorem ipsum dolor sit amet, consectetur adipiscing elit. Proin in dolor nisl, sodales +adipiscing tortor. Integer venenatis +col6 string This comment has a very long single word ABCDEFGHIJKLMNOPQRSTUVXYZabcdefghijklmnopqrstuvzxyz123 which will not fit in a line by itself for small column widths. +col7_nocomment string PREHOOK: query: DESCRIBE PRETTY test_table PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE PRETTY test_table POSTHOOK: type: DESCTABLE -col_name data_type comment - -col1 int col1 one line comment -col2 string col2 - two lines comment -col3 string col3 - three lines - comment -col4 string col4 very long comment that - is greater than 80 chars - and is likely to spill into - multiple lines -col5 string col5 very long multi-line - comment where each line is - very long by itself and is - likely to spill - into multiple lines. Lorem - ipsum dolor sit amet, - consectetur adipiscing - elit. Proin in dolor nisl, - sodales - adipiscing tortor. Integer - venenatis -col6 string This comment has a very - long single word ABCDEFGHIJ - KLMNOPQRSTUVXYZabcdefghijkl - mnopqrstuvzxyz123 which - will not fit in a line by - itself for small column - widths. -col7_nocomment string None +col1 int col1 one line comment +col2 string col2 +two lines comment +col3 string col3 +three lines +comment +col4 string col4 very long comment that is greater than 80 chars and is likely to spill into multiple lines +col5 string col5 very long multi-line comment where each line is very long by itself and is likely to spill +into multiple lines. Lorem ipsum dolor sit amet, consectetur adipiscing elit. Proin in dolor nisl, sodales +adipiscing tortor. Integer venenatis +col6 string This comment has a very long single word ABCDEFGHIJKLMNOPQRSTUVXYZabcdefghijklmnopqrstuvzxyz123 which will not fit in a line by itself for small column widths. +col7_nocomment string PREHOOK: query: CREATE TABLE test_table_very_long_column_name( col1 INT COMMENT 'col1 one line comment', col2_abcdefghiklmnopqrstuvxyz STRING COMMENT 'col2 @@ -214,46 +150,24 @@ PREHOOK: query: DESCRIBE PRETTY test_table_very_long_column_name PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE PRETTY test_table_very_long_column_name POSTHOOK: type: DESCTABLE -col_name data_type comment - -col1 int col1 one line comment -col2_abcdefghiklmnopqrstuvxyz string col2 - two lines comment -col3 string col3 - three lines - comment -col4 string col4 very long comment that is - greater than 80 chars and is - likely to spill into multiple - lines +col1 int col1 one line comment +col2_abcdefghiklmnopqrstuvxyz string col2 +two lines comment +col3 string col3 +three lines +comment +col4 string col4 very long comment that is greater than 80 chars and is likely to spill into multiple lines PREHOOK: query: DESCRIBE PRETTY test_table_very_long_column_name PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE PRETTY test_table_very_long_column_name POSTHOOK: type: DESCTABLE -col_name data_type comment - -col1 int col1 one - line - comment -col2_abcdefghiklmnopqrstuvxyz string col2 - two lines - comment -col3 string col3 - three - lines - comment -col4 string col4 very - long - comment - that is - greater - than 80 - chars and - is likely - to spill - into - multiple - lines +col1 int col1 one line comment +col2_abcdefghiklmnopqrstuvxyz string col2 +two lines comment +col3 string col3 +three lines +comment +col4 string col4 very long comment that is greater than 80 chars and is likely to spill into multiple lines PREHOOK: query: CREATE TABLE test_table_partitioned( col1 INT COMMENT 'col1 one line comment', col2 STRING COMMENT 'col2 @@ -293,40 +207,16 @@ PREHOOK: query: DESCRIBE PRETTY test_table_partitioned PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE PRETTY test_table_partitioned POSTHOOK: type: DESCTABLE -col_name data_type comment - -col1 int col1 one line comment -col2 string col2 - two lines comment -col3 string col3 - three lines - comment -col4 string col4 very long comment that - is greater than 80 chars - and is likely to spill into - multiple lines -col5 string col5 very long multi-line - comment where each line is - very long by itself and is - likely to spill - into multiple lines. Lorem - ipsum dolor sit amet, - consectetur adipiscing - elit. Proin in dolor nisl, - sodales - adipiscing tortor. Integer - venenatis -col6 string This comment has a very - long single word ABCDEFGHIJ - KLMNOPQRSTUVXYZabcdefghijkl - mnopqrstuvzxyz123 which - will not fit in a line by - itself for small column - widths. -col7_nocomment string None -ds string None - -# Partition Information -col_name data_type comment - -ds string None +col1 int col1 one line comment +col2 string col2 +two lines comment +col3 string col3 +three lines +comment +col4 string col4 very long comment that is greater than 80 chars and is likely to spill into multiple lines +col5 string col5 very long multi-line comment where each line is very long by itself and is likely to spill +into multiple lines. Lorem ipsum dolor sit amet, consectetur adipiscing elit. Proin in dolor nisl, sodales +adipiscing tortor. Integer venenatis +col6 string This comment has a very long single word ABCDEFGHIJKLMNOPQRSTUVXYZabcdefghijklmnopqrstuvzxyz123 which will not fit in a line by itself for small column widths. +col7_nocomment string +ds string diff --git ql/src/test/results/clientpositive/describe_syntax.q.out ql/src/test/results/clientpositive/describe_syntax.q.out index fefd983..40f22b1 100644 --- ql/src/test/results/clientpositive/describe_syntax.q.out +++ ql/src/test/results/clientpositive/describe_syntax.q.out @@ -38,34 +38,18 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: -- describe table DESCRIBE t1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key1 int None -value1 string None -ds string None -part string None - -# Partition Information -# col_name data_type comment - -ds string None -part string None +key1 int +value1 string +ds string +part string PREHOOK: query: DESCRIBE EXTENDED t1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED t1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key1 int None -value1 string None -ds string None -part string None - -# Partition Information -# col_name data_type comment - -ds string None -part string None +key1 int +value1 string +ds string +part string #### A masked pattern was here #### PREHOOK: query: DESCRIBE FORMATTED t1 @@ -109,34 +93,18 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: -- describe database.table DESCRIBE db1.t1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key1 int None -value1 string None -ds string None -part string None - -# Partition Information -# col_name data_type comment - -ds string None -part string None +key1 int +value1 string +ds string +part string PREHOOK: query: DESCRIBE EXTENDED db1.t1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED db1.t1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key1 int None -value1 string None -ds string None -part string None - -# Partition Information -# col_name data_type comment - -ds string None -part string None +key1 int +value1 string +ds string +part string #### A masked pattern was here #### PREHOOK: query: DESCRIBE FORMATTED db1.t1 @@ -180,16 +148,12 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: -- describe table column DESCRIBE t1 key1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key1 int from deserializer +key1 int from deserializer PREHOOK: query: DESCRIBE EXTENDED t1 key1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED t1 key1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key1 int from deserializer +key1 int from deserializer PREHOOK: query: DESCRIBE FORMATTED t1 key1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE FORMATTED t1 key1 @@ -203,16 +167,12 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: -- describe database.tabe column DESCRIBE db1.t1 key1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key1 int from deserializer +key1 int from deserializer PREHOOK: query: DESCRIBE EXTENDED db1.t1 key1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED db1.t1 key1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key1 int from deserializer +key1 int from deserializer PREHOOK: query: DESCRIBE FORMATTED db1.t1 key1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE FORMATTED db1.t1 key1 @@ -230,16 +190,12 @@ POSTHOOK: query: -- describe table.column -- fall back to the old syntax table.column DESCRIBE t1.key1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key1 int from deserializer +key1 int from deserializer PREHOOK: query: DESCRIBE EXTENDED t1.key1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED t1.key1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key1 int from deserializer +key1 int from deserializer PREHOOK: query: DESCRIBE FORMATTED t1.key1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE FORMATTED t1.key1 @@ -253,34 +209,18 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: -- describe table partition DESCRIBE t1 PARTITION(ds='4', part='5') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key1 int None -value1 string None -ds string None -part string None - -# Partition Information -# col_name data_type comment - -ds string None -part string None +key1 int +value1 string +ds string +part string PREHOOK: query: DESCRIBE EXTENDED t1 PARTITION(ds='4', part='5') PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED t1 PARTITION(ds='4', part='5') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key1 int None -value1 string None -ds string None -part string None - -# Partition Information -# col_name data_type comment - -ds string None -part string None +key1 int +value1 string +ds string +part string #### A masked pattern was here #### PREHOOK: query: DESCRIBE FORMATTED t1 PARTITION(ds='4', part='5') @@ -324,34 +264,18 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: -- describe database.table partition DESCRIBE db1.t1 PARTITION(ds='4', part='5') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key1 int None -value1 string None -ds string None -part string None - -# Partition Information -# col_name data_type comment - -ds string None -part string None +key1 int +value1 string +ds string +part string PREHOOK: query: DESCRIBE EXTENDED db1.t1 PARTITION(ds='4', part='5') PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED db1.t1 PARTITION(ds='4', part='5') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key1 int None -value1 string None -ds string None -part string None - -# Partition Information -# col_name data_type comment - -ds string None -part string None +key1 int +value1 string +ds string +part string #### A masked pattern was here #### PREHOOK: query: DESCRIBE FORMATTED db1.t1 PARTITION(ds='4', part='5') diff --git ql/src/test/results/clientpositive/describe_table.q.out ql/src/test/results/clientpositive/describe_table.q.out index ea18601..1143f59 100644 --- ql/src/test/results/clientpositive/describe_table.q.out +++ ql/src/test/results/clientpositive/describe_table.q.out @@ -2,82 +2,46 @@ PREHOOK: query: describe srcpart PREHOOK: type: DESCTABLE POSTHOOK: query: describe srcpart POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string PREHOOK: query: describe srcpart.key PREHOOK: type: DESCTABLE POSTHOOK: query: describe srcpart.key POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string from deserializer +key string from deserializer PREHOOK: query: describe srcpart PARTITION(ds='2008-04-08', hr='12') PREHOOK: type: DESCTABLE POSTHOOK: query: describe srcpart PARTITION(ds='2008-04-08', hr='12') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string PREHOOK: query: describe extended srcpart PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended srcpart POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: describe extended srcpart.key PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended srcpart.key POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string from deserializer +key string from deserializer PREHOOK: query: describe extended srcpart PARTITION(ds='2008-04-08', hr='12') PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended srcpart PARTITION(ds='2008-04-08', hr='12') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: describe formatted srcpart diff --git ql/src/test/results/clientpositive/describe_xpath.q.out ql/src/test/results/clientpositive/describe_xpath.q.out index 282d325..1c3a46a 100644 --- ql/src/test/results/clientpositive/describe_xpath.q.out +++ ql/src/test/results/clientpositive/describe_xpath.q.out @@ -4,53 +4,41 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: -- Describe a list structure in a thrift table describe src_thrift.lint POSTHOOK: type: DESCTABLE -# col_name data_type comment - -lint array from deserializer +lint array from deserializer PREHOOK: query: -- Describe the element of a list describe src_thrift.lint.$elem$ PREHOOK: type: DESCTABLE POSTHOOK: query: -- Describe the element of a list describe src_thrift.lint.$elem$ POSTHOOK: type: DESCTABLE -# col_name data_type comment - -$elem$ int from deserializer +$elem$ int from deserializer PREHOOK: query: -- Describe the key of a map describe src_thrift.mStringString.$key$ PREHOOK: type: DESCTABLE POSTHOOK: query: -- Describe the key of a map describe src_thrift.mStringString.$key$ POSTHOOK: type: DESCTABLE -# col_name data_type comment - -$key$ string from deserializer +$key$ string from deserializer PREHOOK: query: -- Describe the value of a map describe src_thrift.mStringString.$value$ PREHOOK: type: DESCTABLE POSTHOOK: query: -- Describe the value of a map describe src_thrift.mStringString.$value$ POSTHOOK: type: DESCTABLE -# col_name data_type comment - -$value$ string from deserializer +$value$ string from deserializer PREHOOK: query: -- Describe a complex element of a list describe src_thrift.lintString.$elem$ PREHOOK: type: DESCTABLE POSTHOOK: query: -- Describe a complex element of a list describe src_thrift.lintString.$elem$ POSTHOOK: type: DESCTABLE -# col_name data_type comment - -myint int from deserializer -mystring string from deserializer -underscore_int int from deserializer +myint int from deserializer +mystring string from deserializer +underscore_int int from deserializer PREHOOK: query: -- Describe a member of an element of a list describe src_thrift.lintString.$elem$.myint PREHOOK: type: DESCTABLE POSTHOOK: query: -- Describe a member of an element of a list describe src_thrift.lintString.$elem$.myint POSTHOOK: type: DESCTABLE -# col_name data_type comment - -myint int from deserializer +myint int from deserializer diff --git ql/src/test/results/clientpositive/drop_partitions_filter.q.out ql/src/test/results/clientpositive/drop_partitions_filter.q.out index 9b36f19..9ebc603 100644 --- ql/src/test/results/clientpositive/drop_partitions_filter.q.out +++ ql/src/test/results/clientpositive/drop_partitions_filter.q.out @@ -7,18 +7,10 @@ PREHOOK: query: describe ptestfilter PREHOOK: type: DESCTABLE POSTHOOK: query: describe ptestfilter POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b int None -c string None -d string None - -# Partition Information -# col_name data_type comment - -c string None -d string None +a string +b int +c string +d string PREHOOK: query: alter table ptestfilter add partition (c='US', d=1) PREHOOK: type: ALTERTABLE_ADDPARTS PREHOOK: Input: default@ptestfilter diff --git ql/src/test/results/clientpositive/drop_partitions_filter2.q.out ql/src/test/results/clientpositive/drop_partitions_filter2.q.out index fb9ae4a..10f45b3 100644 --- ql/src/test/results/clientpositive/drop_partitions_filter2.q.out +++ ql/src/test/results/clientpositive/drop_partitions_filter2.q.out @@ -7,18 +7,10 @@ PREHOOK: query: describe ptestfilter PREHOOK: type: DESCTABLE POSTHOOK: query: describe ptestfilter POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b int None -c int None -d int None - -# Partition Information -# col_name data_type comment - -c int None -d int None +a string +b int +c int +d int PREHOOK: query: alter table ptestfilter add partition (c=1, d=1) PREHOOK: type: ALTERTABLE_ADDPARTS PREHOOK: Input: default@ptestfilter diff --git ql/src/test/results/clientpositive/drop_partitions_filter3.q.out ql/src/test/results/clientpositive/drop_partitions_filter3.q.out index 2033faa..d32721b 100644 --- ql/src/test/results/clientpositive/drop_partitions_filter3.q.out +++ ql/src/test/results/clientpositive/drop_partitions_filter3.q.out @@ -7,18 +7,10 @@ PREHOOK: query: describe ptestfilter PREHOOK: type: DESCTABLE POSTHOOK: query: describe ptestfilter POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b int None -c string None -d int None - -# Partition Information -# col_name data_type comment - -c string None -d int None +a string +b int +c string +d int PREHOOK: query: alter table ptestfilter add partition (c='1', d=1) PREHOOK: type: ALTERTABLE_ADDPARTS PREHOOK: Input: default@ptestfilter diff --git ql/src/test/results/clientpositive/drop_partitions_ignore_protection.q.out ql/src/test/results/clientpositive/drop_partitions_ignore_protection.q.out index e783982..83f4047 100644 --- ql/src/test/results/clientpositive/drop_partitions_ignore_protection.q.out +++ ql/src/test/results/clientpositive/drop_partitions_ignore_protection.q.out @@ -23,16 +23,9 @@ PREHOOK: query: desc extended tbl_protectmode_no_drop partition (p='p1') PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl_protectmode_no_drop partition (p='p1') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -c1 string None -c2 string None -p string None - -# Partition Information -# col_name data_type comment - -p string None +c1 string +c2 string +p string #### A masked pattern was here #### PREHOOK: query: -- The partition will be dropped, even though we have enabled no_drop diff --git ql/src/test/results/clientpositive/exim_00_nonpart_empty.q.out ql/src/test/results/clientpositive/exim_00_nonpart_empty.q.out index 75b7096..b1db059 100644 --- ql/src/test/results/clientpositive/exim_00_nonpart_empty.q.out +++ ql/src/test/results/clientpositive/exim_00_nonpart_empty.q.out @@ -41,9 +41,7 @@ PREHOOK: query: describe extended exim_department PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_department POSTHOOK: type: DESCTABLE -# col_name data_type comment - -dep_id int department id +dep_id int department id #### A masked pattern was here #### PREHOOK: query: show table extended like exim_department diff --git ql/src/test/results/clientpositive/exim_01_nonpart.q.out ql/src/test/results/clientpositive/exim_01_nonpart.q.out index 0c0fd01..910013d 100644 --- ql/src/test/results/clientpositive/exim_01_nonpart.q.out +++ ql/src/test/results/clientpositive/exim_01_nonpart.q.out @@ -47,9 +47,7 @@ PREHOOK: query: describe extended exim_department PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_department POSTHOOK: type: DESCTABLE -# col_name data_type comment - -dep_id int department id +dep_id int department id #### A masked pattern was here #### PREHOOK: query: show table extended like exim_department diff --git ql/src/test/results/clientpositive/exim_02_00_part_empty.q.out ql/src/test/results/clientpositive/exim_02_00_part_empty.q.out index f9c0115..d8d840a 100644 --- ql/src/test/results/clientpositive/exim_02_00_part_empty.q.out +++ ql/src/test/results/clientpositive/exim_02_00_part_empty.q.out @@ -43,17 +43,9 @@ PREHOOK: query: describe extended exim_employee PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_employee POSTHOOK: type: DESCTABLE -# col_name data_type comment - -emp_id int employee id -emp_country string two char iso code -emp_state string free text - -# Partition Information -# col_name data_type comment - -emp_country string two char iso code -emp_state string free text +emp_id int employee id +emp_country string two char iso code +emp_state string free text #### A masked pattern was here #### PREHOOK: query: show table extended like exim_employee diff --git ql/src/test/results/clientpositive/exim_02_part.q.out ql/src/test/results/clientpositive/exim_02_part.q.out index 94949a3..104a1d6 100644 --- ql/src/test/results/clientpositive/exim_02_part.q.out +++ ql/src/test/results/clientpositive/exim_02_part.q.out @@ -55,17 +55,9 @@ PREHOOK: query: describe extended exim_employee PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_employee POSTHOOK: type: DESCTABLE -# col_name data_type comment - -emp_id int employee id -emp_country string two char iso code -emp_state string free text - -# Partition Information -# col_name data_type comment - -emp_country string two char iso code -emp_state string free text +emp_id int employee id +emp_country string two char iso code +emp_state string free text #### A masked pattern was here #### PREHOOK: query: show table extended like exim_employee diff --git ql/src/test/results/clientpositive/exim_03_nonpart_over_compat.q.out ql/src/test/results/clientpositive/exim_03_nonpart_over_compat.q.out index d47b132..0e9d85c 100644 --- ql/src/test/results/clientpositive/exim_03_nonpart_over_compat.q.out +++ ql/src/test/results/clientpositive/exim_03_nonpart_over_compat.q.out @@ -57,9 +57,7 @@ PREHOOK: query: describe extended exim_department PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_department POSTHOOK: type: DESCTABLE -# col_name data_type comment - -dep_id int department identifier +dep_id int department identifier #### A masked pattern was here #### PREHOOK: query: select * from exim_department diff --git ql/src/test/results/clientpositive/exim_04_all_part.q.out ql/src/test/results/clientpositive/exim_04_all_part.q.out index e214dad..6a3acb3 100644 --- ql/src/test/results/clientpositive/exim_04_all_part.q.out +++ ql/src/test/results/clientpositive/exim_04_all_part.q.out @@ -91,17 +91,9 @@ PREHOOK: query: describe extended exim_employee PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_employee POSTHOOK: type: DESCTABLE -# col_name data_type comment - -emp_id int employee id -emp_country string two char iso code -emp_state string free text - -# Partition Information -# col_name data_type comment - -emp_country string two char iso code -emp_state string free text +emp_id int employee id +emp_country string two char iso code +emp_state string free text #### A masked pattern was here #### PREHOOK: query: show table extended like exim_employee diff --git ql/src/test/results/clientpositive/exim_04_evolved_parts.q.out ql/src/test/results/clientpositive/exim_04_evolved_parts.q.out index daf3daf..e510833 100644 --- ql/src/test/results/clientpositive/exim_04_evolved_parts.q.out +++ ql/src/test/results/clientpositive/exim_04_evolved_parts.q.out @@ -101,63 +101,39 @@ PREHOOK: query: describe extended exim_employee PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_employee POSTHOOK: type: DESCTABLE -# col_name data_type comment - -emp_id int from deserializer -emp_name string from deserializer -emp_dob string from deserializer -emp_sex string from deserializer -emp_dept int from deserializer -emp_country string 2-char code -emp_state string 2-char code - -# Partition Information -# col_name data_type comment - -emp_country string 2-char code -emp_state string 2-char code +emp_id int from deserializer +emp_name string from deserializer +emp_dob string from deserializer +emp_sex string from deserializer +emp_dept int from deserializer +emp_country string 2-char code +emp_state string 2-char code #### A masked pattern was here #### PREHOOK: query: describe extended exim_employee partition (emp_country='in', emp_state='tn') PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_employee partition (emp_country='in', emp_state='tn') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -emp_id int employee id -emp_name string None -emp_dob string employee date of birth -emp_sex string M/F -emp_dept int None -emp_country string 2-char code -emp_state string 2-char code - -# Partition Information -# col_name data_type comment - -emp_country string 2-char code -emp_state string 2-char code +emp_id int employee id +emp_name string +emp_dob string employee date of birth +emp_sex string M/F +emp_dept int +emp_country string 2-char code +emp_state string 2-char code #### A masked pattern was here #### PREHOOK: query: describe extended exim_employee partition (emp_country='in', emp_state='ka') PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_employee partition (emp_country='in', emp_state='ka') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -emp_id int from deserializer -emp_name string from deserializer -emp_dob string from deserializer -emp_sex string from deserializer -emp_dept int from deserializer -emp_country string 2-char code -emp_state string 2-char code - -# Partition Information -# col_name data_type comment - -emp_country string 2-char code -emp_state string 2-char code +emp_id int from deserializer +emp_name string from deserializer +emp_dob string from deserializer +emp_sex string from deserializer +emp_dept int from deserializer +emp_country string 2-char code +emp_state string 2-char code #### A masked pattern was here #### PREHOOK: query: show table extended like exim_employee diff --git ql/src/test/results/clientpositive/exim_05_some_part.q.out ql/src/test/results/clientpositive/exim_05_some_part.q.out index 410edfa..2ab36ee 100644 --- ql/src/test/results/clientpositive/exim_05_some_part.q.out +++ ql/src/test/results/clientpositive/exim_05_some_part.q.out @@ -85,17 +85,9 @@ PREHOOK: query: describe extended exim_employee PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_employee POSTHOOK: type: DESCTABLE -# col_name data_type comment - -emp_id int employee id -emp_country string two char iso code -emp_state string free text - -# Partition Information -# col_name data_type comment - -emp_country string two char iso code -emp_state string free text +emp_id int employee id +emp_country string two char iso code +emp_state string free text #### A masked pattern was here #### PREHOOK: query: show table extended like exim_employee diff --git ql/src/test/results/clientpositive/exim_06_one_part.q.out ql/src/test/results/clientpositive/exim_06_one_part.q.out index e1e3a26..b300543 100644 --- ql/src/test/results/clientpositive/exim_06_one_part.q.out +++ ql/src/test/results/clientpositive/exim_06_one_part.q.out @@ -82,17 +82,9 @@ PREHOOK: query: describe extended exim_employee PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_employee POSTHOOK: type: DESCTABLE -# col_name data_type comment - -emp_id int employee id -emp_country string two char iso code -emp_state string free text - -# Partition Information -# col_name data_type comment - -emp_country string two char iso code -emp_state string free text +emp_id int employee id +emp_country string two char iso code +emp_state string free text #### A masked pattern was here #### PREHOOK: query: show table extended like exim_employee diff --git ql/src/test/results/clientpositive/exim_07_all_part_over_nonoverlap.q.out ql/src/test/results/clientpositive/exim_07_all_part_over_nonoverlap.q.out index 4e0ab6d..2decb91 100644 --- ql/src/test/results/clientpositive/exim_07_all_part_over_nonoverlap.q.out +++ ql/src/test/results/clientpositive/exim_07_all_part_over_nonoverlap.q.out @@ -114,17 +114,9 @@ PREHOOK: query: describe extended exim_employee PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_employee POSTHOOK: type: DESCTABLE -# col_name data_type comment - -emp_id int employee id -emp_country string iso code -emp_state string free-form text - -# Partition Information -# col_name data_type comment - -emp_country string iso code -emp_state string free-form text +emp_id int employee id +emp_country string iso code +emp_state string free-form text #### A masked pattern was here #### PREHOOK: query: select * from exim_employee diff --git ql/src/test/results/clientpositive/exim_08_nonpart_rename.q.out ql/src/test/results/clientpositive/exim_08_nonpart_rename.q.out index 1e1d677..b722437 100644 --- ql/src/test/results/clientpositive/exim_08_nonpart_rename.q.out +++ ql/src/test/results/clientpositive/exim_08_nonpart_rename.q.out @@ -65,9 +65,7 @@ PREHOOK: query: describe extended exim_imported_dept PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_imported_dept POSTHOOK: type: DESCTABLE -# col_name data_type comment - -dep_id int department id +dep_id int department id #### A masked pattern was here #### PREHOOK: query: select * from exim_imported_dept diff --git ql/src/test/results/clientpositive/exim_09_part_spec_nonoverlap.q.out ql/src/test/results/clientpositive/exim_09_part_spec_nonoverlap.q.out index 2e20805..47ce26c 100644 --- ql/src/test/results/clientpositive/exim_09_part_spec_nonoverlap.q.out +++ ql/src/test/results/clientpositive/exim_09_part_spec_nonoverlap.q.out @@ -120,17 +120,9 @@ PREHOOK: query: describe extended exim_employee PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_employee POSTHOOK: type: DESCTABLE -# col_name data_type comment - -emp_id int employee id -emp_country string two char iso code -emp_state string free text - -# Partition Information -# col_name data_type comment - -emp_country string two char iso code -emp_state string free text +emp_id int employee id +emp_country string two char iso code +emp_state string free text #### A masked pattern was here #### PREHOOK: query: select * from exim_employee diff --git ql/src/test/results/clientpositive/exim_10_external_managed.q.out ql/src/test/results/clientpositive/exim_10_external_managed.q.out index 4b82efd..530b273 100644 --- ql/src/test/results/clientpositive/exim_10_external_managed.q.out +++ ql/src/test/results/clientpositive/exim_10_external_managed.q.out @@ -51,9 +51,7 @@ PREHOOK: query: describe extended exim_department PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_department POSTHOOK: type: DESCTABLE -# col_name data_type comment - -dep_id int department id +dep_id int department id #### A masked pattern was here #### PREHOOK: query: select * from exim_department diff --git ql/src/test/results/clientpositive/exim_11_managed_external.q.out ql/src/test/results/clientpositive/exim_11_managed_external.q.out index 7c73c96..8b4238d 100644 --- ql/src/test/results/clientpositive/exim_11_managed_external.q.out +++ ql/src/test/results/clientpositive/exim_11_managed_external.q.out @@ -47,9 +47,7 @@ PREHOOK: query: describe extended exim_department PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_department POSTHOOK: type: DESCTABLE -# col_name data_type comment - -dep_id int department id +dep_id int department id #### A masked pattern was here #### PREHOOK: query: select * from exim_department diff --git ql/src/test/results/clientpositive/exim_12_external_location.q.out ql/src/test/results/clientpositive/exim_12_external_location.q.out index 2bd523d..4f40737 100644 --- ql/src/test/results/clientpositive/exim_12_external_location.q.out +++ ql/src/test/results/clientpositive/exim_12_external_location.q.out @@ -50,9 +50,7 @@ PREHOOK: query: describe extended exim_department PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_department POSTHOOK: type: DESCTABLE -# col_name data_type comment - -dep_id int department id +dep_id int department id #### A masked pattern was here #### PREHOOK: query: select * from exim_department diff --git ql/src/test/results/clientpositive/exim_13_managed_location.q.out ql/src/test/results/clientpositive/exim_13_managed_location.q.out index 99f160f..138a396 100644 --- ql/src/test/results/clientpositive/exim_13_managed_location.q.out +++ ql/src/test/results/clientpositive/exim_13_managed_location.q.out @@ -50,9 +50,7 @@ PREHOOK: query: describe extended exim_department PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_department POSTHOOK: type: DESCTABLE -# col_name data_type comment - -dep_id int department id +dep_id int department id #### A masked pattern was here #### PREHOOK: query: select * from exim_department diff --git ql/src/test/results/clientpositive/exim_14_managed_location_over_existing.q.out ql/src/test/results/clientpositive/exim_14_managed_location_over_existing.q.out index 5b22dfa..4042914 100644 --- ql/src/test/results/clientpositive/exim_14_managed_location_over_existing.q.out +++ ql/src/test/results/clientpositive/exim_14_managed_location_over_existing.q.out @@ -62,9 +62,7 @@ PREHOOK: query: describe extended exim_department PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_department POSTHOOK: type: DESCTABLE -# col_name data_type comment - -dep_id int department id +dep_id int department id #### A masked pattern was here #### PREHOOK: query: select * from exim_department diff --git ql/src/test/results/clientpositive/exim_15_external_part.q.out ql/src/test/results/clientpositive/exim_15_external_part.q.out index 7e10f08..98a3700 100644 --- ql/src/test/results/clientpositive/exim_15_external_part.q.out +++ ql/src/test/results/clientpositive/exim_15_external_part.q.out @@ -125,17 +125,9 @@ PREHOOK: query: describe extended exim_employee PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_employee POSTHOOK: type: DESCTABLE -# col_name data_type comment - -emp_id int employee id -emp_country string two char iso code -emp_state string free text - -# Partition Information -# col_name data_type comment - -emp_country string two char iso code -emp_state string free text +emp_id int employee id +emp_country string two char iso code +emp_state string free text #### A masked pattern was here #### PREHOOK: query: select * from exim_employee diff --git ql/src/test/results/clientpositive/exim_18_part_external.q.out ql/src/test/results/clientpositive/exim_18_part_external.q.out index cbf18b2..ceeb342 100644 --- ql/src/test/results/clientpositive/exim_18_part_external.q.out +++ ql/src/test/results/clientpositive/exim_18_part_external.q.out @@ -90,17 +90,9 @@ PREHOOK: query: describe extended exim_employee PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_employee POSTHOOK: type: DESCTABLE -# col_name data_type comment - -emp_id int employee id -emp_country string two char iso code -emp_state string free text - -# Partition Information -# col_name data_type comment - -emp_country string two char iso code -emp_state string free text +emp_id int employee id +emp_country string two char iso code +emp_state string free text #### A masked pattern was here #### PREHOOK: query: show table extended like exim_employee diff --git ql/src/test/results/clientpositive/exim_19_00_part_external_location.q.out ql/src/test/results/clientpositive/exim_19_00_part_external_location.q.out index 66af54e..fc59807 100644 --- ql/src/test/results/clientpositive/exim_19_00_part_external_location.q.out +++ ql/src/test/results/clientpositive/exim_19_00_part_external_location.q.out @@ -72,17 +72,9 @@ PREHOOK: query: describe extended exim_employee PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_employee POSTHOOK: type: DESCTABLE -# col_name data_type comment - -emp_id int employee id -emp_country string two char iso code -emp_state string free text - -# Partition Information -# col_name data_type comment - -emp_country string two char iso code -emp_state string free text +emp_id int employee id +emp_country string two char iso code +emp_state string free text #### A masked pattern was here #### PREHOOK: query: show table extended like exim_employee diff --git ql/src/test/results/clientpositive/exim_19_part_external_location.q.out ql/src/test/results/clientpositive/exim_19_part_external_location.q.out index fe674f6..86a3ea6 100644 --- ql/src/test/results/clientpositive/exim_19_part_external_location.q.out +++ ql/src/test/results/clientpositive/exim_19_part_external_location.q.out @@ -93,17 +93,9 @@ PREHOOK: query: describe extended exim_employee PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_employee POSTHOOK: type: DESCTABLE -# col_name data_type comment - -emp_id int employee id -emp_country string two char iso code -emp_state string free text - -# Partition Information -# col_name data_type comment - -emp_country string two char iso code -emp_state string free text +emp_id int employee id +emp_country string two char iso code +emp_state string free text #### A masked pattern was here #### PREHOOK: query: show table extended like exim_employee diff --git ql/src/test/results/clientpositive/exim_20_part_managed_location.q.out ql/src/test/results/clientpositive/exim_20_part_managed_location.q.out index 1e1866f..b1ba7e0 100644 --- ql/src/test/results/clientpositive/exim_20_part_managed_location.q.out +++ ql/src/test/results/clientpositive/exim_20_part_managed_location.q.out @@ -93,17 +93,9 @@ PREHOOK: query: describe extended exim_employee PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended exim_employee POSTHOOK: type: DESCTABLE -# col_name data_type comment - -emp_id int employee id -emp_country string two char iso code -emp_state string free text - -# Partition Information -# col_name data_type comment - -emp_country string two char iso code -emp_state string free text +emp_id int employee id +emp_country string two char iso code +emp_state string free text #### A masked pattern was here #### PREHOOK: query: show table extended like exim_employee diff --git ql/src/test/results/clientpositive/fileformat_sequencefile.q.out ql/src/test/results/clientpositive/fileformat_sequencefile.q.out index cc86db4..6edd838 100644 --- ql/src/test/results/clientpositive/fileformat_sequencefile.q.out +++ ql/src/test/results/clientpositive/fileformat_sequencefile.q.out @@ -40,10 +40,8 @@ PREHOOK: query: DESCRIBE EXTENDED dest1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED dest1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None +key int +value string #### A masked pattern was here #### PREHOOK: query: FROM src diff --git ql/src/test/results/clientpositive/fileformat_text.q.out ql/src/test/results/clientpositive/fileformat_text.q.out index 0cafd62..6d8a2d5 100644 --- ql/src/test/results/clientpositive/fileformat_text.q.out +++ ql/src/test/results/clientpositive/fileformat_text.q.out @@ -40,10 +40,8 @@ PREHOOK: query: DESCRIBE EXTENDED dest1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED dest1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None +key int +value string #### A masked pattern was here #### PREHOOK: query: FROM src diff --git ql/src/test/results/clientpositive/index_creation.q.out ql/src/test/results/clientpositive/index_creation.q.out index ad2d121..f7106a6 100644 --- ql/src/test/results/clientpositive/index_creation.q.out +++ ql/src/test/results/clientpositive/index_creation.q.out @@ -43,11 +43,9 @@ PREHOOK: query: desc extended default__src_src_index_2__ PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended default__src_src_index_2__ POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -_bucketname string -_offsets array +key string default +_bucketname string +_offsets array #### A masked pattern was here #### PREHOOK: query: create index src_index_3 on table src(key) as 'compact' WITH DEFERRED REBUILD in table src_idx_src_index_3 @@ -59,11 +57,9 @@ PREHOOK: query: desc extended src_idx_src_index_3 PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended src_idx_src_index_3 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -_bucketname string -_offsets array +key string default +_bucketname string +_offsets array #### A masked pattern was here #### PREHOOK: query: create index src_index_4 on table src(key) as 'compact' WITH DEFERRED REBUILD ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t' STORED AS TEXTFILE @@ -75,11 +71,9 @@ PREHOOK: query: desc extended default__src_src_index_4__ PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended default__src_src_index_4__ POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -_bucketname string -_offsets array +key string default +_bucketname string +_offsets array #### A masked pattern was here #### PREHOOK: query: create index src_index_5 on table src(key) as 'compact' WITH DEFERRED REBUILD ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t' ESCAPED BY '\\' @@ -91,11 +85,9 @@ PREHOOK: query: desc extended default__src_src_index_5__ PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended default__src_src_index_5__ POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -_bucketname string -_offsets array +key string default +_bucketname string +_offsets array #### A masked pattern was here #### PREHOOK: query: create index src_index_6 on table src(key) as 'compact' WITH DEFERRED REBUILD STORED AS RCFILE @@ -107,11 +99,9 @@ PREHOOK: query: desc extended default__src_src_index_6__ PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended default__src_src_index_6__ POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -_bucketname string -_offsets array +key string default +_bucketname string +_offsets array #### A masked pattern was here #### PREHOOK: query: create index src_index_7 on table src(key) as 'compact' WITH DEFERRED REBUILD in table src_idx_src_index_7 STORED AS RCFILE @@ -123,11 +113,9 @@ PREHOOK: query: desc extended src_idx_src_index_7 PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended src_idx_src_index_7 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -_bucketname string -_offsets array +key string default +_bucketname string +_offsets array #### A masked pattern was here #### PREHOOK: query: create index src_index_8 on table src(key) as 'compact' WITH DEFERRED REBUILD IDXPROPERTIES ("prop1"="val1", "prop2"="val2") @@ -139,11 +127,9 @@ PREHOOK: query: desc extended default__src_src_index_8__ PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended default__src_src_index_8__ POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -_bucketname string -_offsets array +key string default +_bucketname string +_offsets array #### A masked pattern was here #### PREHOOK: query: create index src_index_9 on table src(key) as 'compact' WITH DEFERRED REBUILD TBLPROPERTIES ("prop1"="val1", "prop2"="val2") @@ -155,11 +141,9 @@ PREHOOK: query: desc extended default__src_src_index_9__ PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended default__src_src_index_9__ POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -_bucketname string -_offsets array +key string default +_bucketname string +_offsets array #### A masked pattern was here #### PREHOOK: query: create table `_t`(`_i` int, `_j` int) diff --git ql/src/test/results/clientpositive/inoutdriver.q.out ql/src/test/results/clientpositive/inoutdriver.q.out index d3a8f98..11a7e50 100644 --- ql/src/test/results/clientpositive/inoutdriver.q.out +++ ql/src/test/results/clientpositive/inoutdriver.q.out @@ -7,8 +7,6 @@ PREHOOK: query: desc extended test PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None +a int #### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/input1.q.out ql/src/test/results/clientpositive/input1.q.out index 2a11967..cc935d9 100644 --- ql/src/test/results/clientpositive/input1.q.out +++ ql/src/test/results/clientpositive/input1.q.out @@ -31,7 +31,5 @@ PREHOOK: query: DESCRIBE TEST1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE TEST1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b double None +a int +b double diff --git ql/src/test/results/clientpositive/input10.q.out ql/src/test/results/clientpositive/input10.q.out index 436119a..e22a2f4 100644 --- ql/src/test/results/clientpositive/input10.q.out +++ ql/src/test/results/clientpositive/input10.q.out @@ -31,15 +31,7 @@ PREHOOK: query: DESCRIBE TEST10 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE TEST10 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key int +value string +ds string +hr string diff --git ql/src/test/results/clientpositive/input15.q.out ql/src/test/results/clientpositive/input15.q.out index 0564184..2967fdb 100644 --- ql/src/test/results/clientpositive/input15.q.out +++ ql/src/test/results/clientpositive/input15.q.out @@ -33,7 +33,5 @@ PREHOOK: query: DESCRIBE TEST15 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE TEST15 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None +key int +value string diff --git ql/src/test/results/clientpositive/input2.q.out ql/src/test/results/clientpositive/input2.q.out index ac51d21..69c898b 100644 --- ql/src/test/results/clientpositive/input2.q.out +++ ql/src/test/results/clientpositive/input2.q.out @@ -7,18 +7,14 @@ PREHOOK: query: DESCRIBE TEST2a PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE TEST2a POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b double None +a int +b double PREHOOK: query: DESC TEST2a PREHOOK: type: DESCTABLE POSTHOOK: query: DESC TEST2a POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b double None +a int +b double PREHOOK: query: CREATE TABLE TEST2b(A ARRAY, B DOUBLE, C MAP) STORED AS TEXTFILE PREHOOK: type: CREATETABLE POSTHOOK: query: CREATE TABLE TEST2b(A ARRAY, B DOUBLE, C MAP) STORED AS TEXTFILE @@ -28,11 +24,9 @@ PREHOOK: query: DESCRIBE TEST2b PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE TEST2b POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a array None -b double None -c map None +a array +b double +c map PREHOOK: query: SHOW TABLES PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES diff --git ql/src/test/results/clientpositive/input3.q.out ql/src/test/results/clientpositive/input3.q.out index 79cf771..40a126c 100644 --- ql/src/test/results/clientpositive/input3.q.out +++ ql/src/test/results/clientpositive/input3.q.out @@ -7,10 +7,8 @@ PREHOOK: query: DESCRIBE TEST3a PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE TEST3a POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b double None +a int +b double PREHOOK: query: CREATE TABLE TEST3b(A ARRAY, B DOUBLE, C MAP) STORED AS TEXTFILE PREHOOK: type: CREATETABLE POSTHOOK: query: CREATE TABLE TEST3b(A ARRAY, B DOUBLE, C MAP) STORED AS TEXTFILE @@ -20,11 +18,9 @@ PREHOOK: query: DESCRIBE TEST3b PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE TEST3b POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a array None -b double None -c map None +a array +b double +c map PREHOOK: query: SHOW TABLES PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES @@ -72,12 +68,10 @@ PREHOOK: query: DESCRIBE TEST3b PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE TEST3b POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a array None -b double None -c map None -x double None +a array +b double +c map +x double PREHOOK: query: EXPLAIN ALTER TABLE TEST3b RENAME TO TEST3c PREHOOK: type: ALTERTABLE_RENAME @@ -112,12 +106,10 @@ PREHOOK: query: DESCRIBE TEST3c PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE TEST3c POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a array None -b double None -c map None -x double None +a array +b double +c map +x double PREHOOK: query: SHOW TABLES PREHOOK: type: SHOWTABLES POSTHOOK: query: SHOW TABLES @@ -165,9 +157,7 @@ PREHOOK: query: DESCRIBE EXTENDED TEST3c PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED TEST3c POSTHOOK: type: DESCTABLE -# col_name data_type comment - -r1 int None -r2 double None +r1 int +r2 double #### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/input_part10.q.out ql/src/test/results/clientpositive/input_part10.q.out index 377f503..73c689b 100644 --- ql/src/test/results/clientpositive/input_part10.q.out +++ ql/src/test/results/clientpositive/input_part10.q.out @@ -107,18 +107,10 @@ POSTHOOK: query: DESCRIBE EXTENDED part_special PARTITION(ds='2008 04 08', ts = POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: part_special PARTITION(ds=2008 04 08,ts=10:11:12=455).a SIMPLE [] POSTHOOK: Lineage: part_special PARTITION(ds=2008 04 08,ts=10:11:12=455).b SIMPLE [] -# col_name data_type comment - -a string None -b string None -ds string None -ts string None - -# Partition Information -# col_name data_type comment - -ds string None -ts string None +a string +b string +ds string +ts string #### A masked pattern was here #### PREHOOK: query: SELECT * FROM part_special WHERE ds='2008 04 08' AND ts = '10:11:12=455' diff --git ql/src/test/results/clientpositive/inputddl2.q.out ql/src/test/results/clientpositive/inputddl2.q.out index 200c600..3521dcf 100644 --- ql/src/test/results/clientpositive/inputddl2.q.out +++ ql/src/test/results/clientpositive/inputddl2.q.out @@ -33,15 +33,7 @@ PREHOOK: query: DESCRIBE INPUTDDL2 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE INPUTDDL2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None -ds string None -country string None - -# Partition Information -# col_name data_type comment - -ds string None -country string None +key int +value string +ds string +country string diff --git ql/src/test/results/clientpositive/inputddl3.q.out ql/src/test/results/clientpositive/inputddl3.q.out index 4b47a46..99e71e3 100644 --- ql/src/test/results/clientpositive/inputddl3.q.out +++ ql/src/test/results/clientpositive/inputddl3.q.out @@ -33,7 +33,5 @@ PREHOOK: query: DESCRIBE INPUTDDL3 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE INPUTDDL3 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key int None -value string None +key int +value string diff --git ql/src/test/results/clientpositive/inputddl4.q.out ql/src/test/results/clientpositive/inputddl4.q.out index 3785444..7b01e8b 100644 --- ql/src/test/results/clientpositive/inputddl4.q.out +++ ql/src/test/results/clientpositive/inputddl4.q.out @@ -23,43 +23,27 @@ PREHOOK: query: DESCRIBE INPUTDDL4 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE INPUTDDL4 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -viewtime string None -userid int None -page_url string None -referrer_url string None -friends array None -properties map None -ip string IP Address of the User -ds string None -country string None - -# Partition Information -# col_name data_type comment - -ds string None -country string None +viewtime string +userid int +page_url string +referrer_url string +friends array +properties map +ip string IP Address of the User +ds string +country string PREHOOK: query: DESCRIBE EXTENDED INPUTDDL4 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED INPUTDDL4 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -viewtime string None -userid int None -page_url string None -referrer_url string None -friends array None -properties map None -ip string IP Address of the User -ds string None -country string None - -# Partition Information -# col_name data_type comment - -ds string None -country string None +viewtime string +userid int +page_url string +referrer_url string +friends array +properties map +ip string IP Address of the User +ds string +country string #### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/inputddl5.q.out ql/src/test/results/clientpositive/inputddl5.q.out index 99a7bfe..ac53b58 100644 --- ql/src/test/results/clientpositive/inputddl5.q.out +++ ql/src/test/results/clientpositive/inputddl5.q.out @@ -17,9 +17,7 @@ PREHOOK: query: DESCRIBE INPUTDDL5 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE INPUTDDL5 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -name string None +name string PREHOOK: query: SELECT INPUTDDL5.name from INPUTDDL5 PREHOOK: type: QUERY PREHOOK: Input: default@inputddl5 diff --git ql/src/test/results/clientpositive/inputddl6.q.out ql/src/test/results/clientpositive/inputddl6.q.out index 52b8f0a..68fa4ad 100644 --- ql/src/test/results/clientpositive/inputddl6.q.out +++ ql/src/test/results/clientpositive/inputddl6.q.out @@ -27,32 +27,18 @@ PREHOOK: query: DESCRIBE EXTENDED INPUTDDL6 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED INPUTDDL6 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key string +value string +ds string #### A masked pattern was here #### PREHOOK: query: DESCRIBE EXTENDED INPUTDDL6 PARTITION (ds='2008-04-08') PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED INPUTDDL6 PARTITION (ds='2008-04-08') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key string +value string +ds string #### A masked pattern was here #### PREHOOK: query: SHOW PARTITIONS INPUTDDL6 diff --git ql/src/test/results/clientpositive/inputddl7.q.out ql/src/test/results/clientpositive/inputddl7.q.out index d8e3215..c219644 100644 --- ql/src/test/results/clientpositive/inputddl7.q.out +++ ql/src/test/results/clientpositive/inputddl7.q.out @@ -96,47 +96,29 @@ PREHOOK: query: DESCRIBE EXTENDED T1 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED T1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -name string None +name string #### A masked pattern was here #### PREHOOK: query: DESCRIBE EXTENDED T2 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED T2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -name string None +name string #### A masked pattern was here #### PREHOOK: query: DESCRIBE EXTENDED T3 PARTITION (ds='2008-04-09') PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED T3 PARTITION (ds='2008-04-09') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -name string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +name string +ds string #### A masked pattern was here #### PREHOOK: query: DESCRIBE EXTENDED T4 PARTITION (ds='2008-04-09') PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED T4 PARTITION (ds='2008-04-09') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -name string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +name string +ds string #### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/inputddl8.q.out ql/src/test/results/clientpositive/inputddl8.q.out index 9fc52de..cff41f5 100644 --- ql/src/test/results/clientpositive/inputddl8.q.out +++ ql/src/test/results/clientpositive/inputddl8.q.out @@ -19,21 +19,13 @@ PREHOOK: query: DESCRIBE EXTENDED INPUTDDL8 PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED INPUTDDL8 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -aint int from deserializer -astring string from deserializer -lint array from deserializer -lstring array from deserializer -lintstring array> from deserializer -mstringstring map from deserializer -ds string None -country string None - -# Partition Information -# col_name data_type comment - -ds string None -country string None +aint int from deserializer +astring string from deserializer +lint array from deserializer +lstring array from deserializer +lintstring array> from deserializer +mstringstring map from deserializer +ds string +country string #### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/join_thrift.q.out ql/src/test/results/clientpositive/join_thrift.q.out index 2ed9f61..da05fc4 100644 --- ql/src/test/results/clientpositive/join_thrift.q.out +++ ql/src/test/results/clientpositive/join_thrift.q.out @@ -2,14 +2,12 @@ PREHOOK: query: DESCRIBE src_thrift PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE src_thrift POSTHOOK: type: DESCTABLE -# col_name data_type comment - -aint int from deserializer -astring string from deserializer -lint array from deserializer -lstring array from deserializer -lintstring array> from deserializer -mstringstring map from deserializer +aint int from deserializer +astring string from deserializer +lint array from deserializer +lstring array from deserializer +lintstring array> from deserializer +mstringstring map from deserializer PREHOOK: query: EXPLAIN SELECT s1.aint, s2.lintstring FROM src_thrift s1 diff --git ql/src/test/results/clientpositive/load_dyn_part1.q.out ql/src/test/results/clientpositive/load_dyn_part1.q.out index f49e436..e93f0c9 100644 --- ql/src/test/results/clientpositive/load_dyn_part1.q.out +++ ql/src/test/results/clientpositive/load_dyn_part1.q.out @@ -20,18 +20,10 @@ PREHOOK: query: describe extended nzhang_part1 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended nzhang_part1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: explain diff --git ql/src/test/results/clientpositive/load_dyn_part10.q.out ql/src/test/results/clientpositive/load_dyn_part10.q.out index b4b1121..254dce4 100644 --- ql/src/test/results/clientpositive/load_dyn_part10.q.out +++ ql/src/test/results/clientpositive/load_dyn_part10.q.out @@ -15,18 +15,10 @@ PREHOOK: query: describe extended nzhang_part10 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended nzhang_part10 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: explain diff --git ql/src/test/results/clientpositive/load_dyn_part11.q.out ql/src/test/results/clientpositive/load_dyn_part11.q.out index 7c229f5..50913c7 100644 --- ql/src/test/results/clientpositive/load_dyn_part11.q.out +++ ql/src/test/results/clientpositive/load_dyn_part11.q.out @@ -15,18 +15,10 @@ PREHOOK: query: describe extended nzhang_part PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended nzhang_part POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: insert overwrite table nzhang_part partition (ds="2010-03-03", hr) select key, value, hr from srcpart where ds is not null and hr is not null diff --git ql/src/test/results/clientpositive/load_dyn_part12.q.out ql/src/test/results/clientpositive/load_dyn_part12.q.out index 0db3ae2..e763cc6 100644 --- ql/src/test/results/clientpositive/load_dyn_part12.q.out +++ ql/src/test/results/clientpositive/load_dyn_part12.q.out @@ -15,18 +15,10 @@ PREHOOK: query: describe extended nzhang_part12 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended nzhang_part12 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: insert overwrite table nzhang_part12 partition (ds="2010-03-03", hr) select key, value, cast(hr*2 as int) from srcpart where ds is not null and hr is not null diff --git ql/src/test/results/clientpositive/load_dyn_part13.q.out ql/src/test/results/clientpositive/load_dyn_part13.q.out index 69672b3..8824ecf 100644 --- ql/src/test/results/clientpositive/load_dyn_part13.q.out +++ ql/src/test/results/clientpositive/load_dyn_part13.q.out @@ -15,18 +15,10 @@ PREHOOK: query: describe extended nzhang_part13 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended nzhang_part13 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: explain diff --git ql/src/test/results/clientpositive/load_dyn_part14.q.out ql/src/test/results/clientpositive/load_dyn_part14.q.out index 23d8010..8a5f5da 100644 --- ql/src/test/results/clientpositive/load_dyn_part14.q.out +++ ql/src/test/results/clientpositive/load_dyn_part14.q.out @@ -15,15 +15,8 @@ PREHOOK: query: describe extended nzhang_part14 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended nzhang_part14 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string None -value string None - -# Partition Information -# col_name data_type comment - -value string None +key string +value string #### A masked pattern was here #### PREHOOK: query: explain diff --git ql/src/test/results/clientpositive/load_dyn_part2.q.out ql/src/test/results/clientpositive/load_dyn_part2.q.out index d6a8905..8d3ad15 100644 --- ql/src/test/results/clientpositive/load_dyn_part2.q.out +++ ql/src/test/results/clientpositive/load_dyn_part2.q.out @@ -11,18 +11,10 @@ PREHOOK: query: describe extended nzhang_part_bucket PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended nzhang_part_bucket POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string None -value string None -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string +value string +ds string +hr string #### A masked pattern was here #### PREHOOK: query: explain diff --git ql/src/test/results/clientpositive/load_dyn_part3.q.out ql/src/test/results/clientpositive/load_dyn_part3.q.out index 323e939..77d3d8f 100644 --- ql/src/test/results/clientpositive/load_dyn_part3.q.out +++ ql/src/test/results/clientpositive/load_dyn_part3.q.out @@ -15,18 +15,10 @@ PREHOOK: query: describe extended nzhang_part3 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended nzhang_part3 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: explain diff --git ql/src/test/results/clientpositive/load_dyn_part4.q.out ql/src/test/results/clientpositive/load_dyn_part4.q.out index d76596e..11f0120 100644 --- ql/src/test/results/clientpositive/load_dyn_part4.q.out +++ ql/src/test/results/clientpositive/load_dyn_part4.q.out @@ -15,18 +15,10 @@ PREHOOK: query: describe extended nzhang_part4 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended nzhang_part4 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: insert overwrite table nzhang_part4 partition (ds='2008-04-08', hr='existing_value') select key, value from src diff --git ql/src/test/results/clientpositive/load_dyn_part5.q.out ql/src/test/results/clientpositive/load_dyn_part5.q.out index 5ef4a4f..8bdc266 100644 --- ql/src/test/results/clientpositive/load_dyn_part5.q.out +++ ql/src/test/results/clientpositive/load_dyn_part5.q.out @@ -7,15 +7,8 @@ PREHOOK: query: describe extended nzhang_part5 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended nzhang_part5 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string None -value string None - -# Partition Information -# col_name data_type comment - -value string None +key string +value string #### A masked pattern was here #### PREHOOK: query: explain diff --git ql/src/test/results/clientpositive/load_dyn_part6.q.out ql/src/test/results/clientpositive/load_dyn_part6.q.out index ae59e49..4fe23b2 100644 --- ql/src/test/results/clientpositive/load_dyn_part6.q.out +++ ql/src/test/results/clientpositive/load_dyn_part6.q.out @@ -15,18 +15,10 @@ PREHOOK: query: describe extended nzhang_part6 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended nzhang_part6 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: insert overwrite table nzhang_part6 partition (ds="2010-03-03", hr) select key, value, hr from srcpart where ds is not null and hr is not null diff --git ql/src/test/results/clientpositive/load_dyn_part7.q.out ql/src/test/results/clientpositive/load_dyn_part7.q.out index fcc2029..9d48f40 100644 --- ql/src/test/results/clientpositive/load_dyn_part7.q.out +++ ql/src/test/results/clientpositive/load_dyn_part7.q.out @@ -15,18 +15,10 @@ PREHOOK: query: describe extended nzhang_part7 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended nzhang_part7 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: insert overwrite table nzhang_part7 partition (ds='2010-03-03', hr='12') select key, value from srcpart where ds = '2008-04-08' and hr = '12' diff --git ql/src/test/results/clientpositive/load_dyn_part8.q.out ql/src/test/results/clientpositive/load_dyn_part8.q.out index db309df..a86966f 100644 --- ql/src/test/results/clientpositive/load_dyn_part8.q.out +++ ql/src/test/results/clientpositive/load_dyn_part8.q.out @@ -15,18 +15,10 @@ PREHOOK: query: describe extended nzhang_part8 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended nzhang_part8 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: explain extended diff --git ql/src/test/results/clientpositive/load_dyn_part9.q.out ql/src/test/results/clientpositive/load_dyn_part9.q.out index 2ad6af3..003eb83 100644 --- ql/src/test/results/clientpositive/load_dyn_part9.q.out +++ ql/src/test/results/clientpositive/load_dyn_part9.q.out @@ -15,18 +15,10 @@ PREHOOK: query: describe extended nzhang_part9 PREHOOK: type: DESCTABLE POSTHOOK: query: describe extended nzhang_part9 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: explain diff --git ql/src/test/results/clientpositive/load_fs.q.out ql/src/test/results/clientpositive/load_fs.q.out index 7d85e4b..7f3d23b 100644 --- ql/src/test/results/clientpositive/load_fs.q.out +++ ql/src/test/results/clientpositive/load_fs.q.out @@ -47,10 +47,8 @@ PREHOOK: query: desc extended load_overwrite PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended load_overwrite POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string None -value string None +key string +value string #### A masked pattern was here #### PREHOOK: query: select count(*) from load_overwrite @@ -89,10 +87,8 @@ PREHOOK: query: desc extended load_overwrite2 PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended load_overwrite2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string None -value string None +key string +value string #### A masked pattern was here #### PREHOOK: query: select count(*) from load_overwrite2 diff --git ql/src/test/results/clientpositive/multi_sahooks.q.out ql/src/test/results/clientpositive/multi_sahooks.q.out index 3e2cd39..8e01ab5 100644 --- ql/src/test/results/clientpositive/multi_sahooks.q.out +++ ql/src/test/results/clientpositive/multi_sahooks.q.out @@ -13,9 +13,7 @@ PREHOOK: query: desc extended tbl_sahook PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl_sahook POSTHOOK: type: DESCTABLE -# col_name data_type comment - -c string None +c string #### A masked pattern was here #### PREHOOK: query: drop table tbl_sahook @@ -41,9 +39,7 @@ PREHOOK: query: desc extended tbl_sahooks PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl_sahooks POSTHOOK: type: DESCTABLE -# col_name data_type comment - -c string None +c string #### A masked pattern was here #### PREHOOK: query: drop table tbl_sahooks @@ -69,9 +65,7 @@ PREHOOK: query: desc extended tbl_sahooks PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl_sahooks POSTHOOK: type: DESCTABLE -# col_name data_type comment - -c string None +c string #### A masked pattern was here #### PREHOOK: query: drop table tbl_sahooks @@ -99,9 +93,7 @@ PREHOOK: query: desc extended tbl_sahooks PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl_sahooks POSTHOOK: type: DESCTABLE -# col_name data_type comment - -c string None +c string #### A masked pattern was here #### PREHOOK: query: drop table tbl_sahooks diff --git ql/src/test/results/clientpositive/partition_schema1.q.out ql/src/test/results/clientpositive/partition_schema1.q.out index 03f366b..f76bb83 100644 --- ql/src/test/results/clientpositive/partition_schema1.q.out +++ ql/src/test/results/clientpositive/partition_schema1.q.out @@ -19,16 +19,9 @@ POSTHOOK: query: desc partition_schema1 partition(dt='100') POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: partition_schema1 PARTITION(dt=100).key SIMPLE [(src1)src1.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: partition_schema1 PARTITION(dt=100).value SIMPLE [(src1)src1.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None -dt string None - -# Partition Information -# col_name data_type comment - -dt string None +key string +value string +dt string PREHOOK: query: alter table partition_schema1 add columns (x string) PREHOOK: type: ALTERTABLE_ADDCOLS PREHOOK: Input: default@partition_schema1 @@ -45,30 +38,16 @@ POSTHOOK: query: desc partition_schema1 POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: partition_schema1 PARTITION(dt=100).key SIMPLE [(src1)src1.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: partition_schema1 PARTITION(dt=100).value SIMPLE [(src1)src1.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None -x string None -dt string None - -# Partition Information -# col_name data_type comment - -dt string None +key string +value string +x string +dt string PREHOOK: query: desc partition_schema1 partition (dt='100') PREHOOK: type: DESCTABLE POSTHOOK: query: desc partition_schema1 partition (dt='100') POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: partition_schema1 PARTITION(dt=100).key SIMPLE [(src1)src1.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: partition_schema1 PARTITION(dt=100).value SIMPLE [(src1)src1.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None -dt string None - -# Partition Information -# col_name data_type comment - -dt string None +key string +value string +dt string diff --git ql/src/test/results/clientpositive/protectmode.q.out ql/src/test/results/clientpositive/protectmode.q.out index 3c7769f..c67e06c 100644 --- ql/src/test/results/clientpositive/protectmode.q.out +++ ql/src/test/results/clientpositive/protectmode.q.out @@ -43,9 +43,7 @@ PREHOOK: query: desc extended tbl1 PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col string None +col string #### A masked pattern was here #### PREHOOK: query: alter table tbl1 disable offline @@ -60,9 +58,7 @@ PREHOOK: query: desc extended tbl1 PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col string None +col string #### A masked pattern was here #### PREHOOK: query: select * from tbl1 @@ -142,15 +138,8 @@ PREHOOK: query: desc extended tbl2 partition (p='p1') PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl2 partition (p='p1') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col string None -p string None - -# Partition Information -# col_name data_type comment - -p string None +col string +p string #### A masked pattern was here #### PREHOOK: query: alter table tbl2 enable offline @@ -165,15 +154,8 @@ PREHOOK: query: desc extended tbl2 PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col string None -p string None - -# Partition Information -# col_name data_type comment - -p string None +col string +p string #### A masked pattern was here #### PREHOOK: query: alter table tbl2 enable no_drop @@ -188,15 +170,8 @@ PREHOOK: query: desc extended tbl2 PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col string None -p string None - -# Partition Information -# col_name data_type comment - -p string None +col string +p string #### A masked pattern was here #### PREHOOK: query: alter table tbl2 drop partition (p='p3') @@ -219,15 +194,8 @@ PREHOOK: query: desc extended tbl2 PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col string None -p string None - -# Partition Information -# col_name data_type comment - -p string None +col string +p string #### A masked pattern was here #### PREHOOK: query: alter table tbl2 disable no_drop @@ -242,15 +210,8 @@ PREHOOK: query: desc extended tbl2 PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl2 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col string None -p string None - -# Partition Information -# col_name data_type comment - -p string None +col string +p string #### A masked pattern was here #### PREHOOK: query: select * from tbl2 where p='p2' @@ -284,15 +245,8 @@ PREHOOK: query: desc extended tbl2 partition (p='p1') PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl2 partition (p='p1') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col string None -p string None - -# Partition Information -# col_name data_type comment - -p string None +col string +p string #### A masked pattern was here #### PREHOOK: query: select * from tbl2 where p='p1' diff --git ql/src/test/results/clientpositive/protectmode2.q.out ql/src/test/results/clientpositive/protectmode2.q.out index 5b0b6db..471620a 100644 --- ql/src/test/results/clientpositive/protectmode2.q.out +++ ql/src/test/results/clientpositive/protectmode2.q.out @@ -19,9 +19,7 @@ PREHOOK: query: desc extended tbl1 PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col string None +col string #### A masked pattern was here #### PREHOOK: query: alter table tbl1 enable no_drop @@ -36,9 +34,7 @@ PREHOOK: query: desc extended tbl1 PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col string None +col string #### A masked pattern was here #### PREHOOK: query: alter table tbl1 disable no_drop cascade @@ -53,9 +49,7 @@ PREHOOK: query: desc extended tbl1 PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tbl1 POSTHOOK: type: DESCTABLE -# col_name data_type comment - -col string None +col string #### A masked pattern was here #### PREHOOK: query: alter table tbl1 disable no_drop diff --git ql/src/test/results/clientpositive/rcfile_bigdata.q.out ql/src/test/results/clientpositive/rcfile_bigdata.q.out index 08a62a0..932dfcc 100644 --- ql/src/test/results/clientpositive/rcfile_bigdata.q.out +++ ql/src/test/results/clientpositive/rcfile_bigdata.q.out @@ -31,10 +31,8 @@ POSTHOOK: query: describe columnTable_Bigdata POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: columntable_bigdata.key SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: columntable_bigdata.value SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None +key string +value string PREHOOK: query: select count(columnTable_Bigdata.key) from columnTable_Bigdata PREHOOK: type: QUERY PREHOOK: Input: default@columntable_bigdata diff --git ql/src/test/results/clientpositive/rcfile_columnar.q.out ql/src/test/results/clientpositive/rcfile_columnar.q.out index 7e67bb8..6f05ab3 100644 --- ql/src/test/results/clientpositive/rcfile_columnar.q.out +++ ql/src/test/results/clientpositive/rcfile_columnar.q.out @@ -31,10 +31,8 @@ POSTHOOK: query: describe columnTable POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: columntable.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: columntable.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None +key string +value string PREHOOK: query: SELECT columnTable.* FROM columnTable ORDER BY key ASC, value ASC PREHOOK: type: QUERY PREHOOK: Input: default@columntable diff --git ql/src/test/results/clientpositive/rcfile_default_format.q.out ql/src/test/results/clientpositive/rcfile_default_format.q.out index 78f06e8..45c52cb 100644 --- ql/src/test/results/clientpositive/rcfile_default_format.q.out +++ ql/src/test/results/clientpositive/rcfile_default_format.q.out @@ -7,9 +7,7 @@ PREHOOK: query: DESCRIBE EXTENDED rcfile_default_format PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED rcfile_default_format POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string None +key string #### A masked pattern was here #### PREHOOK: query: CREATE TABLE rcfile_default_format_ctas AS SELECT key,value FROM src @@ -23,10 +21,8 @@ PREHOOK: query: DESCRIBE EXTENDED rcfile_default_format_ctas PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED rcfile_default_format_ctas POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string None -value string None +key string +value string #### A masked pattern was here #### PREHOOK: query: CREATE TABLE rcfile_default_format_txtfile (key STRING) STORED AS TEXTFILE @@ -48,9 +44,7 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED rcfile_default_format_txtfile POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: rcfile_default_format_txtfile.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -key string None +key string #### A masked pattern was here #### PREHOOK: query: CREATE TABLE textfile_default_format_ctas AS SELECT key,value FROM rcfile_default_format_ctas @@ -66,9 +60,7 @@ PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE EXTENDED textfile_default_format_ctas POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: rcfile_default_format_txtfile.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None +key string +value string #### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/rename_column.q.out ql/src/test/results/clientpositive/rename_column.q.out index f77ba41..0289738 100644 --- ql/src/test/results/clientpositive/rename_column.q.out +++ ql/src/test/results/clientpositive/rename_column.q.out @@ -7,11 +7,9 @@ PREHOOK: query: DESCRIBE kv_rename_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE kv_rename_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None -c int None +a int +b int +c int PREHOOK: query: ALTER TABLE kv_rename_test CHANGE a a STRING PREHOOK: type: ALTERTABLE_RENAMECOL PREHOOK: Input: default@kv_rename_test @@ -24,11 +22,9 @@ PREHOOK: query: DESCRIBE kv_rename_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE kv_rename_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b int None -c int None +a string +b int +c int PREHOOK: query: ALTER TABLE kv_rename_test CHANGE a a1 INT PREHOOK: type: ALTERTABLE_RENAMECOL PREHOOK: Input: default@kv_rename_test @@ -41,11 +37,9 @@ PREHOOK: query: DESCRIBE kv_rename_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE kv_rename_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a1 int None -b int None -c int None +a1 int +b int +c int PREHOOK: query: ALTER TABLE kv_rename_test CHANGE a1 a2 INT FIRST PREHOOK: type: ALTERTABLE_RENAMECOL PREHOOK: Input: default@kv_rename_test @@ -58,11 +52,9 @@ PREHOOK: query: DESCRIBE kv_rename_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE kv_rename_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a2 int None -b int None -c int None +a2 int +b int +c int PREHOOK: query: ALTER TABLE kv_rename_test CHANGE a2 a INT AFTER b PREHOOK: type: ALTERTABLE_RENAMECOL PREHOOK: Input: default@kv_rename_test @@ -75,11 +67,9 @@ PREHOOK: query: DESCRIBE kv_rename_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE kv_rename_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -b int None -a int None -c int None +b int +a int +c int PREHOOK: query: ALTER TABLE kv_rename_test CHANGE a a1 INT COMMENT 'test comment1' PREHOOK: type: ALTERTABLE_RENAMECOL PREHOOK: Input: default@kv_rename_test @@ -92,11 +82,9 @@ PREHOOK: query: DESCRIBE kv_rename_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE kv_rename_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -b int None -a1 int test comment1 -c int None +b int +a1 int test comment1 +c int PREHOOK: query: ALTER TABLE kv_rename_test CHANGE a1 a2 INT COMMENT 'test comment2' FIRST PREHOOK: type: ALTERTABLE_RENAMECOL PREHOOK: Input: default@kv_rename_test @@ -109,11 +97,9 @@ PREHOOK: query: DESCRIBE kv_rename_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE kv_rename_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a2 int test comment2 -b int None -c int None +a2 int test comment2 +b int +c int PREHOOK: query: ALTER TABLE kv_rename_test CHANGE COLUMN a2 a INT AFTER b PREHOOK: type: ALTERTABLE_RENAMECOL PREHOOK: Input: default@kv_rename_test @@ -126,11 +112,9 @@ PREHOOK: query: DESCRIBE kv_rename_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE kv_rename_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -b int None -a int test comment2 -c int None +b int +a int test comment2 +c int PREHOOK: query: DROP TABLE kv_rename_test PREHOOK: type: DROPTABLE PREHOOK: Input: default@kv_rename_test @@ -170,11 +154,9 @@ PREHOOK: query: DESCRIBE kv_rename_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE kv_rename_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b int None -c int None +a int +b int +c int PREHOOK: query: ALTER TABLE kv_rename_test CHANGE a a STRING PREHOOK: type: ALTERTABLE_RENAMECOL PREHOOK: Input: kv_rename_test_db@kv_rename_test @@ -187,11 +169,9 @@ PREHOOK: query: DESCRIBE kv_rename_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE kv_rename_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a string None -b int None -c int None +a string +b int +c int PREHOOK: query: ALTER TABLE kv_rename_test CHANGE a a1 INT PREHOOK: type: ALTERTABLE_RENAMECOL PREHOOK: Input: kv_rename_test_db@kv_rename_test @@ -204,11 +184,9 @@ PREHOOK: query: DESCRIBE kv_rename_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE kv_rename_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a1 int None -b int None -c int None +a1 int +b int +c int PREHOOK: query: ALTER TABLE kv_rename_test CHANGE a1 a2 INT FIRST PREHOOK: type: ALTERTABLE_RENAMECOL PREHOOK: Input: kv_rename_test_db@kv_rename_test @@ -221,11 +199,9 @@ PREHOOK: query: DESCRIBE kv_rename_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE kv_rename_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a2 int None -b int None -c int None +a2 int +b int +c int PREHOOK: query: ALTER TABLE kv_rename_test CHANGE a2 a INT AFTER b PREHOOK: type: ALTERTABLE_RENAMECOL PREHOOK: Input: kv_rename_test_db@kv_rename_test @@ -238,11 +214,9 @@ PREHOOK: query: DESCRIBE kv_rename_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE kv_rename_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -b int None -a int None -c int None +b int +a int +c int PREHOOK: query: ALTER TABLE kv_rename_test CHANGE a a1 INT COMMENT 'test comment1' PREHOOK: type: ALTERTABLE_RENAMECOL PREHOOK: Input: kv_rename_test_db@kv_rename_test @@ -255,11 +229,9 @@ PREHOOK: query: DESCRIBE kv_rename_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE kv_rename_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -b int None -a1 int test comment1 -c int None +b int +a1 int test comment1 +c int PREHOOK: query: ALTER TABLE kv_rename_test CHANGE a1 a2 INT COMMENT 'test comment2' FIRST PREHOOK: type: ALTERTABLE_RENAMECOL PREHOOK: Input: kv_rename_test_db@kv_rename_test @@ -272,11 +244,9 @@ PREHOOK: query: DESCRIBE kv_rename_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE kv_rename_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a2 int test comment2 -b int None -c int None +a2 int test comment2 +b int +c int PREHOOK: query: ALTER TABLE kv_rename_test CHANGE COLUMN a2 a INT AFTER b PREHOOK: type: ALTERTABLE_RENAMECOL PREHOOK: Input: kv_rename_test_db@kv_rename_test @@ -289,11 +259,9 @@ PREHOOK: query: DESCRIBE kv_rename_test PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE kv_rename_test POSTHOOK: type: DESCTABLE -# col_name data_type comment - -b int None -a int test comment2 -c int None +b int +a int test comment2 +c int PREHOOK: query: DROP TABLE kv_rename_test PREHOOK: type: DROPTABLE PREHOOK: Input: kv_rename_test_db@kv_rename_test diff --git ql/src/test/results/clientpositive/serde_reported_schema.q.out ql/src/test/results/clientpositive/serde_reported_schema.q.out index 5dd81b4..6fac935 100644 --- ql/src/test/results/clientpositive/serde_reported_schema.q.out +++ ql/src/test/results/clientpositive/serde_reported_schema.q.out @@ -17,17 +17,10 @@ PREHOOK: query: describe int_string PREHOOK: type: DESCTABLE POSTHOOK: query: describe int_string POSTHOOK: type: DESCTABLE -# col_name data_type comment - -myint int from deserializer -mystring string from deserializer -underscore_int int from deserializer -b string None - -# Partition Information -# col_name data_type comment - -b string None +myint int from deserializer +mystring string from deserializer +underscore_int int from deserializer +b string PREHOOK: query: alter table int_string add partition (b='part1') PREHOOK: type: ALTERTABLE_ADDPARTS PREHOOK: Input: default@int_string @@ -39,14 +32,7 @@ PREHOOK: query: describe int_string partition (b='part1') PREHOOK: type: DESCTABLE POSTHOOK: query: describe int_string partition (b='part1') POSTHOOK: type: DESCTABLE -# col_name data_type comment - -myint int from deserializer -mystring string from deserializer -underscore_int int from deserializer -b string None - -# Partition Information -# col_name data_type comment - -b string None +myint int from deserializer +mystring string from deserializer +underscore_int int from deserializer +b string diff --git ql/src/test/results/clientpositive/show_columns.q.out ql/src/test/results/clientpositive/show_columns.q.out index 10bd3ce..3302741 100644 --- ql/src/test/results/clientpositive/show_columns.q.out +++ ql/src/test/results/clientpositive/show_columns.q.out @@ -32,11 +32,9 @@ PREHOOK: Input: default@shcol_test POSTHOOK: query: SHOW COLUMNS from shcol_test POSTHOOK: type: SHOWCOLUMNS POSTHOOK: Input: default@shcol_test -# col_name - -key -value -ds +key +value +ds PREHOOK: query: -- SHOW COLUMNS CREATE DATABASE test_db PREHOOK: type: CREATEDATABASE @@ -64,18 +62,14 @@ PREHOOK: Input: test_db@foo POSTHOOK: query: SHOW COLUMNS from foo POSTHOOK: type: SHOWCOLUMNS POSTHOOK: Input: test_db@foo -# col_name - -a +a PREHOOK: query: SHOW COLUMNS in foo PREHOOK: type: SHOWCOLUMNS PREHOOK: Input: test_db@foo POSTHOOK: query: SHOW COLUMNS in foo POSTHOOK: type: SHOWCOLUMNS POSTHOOK: Input: test_db@foo -# col_name - -a +a PREHOOK: query: -- SHOW COLUMNS from a database with a name that requires escaping CREATE DATABASE `database` PREHOOK: type: CREATEDATABASE @@ -97,9 +91,7 @@ PREHOOK: Input: database@foo POSTHOOK: query: SHOW COLUMNS from foo POSTHOOK: type: SHOWCOLUMNS POSTHOOK: Input: database@foo -# col_name - -a +a PREHOOK: query: use default PREHOOK: type: SWITCHDATABASE POSTHOOK: query: use default @@ -110,15 +102,11 @@ PREHOOK: Input: test_db@foo POSTHOOK: query: SHOW COLUMNS from test_db.foo POSTHOOK: type: SHOWCOLUMNS POSTHOOK: Input: test_db@foo -# col_name - -a +a PREHOOK: query: SHOW COLUMNS from foo from test_db PREHOOK: type: SHOWCOLUMNS PREHOOK: Input: test_db@foo POSTHOOK: query: SHOW COLUMNS from foo from test_db POSTHOOK: type: SHOWCOLUMNS POSTHOOK: Input: test_db@foo -# col_name - -a +a diff --git ql/src/test/results/clientpositive/split_sample.q.out ql/src/test/results/clientpositive/split_sample.q.out index bb00b90..ef641d0 100644 --- ql/src/test/results/clientpositive/split_sample.q.out +++ ql/src/test/results/clientpositive/split_sample.q.out @@ -107,10 +107,8 @@ POSTHOOK: Lineage: ss_i_part PARTITION(p=2).key EXPRESSION [(src)src.FieldSchema POSTHOOK: Lineage: ss_i_part PARTITION(p=2).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: ss_i_part PARTITION(p=3).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: ss_i_part PARTITION(p=3).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key int None -value string None +key int +value string PREHOOK: query: explain select key, value from ss_src2 tablesample(1 percent) limit 10 PREHOOK: type: QUERY POSTHOOK: query: explain select key, value from ss_src2 tablesample(1 percent) limit 10 diff --git ql/src/test/results/clientpositive/stats0.q.out ql/src/test/results/clientpositive/stats0.q.out index cfda06c..b934f0a 100644 --- ql/src/test/results/clientpositive/stats0.q.out +++ ql/src/test/results/clientpositive/stats0.q.out @@ -152,10 +152,8 @@ POSTHOOK: query: desc extended stats_non_partitioned POSTHOOK: type: DESCTABLE POSTHOOK: Lineage: stats_non_partitioned.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: stats_non_partitioned.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None +key string +value string #### A masked pattern was here #### PREHOOK: query: select * from stats_non_partitioned @@ -1275,16 +1273,9 @@ POSTHOOK: Lineage: stats_non_partitioned.key SIMPLE [(src)src.FieldSchema(name:k POSTHOOK: Lineage: stats_non_partitioned.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: stats_partitioned PARTITION(ds=1).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: stats_partitioned PARTITION(ds=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key string +value string +ds string #### A masked pattern was here #### PREHOOK: query: describe extended stats_partitioned @@ -1295,16 +1286,9 @@ POSTHOOK: Lineage: stats_non_partitioned.key SIMPLE [(src)src.FieldSchema(name:k POSTHOOK: Lineage: stats_non_partitioned.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: stats_partitioned PARTITION(ds=1).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: stats_partitioned PARTITION(ds=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key string +value string +ds string #### A masked pattern was here #### PREHOOK: query: drop table stats_non_partitioned @@ -1653,10 +1637,8 @@ POSTHOOK: Lineage: stats_non_partitioned.value SIMPLE [(src)src.FieldSchema(name POSTHOOK: Lineage: stats_non_partitioned.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: stats_partitioned PARTITION(ds=1).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: stats_partitioned PARTITION(ds=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None +key string +value string #### A masked pattern was here #### PREHOOK: query: select * from stats_non_partitioned @@ -2850,16 +2832,9 @@ POSTHOOK: Lineage: stats_partitioned PARTITION(ds=1).key SIMPLE [(src)src.FieldS POSTHOOK: Lineage: stats_partitioned PARTITION(ds=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: stats_partitioned PARTITION(ds=1).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: stats_partitioned PARTITION(ds=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key string +value string +ds string #### A masked pattern was here #### PREHOOK: query: describe extended stats_partitioned @@ -2874,15 +2849,8 @@ POSTHOOK: Lineage: stats_partitioned PARTITION(ds=1).key SIMPLE [(src)src.FieldS POSTHOOK: Lineage: stats_partitioned PARTITION(ds=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: stats_partitioned PARTITION(ds=1).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: stats_partitioned PARTITION(ds=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string None -value string None -ds string None - -# Partition Information -# col_name data_type comment - -ds string None +key string +value string +ds string #### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/tablename_with_select.q.out ql/src/test/results/clientpositive/tablename_with_select.q.out index 3a34ba1..8fd14ee 100644 --- ql/src/test/results/clientpositive/tablename_with_select.q.out +++ ql/src/test/results/clientpositive/tablename_with_select.q.out @@ -7,10 +7,8 @@ PREHOOK: query: DESCRIBE tmp_select PREHOOK: type: DESCTABLE POSTHOOK: query: DESCRIBE tmp_select POSTHOOK: type: DESCTABLE -# col_name data_type comment - -a int None -b string None +a int +b string PREHOOK: query: INSERT OVERWRITE TABLE tmp_select SELECT key, value FROM src PREHOOK: type: QUERY PREHOOK: Input: default@src diff --git ql/src/test/results/clientpositive/updateAccessTime.q.out ql/src/test/results/clientpositive/updateAccessTime.q.out index c1570c4..734ec93 100644 --- ql/src/test/results/clientpositive/updateAccessTime.q.out +++ ql/src/test/results/clientpositive/updateAccessTime.q.out @@ -13,10 +13,8 @@ PREHOOK: query: desc extended tstsrc PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tstsrc POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string None -value string None +key string +value string #### A masked pattern was here #### PREHOOK: query: select count(1) from tstsrc @@ -32,10 +30,8 @@ PREHOOK: query: desc extended tstsrc PREHOOK: type: DESCTABLE POSTHOOK: query: desc extended tstsrc POSTHOOK: type: DESCTABLE -# col_name data_type comment - -key string None -value string None +key string +value string #### A masked pattern was here #### PREHOOK: query: drop table tstsrc @@ -94,18 +90,10 @@ POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=11).key SIMPLE [(srcpar POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: desc extended tstsrcpart partition (ds='2008-04-08', hr='11') @@ -120,18 +108,10 @@ POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=11).key SIMPLE [(srcpar POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: desc extended tstsrcpart partition (ds='2008-04-08', hr='12') @@ -146,18 +126,10 @@ POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=11).key SIMPLE [(srcpar POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: select count(1) from tstsrcpart where ds = '2008-04-08' and hr = '11' @@ -191,18 +163,10 @@ POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=11).key SIMPLE [(srcpar POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: desc extended tstsrcpart partition (ds='2008-04-08', hr='11') @@ -217,18 +181,10 @@ POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=11).key SIMPLE [(srcpar POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: desc extended tstsrcpart partition (ds='2008-04-08', hr='12') @@ -243,18 +199,10 @@ POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=11).key SIMPLE [(srcpar POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-09,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default -ds string None -hr string None - -# Partition Information -# col_name data_type comment - -ds string None -hr string None +key string default +value string default +ds string +hr string #### A masked pattern was here #### PREHOOK: query: drop table tstsrcpart