diff --git data/files/alltypes.txt data/files/alltypes.txt index d86a7bc..358cf40 100644 --- data/files/alltypes.txt +++ data/files/alltypes.txt @@ -1,2 +1,2 @@ -true|10|100|1000|10000|4.0|20.0|2.2222|1969-12-31 15:59:58.174|1970-01-01 00:00:00|hello|k1:v1,k2:v2|100,200|{10, "foo"} -true|20|200|2000|20000|8.0|40.0|4.2222|1970-12-31 15:59:58.174|1971-01-01 00:00:00||k3:v3,k4:v4|200,300|{20, "bar"} +true|10|100|1000|10000|4.0|20.0|2.2222|1969-12-31 15:59:58.174|1970-01-01 00:00:00|hello|hello|k1:v1,k2:v2|100,200|{10, "foo"} +true|20|200|2000|20000|8.0|40.0|4.2222|1970-12-31 15:59:58.174|1971-01-01 00:00:00|||k3:v3,k4:v4|200,300|{20, "bar"} diff --git metastore/src/java/org/apache/hadoop/hive/metastore/ObjectStore.java metastore/src/java/org/apache/hadoop/hive/metastore/ObjectStore.java index 4f158bc..04d399f 100644 --- metastore/src/java/org/apache/hadoop/hive/metastore/ObjectStore.java +++ metastore/src/java/org/apache/hadoop/hive/metastore/ObjectStore.java @@ -4943,38 +4943,39 @@ private ColumnStatisticsObj getTableColumnStatisticsObj(MTableColumnStatistics m ColumnStatisticsObj statsObj = new ColumnStatisticsObj(); statsObj.setColType(mStatsObj.getColType()); statsObj.setColName(mStatsObj.getColName()); - String colType = mStatsObj.getColType(); + String colType = mStatsObj.getColType().toLowerCase(); ColumnStatisticsData colStatsData = new ColumnStatisticsData(); - if (colType.equalsIgnoreCase("boolean")) { + if (colType.equals("boolean")) { BooleanColumnStatsData boolStats = new BooleanColumnStatsData(); boolStats.setNumFalses(mStatsObj.getNumFalses()); boolStats.setNumTrues(mStatsObj.getNumTrues()); boolStats.setNumNulls(mStatsObj.getNumNulls()); colStatsData.setBooleanStats(boolStats); - } else if (colType.equalsIgnoreCase("string")) { + } else if (colType.equals("string") || + colType.startsWith("varchar") || colType.startsWith("char")) { StringColumnStatsData stringStats = new StringColumnStatsData(); stringStats.setNumNulls(mStatsObj.getNumNulls()); stringStats.setAvgColLen(mStatsObj.getAvgColLen()); stringStats.setMaxColLen(mStatsObj.getMaxColLen()); stringStats.setNumDVs(mStatsObj.getNumDVs()); colStatsData.setStringStats(stringStats); - } else if (colType.equalsIgnoreCase("binary")) { + } else if (colType.equals("binary")) { BinaryColumnStatsData binaryStats = new BinaryColumnStatsData(); binaryStats.setNumNulls(mStatsObj.getNumNulls()); binaryStats.setAvgColLen(mStatsObj.getAvgColLen()); binaryStats.setMaxColLen(mStatsObj.getMaxColLen()); colStatsData.setBinaryStats(binaryStats); - } else if (colType.equalsIgnoreCase("bigint") || colType.equalsIgnoreCase("int") || - colType.equalsIgnoreCase("smallint") || colType.equalsIgnoreCase("tinyint") || - colType.equalsIgnoreCase("timestamp")) { + } else if (colType.equals("bigint") || colType.equals("int") || + colType.equals("smallint") || colType.equals("tinyint") || + colType.equals("timestamp")) { LongColumnStatsData longStats = new LongColumnStatsData(); longStats.setNumNulls(mStatsObj.getNumNulls()); longStats.setHighValue(mStatsObj.getLongHighValue()); longStats.setLowValue(mStatsObj.getLongLowValue()); longStats.setNumDVs(mStatsObj.getNumDVs()); colStatsData.setLongStats(longStats); - } else if (colType.equalsIgnoreCase("double") || colType.equalsIgnoreCase("float")) { + } else if (colType.equals("double") || colType.equals("float")) { DoubleColumnStatsData doubleStats = new DoubleColumnStatsData(); doubleStats.setNumNulls(mStatsObj.getNumNulls()); doubleStats.setHighValue(mStatsObj.getDoubleHighValue()); @@ -5119,38 +5120,39 @@ private ColumnStatisticsObj getPartitionColumnStatisticsObj(MPartitionColumnStat ColumnStatisticsObj statsObj = new ColumnStatisticsObj(); statsObj.setColType(mStatsObj.getColType()); statsObj.setColName(mStatsObj.getColName()); - String colType = mStatsObj.getColType(); + String colType = mStatsObj.getColType().toLowerCase(); ColumnStatisticsData colStatsData = new ColumnStatisticsData(); - if (colType.equalsIgnoreCase("boolean")) { + if (colType.equals("boolean")) { BooleanColumnStatsData boolStats = new BooleanColumnStatsData(); boolStats.setNumFalses(mStatsObj.getNumFalses()); boolStats.setNumTrues(mStatsObj.getNumTrues()); boolStats.setNumNulls(mStatsObj.getNumNulls()); colStatsData.setBooleanStats(boolStats); - } else if (colType.equalsIgnoreCase("string")) { + } else if (colType.equals("string") || + colType.startsWith("varchar") || colType.startsWith("char")) { StringColumnStatsData stringStats = new StringColumnStatsData(); stringStats.setNumNulls(mStatsObj.getNumNulls()); stringStats.setAvgColLen(mStatsObj.getAvgColLen()); stringStats.setMaxColLen(mStatsObj.getMaxColLen()); stringStats.setNumDVs(mStatsObj.getNumDVs()); colStatsData.setStringStats(stringStats); - } else if (colType.equalsIgnoreCase("binary")) { + } else if (colType.equals("binary")) { BinaryColumnStatsData binaryStats = new BinaryColumnStatsData(); binaryStats.setNumNulls(mStatsObj.getNumNulls()); binaryStats.setAvgColLen(mStatsObj.getAvgColLen()); binaryStats.setMaxColLen(mStatsObj.getMaxColLen()); colStatsData.setBinaryStats(binaryStats); - } else if (colType.equalsIgnoreCase("tinyint") || colType.equalsIgnoreCase("smallint") || - colType.equalsIgnoreCase("int") || colType.equalsIgnoreCase("bigint") || - colType.equalsIgnoreCase("timestamp")) { + } else if (colType.equals("tinyint") || colType.equals("smallint") || + colType.equals("int") || colType.equals("bigint") || + colType.equals("timestamp")) { LongColumnStatsData longStats = new LongColumnStatsData(); longStats.setNumNulls(mStatsObj.getNumNulls()); longStats.setHighValue(mStatsObj.getLongHighValue()); longStats.setLowValue(mStatsObj.getLongLowValue()); longStats.setNumDVs(mStatsObj.getNumDVs()); colStatsData.setLongStats(longStats); - } else if (colType.equalsIgnoreCase("double") || colType.equalsIgnoreCase("float")) { + } else if (colType.equals("double") || colType.equals("float")) { DoubleColumnStatsData doubleStats = new DoubleColumnStatsData(); doubleStats.setNumNulls(mStatsObj.getNumNulls()); doubleStats.setHighValue(mStatsObj.getDoubleHighValue()); diff --git ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java index 4658049..6ffef30 100644 --- ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java +++ ql/src/java/org/apache/hadoop/hive/ql/io/orc/WriterImpl.java @@ -1830,6 +1830,8 @@ private long getRawDataSizeFromPrimitives(TreeWriter child, ObjectInspector oi) case DOUBLE: return numVals * JavaDataModel.get().primitive2(); case STRING: + case VARCHAR: + case CHAR: // ORC strings are converted to java Strings. so use JavaDataModel to // compute the overall size of strings child = (StringTreeWriter) child; diff --git ql/src/test/queries/clientpositive/annotate_stats_select.q ql/src/test/queries/clientpositive/annotate_stats_select.q index 93492f5..b94fc0c 100644 --- ql/src/test/queries/clientpositive/annotate_stats_select.q +++ ql/src/test/queries/clientpositive/annotate_stats_select.q @@ -10,6 +10,7 @@ create table if not exists alltypes ( ts1 timestamp, da1 timestamp, s1 string, + vc1 varchar(5), m1 map, l1 array, st1 struct @@ -28,7 +29,7 @@ insert overwrite table alltypes_orc select * from alltypes; explain extended select * from alltypes_orc; -- statistics for complex types are not supported yet -analyze table alltypes_orc compute statistics for columns bo1, ti1, si1, i1, bi1, f1, d1,s1; +analyze table alltypes_orc compute statistics for columns bo1, ti1, si1, i1, bi1, f1, d1, s1, vc1; -- numRows: 2 rawDataSize: 1514 explain extended select * from alltypes_orc; diff --git ql/src/test/results/clientpositive/annotate_stats_select.q.out ql/src/test/results/clientpositive/annotate_stats_select.q.out index 9cdfa74..66d11ea 100644 --- ql/src/test/results/clientpositive/annotate_stats_select.q.out +++ ql/src/test/results/clientpositive/annotate_stats_select.q.out @@ -10,6 +10,7 @@ PREHOOK: query: create table if not exists alltypes ( ts1 timestamp, da1 timestamp, s1 string, + vc1 varchar(5), m1 map, l1 array, st1 struct @@ -29,6 +30,7 @@ POSTHOOK: query: create table if not exists alltypes ( ts1 timestamp, da1 timestamp, s1 string, + vc1 varchar(5), m1 map, l1 array, st1 struct @@ -78,6 +80,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] PREHOOK: query: -- basicStatState: COMPLETE colStatState: NONE numRows: 2 rawDataSize: 1514 explain extended select * from alltypes_orc PREHOOK: type: QUERY @@ -98,6 +101,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)))) @@ -112,7 +116,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: NONE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: NONE GatherStats: false Select Operator expressions: @@ -138,25 +142,27 @@ STAGE PLANS: type: timestamp expr: s1 type: string + expr: vc1 + type: varchar(5) expr: m1 type: map expr: l1 type: array expr: st1 type: struct - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14 Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: NONE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: NONE ListSink PREHOOK: query: -- statistics for complex types are not supported yet -analyze table alltypes_orc compute statistics for columns bo1, ti1, si1, i1, bi1, f1, d1,s1 +analyze table alltypes_orc compute statistics for columns bo1, ti1, si1, i1, bi1, f1, d1, s1, vc1 PREHOOK: type: QUERY PREHOOK: Input: default@alltypes_orc #### A masked pattern was here #### POSTHOOK: query: -- statistics for complex types are not supported yet -analyze table alltypes_orc compute statistics for columns bo1, ti1, si1, i1, bi1, f1, d1,s1 +analyze table alltypes_orc compute statistics for columns bo1, ti1, si1, i1, bi1, f1, d1, s1, vc1 POSTHOOK: type: QUERY POSTHOOK: Input: default@alltypes_orc #### A masked pattern was here #### @@ -174,6 +180,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] PREHOOK: query: -- numRows: 2 rawDataSize: 1514 explain extended select * from alltypes_orc PREHOOK: type: QUERY @@ -194,6 +201,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)))) @@ -208,7 +216,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: PARTIAL + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: PARTIAL GatherStats: false Select Operator expressions: @@ -234,15 +242,17 @@ STAGE PLANS: type: timestamp expr: s1 type: string + expr: vc1 + type: varchar(5) expr: m1 type: map expr: l1 type: array expr: st1 type: struct - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14 Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: PARTIAL + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: PARTIAL ListSink @@ -266,6 +276,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL bo1))))) @@ -281,7 +292,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -324,19 +335,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -346,19 +357,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -393,6 +404,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL i1) int1)))) @@ -408,7 +420,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -451,19 +463,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -473,19 +485,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -518,6 +530,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL s1))))) @@ -533,7 +546,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -576,19 +589,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -598,19 +611,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -645,6 +658,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL m1))))) @@ -660,7 +674,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: NONE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: NONE GatherStats: false Select Operator expressions: @@ -668,14 +682,14 @@ STAGE PLANS: type: map outputColumnNames: _col0 Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: NONE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: NONE File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: NONE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: NONE #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.TextInputFormat @@ -703,19 +717,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -725,19 +739,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -770,6 +784,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL bo1)) (TOK_SELEXPR (TOK_TABLE_OR_COL ti1)) (TOK_SELEXPR (TOK_TABLE_OR_COL si1)) (TOK_SELEXPR (TOK_TABLE_OR_COL i1)) (TOK_SELEXPR (TOK_TABLE_OR_COL bi1)) (TOK_SELEXPR (TOK_TABLE_OR_COL f1)) (TOK_SELEXPR (TOK_TABLE_OR_COL d1)) (TOK_SELEXPR (TOK_TABLE_OR_COL s1))))) @@ -785,7 +800,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -842,19 +857,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -864,19 +879,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -909,6 +924,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_NULL)))) @@ -924,7 +940,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -967,19 +983,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -989,19 +1005,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -1034,6 +1050,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR 11)))) @@ -1049,7 +1066,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -1092,19 +1109,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -1114,19 +1131,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -1159,6 +1176,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR 11L)))) @@ -1174,7 +1192,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -1217,19 +1235,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -1239,19 +1257,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -1284,6 +1302,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR 11.0)))) @@ -1299,7 +1318,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -1342,19 +1361,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -1364,19 +1383,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -1409,6 +1428,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR "hello")))) @@ -1424,7 +1444,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -1467,19 +1487,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -1489,19 +1509,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -1532,6 +1552,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION (TOK_CHAR 5) "hello"))))) @@ -1547,7 +1568,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -1590,19 +1611,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -1612,19 +1633,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -1655,6 +1676,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION (TOK_VARCHAR 5) "hello"))))) @@ -1670,7 +1692,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -1713,19 +1735,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -1735,19 +1757,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -1780,6 +1802,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION unbase64 "0xe23"))))) @@ -1795,7 +1818,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -1838,19 +1861,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -1860,19 +1883,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -1905,6 +1928,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION TOK_TINYINT "1")) (TOK_SELEXPR (TOK_FUNCTION TOK_SMALLINT "20"))))) @@ -1920,7 +1944,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -1965,19 +1989,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -1987,19 +2011,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -2032,6 +2056,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION TOK_TIMESTAMP "1970-12-31 15:59:58.174"))))) @@ -2047,7 +2072,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -2090,19 +2115,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -2112,19 +2137,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -2157,6 +2182,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION TOK_DATE "1970-12-31 15:59:58.174"))))) @@ -2172,7 +2198,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -2215,19 +2241,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -2237,19 +2263,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -2282,6 +2308,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION TOK_DECIMAL "58.174"))))) @@ -2297,7 +2324,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -2340,19 +2367,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -2362,19 +2389,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -2407,6 +2434,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION array 1 2 3))))) @@ -2422,7 +2450,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -2465,19 +2493,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -2487,19 +2515,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -2532,6 +2560,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION str_to_map "a=1 b=2 c=3" " " "="))))) @@ -2547,7 +2576,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -2590,19 +2619,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -2612,19 +2641,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -2657,6 +2686,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION NAMED_STRUCT "a" 11 "b" 11))))) @@ -2672,7 +2702,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -2715,19 +2745,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -2737,19 +2767,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -2782,6 +2812,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION CREATE_UNION 0 "hello"))))) @@ -2797,7 +2828,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -2840,19 +2871,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -2862,19 +2893,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -2909,6 +2940,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTIONSTAR count))))) @@ -2924,11 +2956,11 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE Group By Operator aggregations: expr: count() @@ -2957,19 +2989,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -2979,19 +3011,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -3065,6 +3097,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_FUNCTION count 1))))) @@ -3080,11 +3113,11 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE Group By Operator aggregations: expr: count(1) @@ -3113,19 +3146,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -3135,19 +3168,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -3221,6 +3254,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF) (TOK_SELEXPR 11)))) @@ -3236,7 +3270,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: PARTIAL + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: PARTIAL GatherStats: false Select Operator expressions: @@ -3262,6 +3296,8 @@ STAGE PLANS: type: timestamp expr: s1 type: string + expr: vc1 + type: varchar(5) expr: m1 type: map expr: l1 @@ -3270,23 +3306,23 @@ STAGE PLANS: type: struct expr: 11 type: int - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15 Statistics: - numRows: 2 dataSize: 254 basicStatsState: COMPLETE colStatsState: PARTIAL + numRows: 2 dataSize: 428 basicStatsState: COMPLETE colStatsState: PARTIAL File Output Operator compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 Statistics: - numRows: 2 dataSize: 254 basicStatsState: COMPLETE colStatsState: PARTIAL + numRows: 2 dataSize: 428 basicStatsState: COMPLETE colStatsState: PARTIAL #### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - columns _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct:int + columns _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct:int escape.delim \ hive.serialization.extend.nesting.levels true serialization.format 1 @@ -3307,19 +3343,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -3329,19 +3365,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -3378,6 +3414,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL i1))) (TOK_LIMIT 10))) temp)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL i1))))) @@ -3393,7 +3430,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -3425,19 +3462,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -3447,19 +3484,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -3531,6 +3568,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL i1)) (TOK_SELEXPR 11)) (TOK_LIMIT 10))) temp)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL i1))))) @@ -3546,7 +3584,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -3582,19 +3620,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -3604,19 +3642,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -3688,6 +3726,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL i1)) (TOK_SELEXPR 11)) (TOK_LIMIT 10))) temp)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL i1)) (TOK_SELEXPR "hello")))) @@ -3703,7 +3742,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -3739,19 +3778,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -3761,19 +3800,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -3847,6 +3886,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL i1)) (TOK_SELEXPR 11.0 x)) (TOK_LIMIT 10))) temp)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL x))))) @@ -3862,7 +3902,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -3898,19 +3938,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -3920,19 +3960,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -4004,6 +4044,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL i1) x) (TOK_SELEXPR (TOK_FUNCTION unbase64 "0xe23") ub)) (TOK_LIMIT 10))) temp)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL x)) (TOK_SELEXPR "hello")))) @@ -4019,7 +4060,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -4055,19 +4096,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -4077,19 +4118,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -4165,6 +4206,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL i1)) (TOK_SELEXPR "hello" hell)) (TOK_LIMIT 10))) in1)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL hell) h)) (TOK_LIMIT 10))) in2)) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL h)) (TOK_SELEXPR 11.0)))) @@ -4181,7 +4223,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Select Operator expressions: @@ -4217,19 +4259,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -4239,19 +4281,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -4400,6 +4442,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL bo1))) (TOK_WHERE (TOK_TABLE_OR_COL bo1)))) @@ -4415,7 +4458,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Filter Operator isSamplingPred: false @@ -4465,19 +4508,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -4487,19 +4530,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc @@ -4532,6 +4575,7 @@ POSTHOOK: Lineage: alltypes_orc.si1 SIMPLE [(alltypes)alltypes.FieldSchema(name: POSTHOOK: Lineage: alltypes_orc.st1 SIMPLE [(alltypes)alltypes.FieldSchema(name:st1, type:struct, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ti1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ti1, type:tinyint, comment:null), ] POSTHOOK: Lineage: alltypes_orc.ts1 SIMPLE [(alltypes)alltypes.FieldSchema(name:ts1, type:timestamp, comment:null), ] +POSTHOOK: Lineage: alltypes_orc.vc1 SIMPLE [(alltypes)alltypes.FieldSchema(name:vc1, type:varchar(5), comment:null), ] ABSTRACT SYNTAX TREE: (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME alltypes_orc))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL bo1))) (TOK_WHERE (! (TOK_TABLE_OR_COL bo1))))) @@ -4547,7 +4591,7 @@ STAGE PLANS: TableScan alias: alltypes_orc Statistics: - numRows: 2 dataSize: 1514 basicStatsState: COMPLETE colStatsState: COMPLETE + numRows: 2 dataSize: 1686 basicStatsState: COMPLETE colStatsState: COMPLETE GatherStats: false Filter Operator isSamplingPred: false @@ -4597,19 +4641,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -4619,19 +4663,19 @@ STAGE PLANS: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim , - columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,m1,l1,st1 - columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:map:array:struct + columns bo1,ti1,si1,i1,bi1,f1,d1,de1,ts1,da1,s1,vc1,m1,l1,st1 + columns.types boolean:tinyint:smallint:int:bigint:float:double:decimal(10,0):timestamp:timestamp:string:varchar(5):map:array:struct field.delim | #### A masked pattern was here #### mapkey.delim : name default.alltypes_orc numFiles 1 numRows 2 - rawDataSize 1514 - serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, map m1, list l1, struct st1} + rawDataSize 1686 + serialization.ddl struct alltypes_orc { bool bo1, byte ti1, i16 si1, i32 i1, i64 bi1, float f1, double d1, decimal(10,0) de1, timestamp ts1, timestamp da1, string s1, varchar(5) vc1, map m1, list l1, struct st1} serialization.format | serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 1409 + totalSize 1475 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypes_orc