diff --git itests/hive-blobstore/src/test/queries/clientpositive/create_table_timestamp_partition.q itests/hive-blobstore/src/test/queries/clientpositive/create_table_timestamp_partition.q new file mode 100644 index 0000000..17186a3 --- /dev/null +++ itests/hive-blobstore/src/test/queries/clientpositive/create_table_timestamp_partition.q @@ -0,0 +1,34 @@ +set hive.exec.dynamic.partition.mode=nonstrict; + +DROP TABLE blobstore_partitioned_source_table; +DROP TABLE blobstore_partitioned_target_table; + +CREATE EXTERNAL TABLE blobstore_partitioned_source_table ( + tsbucket TIMESTAMP,geo_country STRING,aid BIGINT) +STORED AS ORC +LOCATION '${hiveconf:test.blobstore.path.unique}/source_table/data' +TBLPROPERTIES("orc.compress"="ZLIB"); + +INSERT INTO TABLE blobstore_partitioned_source_table VALUES + ('2016-11-02 17:00:00','France',74530), + ('2016-11-02 18:00:00','Canada',57008), + ('2016-11-02 17:00:00','Morocco',58097); + +CREATE EXTERNAL TABLE blobstore_partitioned_target_table ( + geo_country STRING,aid BIGINT) +PARTITIONED BY (tsbucket TIMESTAMP) +STORED AS ORC +LOCATION '${hiveconf:test.blobstore.path.unique}/target_table/data' +TBLPROPERTIES("orc.compress"="ZLIB"); + +INSERT INTO TABLE blobstore_partitioned_target_table PARTITION (tsbucket) +SELECT geo_country,aid,tsbucket FROM blobstore_partitioned_source_table; + +SHOW PARTITIONS blobstore_partitioned_target_table; + +DESCRIBE formatted blobstore_partitioned_target_table PARTITION (tsbucket='2016-11-02 17:00:00'); + +DESCRIBE formatted blobstore_partitioned_target_table PARTITION (tsbucket='2016-11-02 17:00:00.0'); + +DROP TABLE blobstore_partitioned_source_table; +DROP TABLE blobstore_partitioned_target_table; diff --git itests/hive-blobstore/src/test/results/clientpositive/create_table_timestamp_partition.q.out itests/hive-blobstore/src/test/results/clientpositive/create_table_timestamp_partition.q.out new file mode 100644 index 0000000..acb25e4 --- /dev/null +++ itests/hive-blobstore/src/test/results/clientpositive/create_table_timestamp_partition.q.out @@ -0,0 +1,174 @@ +PREHOOK: query: DROP TABLE blobstore_partitioned_source_table +PREHOOK: type: DROPTABLE +POSTHOOK: query: DROP TABLE blobstore_partitioned_source_table +POSTHOOK: type: DROPTABLE +PREHOOK: query: DROP TABLE blobstore_partitioned_target_table +PREHOOK: type: DROPTABLE +POSTHOOK: query: DROP TABLE blobstore_partitioned_target_table +POSTHOOK: type: DROPTABLE +PREHOOK: query: CREATE EXTERNAL TABLE blobstore_partitioned_source_table ( + tsbucket TIMESTAMP,geo_country STRING,aid BIGINT) +STORED AS ORC +#### A masked pattern was here #### +TBLPROPERTIES("orc.compress"="ZLIB") +PREHOOK: type: CREATETABLE +PREHOOK: Input: ### test.blobstore.path ###/source_table/data +PREHOOK: Output: database:default +PREHOOK: Output: default@blobstore_partitioned_source_table +POSTHOOK: query: CREATE EXTERNAL TABLE blobstore_partitioned_source_table ( + tsbucket TIMESTAMP,geo_country STRING,aid BIGINT) +STORED AS ORC +#### A masked pattern was here #### +TBLPROPERTIES("orc.compress"="ZLIB") +POSTHOOK: type: CREATETABLE +POSTHOOK: Input: ### test.blobstore.path ###/source_table/data +POSTHOOK: Output: database:default +POSTHOOK: Output: default@blobstore_partitioned_source_table +PREHOOK: query: INSERT INTO TABLE blobstore_partitioned_source_table VALUES + ('2016-11-02 17:00:00','France',74530), + ('2016-11-02 18:00:00','Canada',57008), + ('2016-11-02 17:00:00','Morocco',58097) +PREHOOK: type: QUERY +PREHOOK: Output: default@blobstore_partitioned_source_table +POSTHOOK: query: INSERT INTO TABLE blobstore_partitioned_source_table VALUES + ('2016-11-02 17:00:00','France',74530), + ('2016-11-02 18:00:00','Canada',57008), + ('2016-11-02 17:00:00','Morocco',58097) +POSTHOOK: type: QUERY +POSTHOOK: Output: default@blobstore_partitioned_source_table +POSTHOOK: Lineage: blobstore_partitioned_source_table.aid EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col3, type:string, comment:), ] +POSTHOOK: Lineage: blobstore_partitioned_source_table.geo_country SIMPLE [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col2, type:string, comment:), ] +POSTHOOK: Lineage: blobstore_partitioned_source_table.tsbucket EXPRESSION [(values__tmp__table__1)values__tmp__table__1.FieldSchema(name:tmp_values_col1, type:string, comment:), ] +PREHOOK: query: CREATE EXTERNAL TABLE blobstore_partitioned_target_table ( + geo_country STRING,aid BIGINT) +PARTITIONED BY (tsbucket TIMESTAMP) +STORED AS ORC +#### A masked pattern was here #### +TBLPROPERTIES("orc.compress"="ZLIB") +PREHOOK: type: CREATETABLE +PREHOOK: Input: ### test.blobstore.path ###/target_table/data +PREHOOK: Output: database:default +PREHOOK: Output: default@blobstore_partitioned_target_table +POSTHOOK: query: CREATE EXTERNAL TABLE blobstore_partitioned_target_table ( + geo_country STRING,aid BIGINT) +PARTITIONED BY (tsbucket TIMESTAMP) +STORED AS ORC +#### A masked pattern was here #### +TBLPROPERTIES("orc.compress"="ZLIB") +POSTHOOK: type: CREATETABLE +POSTHOOK: Input: ### test.blobstore.path ###/target_table/data +POSTHOOK: Output: database:default +POSTHOOK: Output: default@blobstore_partitioned_target_table +PREHOOK: query: INSERT INTO TABLE blobstore_partitioned_target_table PARTITION (tsbucket) +SELECT geo_country,aid,tsbucket FROM blobstore_partitioned_source_table +PREHOOK: type: QUERY +PREHOOK: Input: default@blobstore_partitioned_source_table +PREHOOK: Output: default@blobstore_partitioned_target_table +POSTHOOK: query: INSERT INTO TABLE blobstore_partitioned_target_table PARTITION (tsbucket) +SELECT geo_country,aid,tsbucket FROM blobstore_partitioned_source_table +POSTHOOK: type: QUERY +POSTHOOK: Input: default@blobstore_partitioned_source_table +POSTHOOK: Output: default@blobstore_partitioned_target_table@tsbucket=2016-11-02 17%3A00%3A00 +POSTHOOK: Output: default@blobstore_partitioned_target_table@tsbucket=2016-11-02 18%3A00%3A00 +POSTHOOK: Lineage: blobstore_partitioned_target_table PARTITION(tsbucket=2016-11-02 17:00:00).aid SIMPLE [(blobstore_partitioned_source_table)blobstore_partitioned_source_table.FieldSchema(name:aid, type:bigint, comment:null), ] +POSTHOOK: Lineage: blobstore_partitioned_target_table PARTITION(tsbucket=2016-11-02 17:00:00).geo_country SIMPLE [(blobstore_partitioned_source_table)blobstore_partitioned_source_table.FieldSchema(name:geo_country, type:string, comment:null), ] +POSTHOOK: Lineage: blobstore_partitioned_target_table PARTITION(tsbucket=2016-11-02 18:00:00).aid SIMPLE [(blobstore_partitioned_source_table)blobstore_partitioned_source_table.FieldSchema(name:aid, type:bigint, comment:null), ] +POSTHOOK: Lineage: blobstore_partitioned_target_table PARTITION(tsbucket=2016-11-02 18:00:00).geo_country SIMPLE [(blobstore_partitioned_source_table)blobstore_partitioned_source_table.FieldSchema(name:geo_country, type:string, comment:null), ] +PREHOOK: query: SHOW PARTITIONS blobstore_partitioned_target_table +PREHOOK: type: SHOWPARTITIONS +PREHOOK: Input: default@blobstore_partitioned_target_table +POSTHOOK: query: SHOW PARTITIONS blobstore_partitioned_target_table +POSTHOOK: type: SHOWPARTITIONS +POSTHOOK: Input: default@blobstore_partitioned_target_table +tsbucket=2016-11-02 17%3A00%3A00 +tsbucket=2016-11-02 18%3A00%3A00 +PREHOOK: query: DESCRIBE formatted blobstore_partitioned_target_table PARTITION (tsbucket='2016-11-02 17:00:00') +PREHOOK: type: DESCTABLE +PREHOOK: Input: default@blobstore_partitioned_target_table +POSTHOOK: query: DESCRIBE formatted blobstore_partitioned_target_table PARTITION (tsbucket='2016-11-02 17:00:00') +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: default@blobstore_partitioned_target_table +# col_name data_type comment +geo_country string +aid bigint + +# Partition Information +# col_name data_type comment +tsbucket timestamp + +# Detailed Partition Information +Partition Value: [2016-11-02 17:00:00] +Database: default +Table: blobstore_partitioned_target_table +#### A masked pattern was here #### +Partition Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + numFiles 1 + numRows 2 + rawDataSize 196 + totalSize 355 +#### A masked pattern was here #### + +# Storage Information +SerDe Library: org.apache.hadoop.hive.ql.io.orc.OrcSerde +InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat +OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat +Compressed: No +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] +Storage Desc Params: + serialization.format 1 +PREHOOK: query: DESCRIBE formatted blobstore_partitioned_target_table PARTITION (tsbucket='2016-11-02 17:00:00.0') +PREHOOK: type: DESCTABLE +PREHOOK: Input: default@blobstore_partitioned_target_table +POSTHOOK: query: DESCRIBE formatted blobstore_partitioned_target_table PARTITION (tsbucket='2016-11-02 17:00:00.0') +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: default@blobstore_partitioned_target_table +# col_name data_type comment +geo_country string +aid bigint + +# Partition Information +# col_name data_type comment +tsbucket timestamp + +# Detailed Partition Information +Partition Value: [2016-11-02 17:00:00] +Database: default +Table: blobstore_partitioned_target_table +#### A masked pattern was here #### +Partition Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} + numFiles 1 + numRows 2 + rawDataSize 196 + totalSize 355 +#### A masked pattern was here #### + +# Storage Information +SerDe Library: org.apache.hadoop.hive.ql.io.orc.OrcSerde +InputFormat: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat +OutputFormat: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat +Compressed: No +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] +Storage Desc Params: + serialization.format 1 +PREHOOK: query: DROP TABLE blobstore_partitioned_source_table +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@blobstore_partitioned_source_table +PREHOOK: Output: default@blobstore_partitioned_source_table +POSTHOOK: query: DROP TABLE blobstore_partitioned_source_table +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@blobstore_partitioned_source_table +POSTHOOK: Output: default@blobstore_partitioned_source_table +PREHOOK: query: DROP TABLE blobstore_partitioned_target_table +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@blobstore_partitioned_target_table +PREHOOK: Output: default@blobstore_partitioned_target_table +POSTHOOK: query: DROP TABLE blobstore_partitioned_target_table +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@blobstore_partitioned_target_table +POSTHOOK: Output: default@blobstore_partitioned_target_table diff --git ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java index 0f7ef8b..d8d7554 100644 --- ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java +++ ql/src/java/org/apache/hadoop/hive/ql/parse/BaseSemanticAnalyzer.java @@ -1701,7 +1701,7 @@ public static void validatePartColumnType(Table tbl, Map partSpe TypeInfo expectedType = TypeInfoUtils.getTypeInfoFromTypeString(colType); ObjectInspector outputOI = - TypeInfoUtils.getStandardJavaObjectInspectorFromTypeInfo(expectedType); + TypeInfoUtils.getStandardWritableObjectInspectorFromTypeInfo(expectedType); // Since partVal is a constant, it is safe to cast ExprNodeDesc to ExprNodeConstantDesc. // Its value should be in normalized format (e.g. no leading zero in integer, date is in // format of YYYY-MM-DD etc) diff --git ql/src/test/results/clientpositive/partition_timestamp.q.out ql/src/test/results/clientpositive/partition_timestamp.q.out index 34f70a5..a80ed2b 100644 --- ql/src/test/results/clientpositive/partition_timestamp.q.out +++ ql/src/test/results/clientpositive/partition_timestamp.q.out @@ -14,79 +14,79 @@ PREHOOK: query: insert overwrite table partition_timestamp_1 partition(dt='2000- select * from src tablesample (10 rows) PREHOOK: type: QUERY PREHOOK: Input: default@src -PREHOOK: Output: default@partition_timestamp_1@dt=2000-01-01 01%3A00%3A00.0/region=1 +PREHOOK: Output: default@partition_timestamp_1@dt=2000-01-01 01%3A00%3A00/region=1 POSTHOOK: query: insert overwrite table partition_timestamp_1 partition(dt='2000-01-01 01:00:00', region= '1') select * from src tablesample (10 rows) POSTHOOK: type: QUERY POSTHOOK: Input: default@src -POSTHOOK: Output: default@partition_timestamp_1@dt=2000-01-01 01%3A00%3A00.0/region=1 -POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2000-01-01 01:00:00.0,region=1).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2000-01-01 01:00:00.0,region=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Output: default@partition_timestamp_1@dt=2000-01-01 01%3A00%3A00/region=1 +POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2000-01-01 01:00:00,region=1).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2000-01-01 01:00:00,region=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] PREHOOK: query: insert overwrite table partition_timestamp_1 partition(dt='2000-01-01 02:00:00', region= '2') select * from src tablesample (5 rows) PREHOOK: type: QUERY PREHOOK: Input: default@src -PREHOOK: Output: default@partition_timestamp_1@dt=2000-01-01 02%3A00%3A00.0/region=2 +PREHOOK: Output: default@partition_timestamp_1@dt=2000-01-01 02%3A00%3A00/region=2 POSTHOOK: query: insert overwrite table partition_timestamp_1 partition(dt='2000-01-01 02:00:00', region= '2') select * from src tablesample (5 rows) POSTHOOK: type: QUERY POSTHOOK: Input: default@src -POSTHOOK: Output: default@partition_timestamp_1@dt=2000-01-01 02%3A00%3A00.0/region=2 -POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2000-01-01 02:00:00.0,region=2).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2000-01-01 02:00:00.0,region=2).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Output: default@partition_timestamp_1@dt=2000-01-01 02%3A00%3A00/region=2 +POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2000-01-01 02:00:00,region=2).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2000-01-01 02:00:00,region=2).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] PREHOOK: query: insert overwrite table partition_timestamp_1 partition(dt='2001-01-01 01:00:00', region= '2020-20-20') select * from src tablesample (5 rows) PREHOOK: type: QUERY PREHOOK: Input: default@src -PREHOOK: Output: default@partition_timestamp_1@dt=2001-01-01 01%3A00%3A00.0/region=2020-20-20 +PREHOOK: Output: default@partition_timestamp_1@dt=2001-01-01 01%3A00%3A00/region=2020-20-20 POSTHOOK: query: insert overwrite table partition_timestamp_1 partition(dt='2001-01-01 01:00:00', region= '2020-20-20') select * from src tablesample (5 rows) POSTHOOK: type: QUERY POSTHOOK: Input: default@src -POSTHOOK: Output: default@partition_timestamp_1@dt=2001-01-01 01%3A00%3A00.0/region=2020-20-20 -POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2001-01-01 01:00:00.0,region=2020-20-20).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2001-01-01 01:00:00.0,region=2020-20-20).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Output: default@partition_timestamp_1@dt=2001-01-01 01%3A00%3A00/region=2020-20-20 +POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2001-01-01 01:00:00,region=2020-20-20).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2001-01-01 01:00:00,region=2020-20-20).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] PREHOOK: query: insert overwrite table partition_timestamp_1 partition(dt='2001-01-01 02:00:00', region= '1') select * from src tablesample (20 rows) PREHOOK: type: QUERY PREHOOK: Input: default@src -PREHOOK: Output: default@partition_timestamp_1@dt=2001-01-01 02%3A00%3A00.0/region=1 +PREHOOK: Output: default@partition_timestamp_1@dt=2001-01-01 02%3A00%3A00/region=1 POSTHOOK: query: insert overwrite table partition_timestamp_1 partition(dt='2001-01-01 02:00:00', region= '1') select * from src tablesample (20 rows) POSTHOOK: type: QUERY POSTHOOK: Input: default@src -POSTHOOK: Output: default@partition_timestamp_1@dt=2001-01-01 02%3A00%3A00.0/region=1 -POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2001-01-01 02:00:00.0,region=1).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2001-01-01 02:00:00.0,region=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Output: default@partition_timestamp_1@dt=2001-01-01 02%3A00%3A00/region=1 +POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2001-01-01 02:00:00,region=1).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2001-01-01 02:00:00,region=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] PREHOOK: query: insert overwrite table partition_timestamp_1 partition(dt='2001-01-01 03:00:00', region= '10') select * from src tablesample (11 rows) PREHOOK: type: QUERY PREHOOK: Input: default@src -PREHOOK: Output: default@partition_timestamp_1@dt=2001-01-01 03%3A00%3A00.0/region=10 +PREHOOK: Output: default@partition_timestamp_1@dt=2001-01-01 03%3A00%3A00/region=10 POSTHOOK: query: insert overwrite table partition_timestamp_1 partition(dt='2001-01-01 03:00:00', region= '10') select * from src tablesample (11 rows) POSTHOOK: type: QUERY POSTHOOK: Input: default@src -POSTHOOK: Output: default@partition_timestamp_1@dt=2001-01-01 03%3A00%3A00.0/region=10 -POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2001-01-01 03:00:00.0,region=10).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2001-01-01 03:00:00.0,region=10).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Output: default@partition_timestamp_1@dt=2001-01-01 03%3A00%3A00/region=10 +POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2001-01-01 03:00:00,region=10).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: partition_timestamp_1 PARTITION(dt=2001-01-01 03:00:00,region=10).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] PREHOOK: query: select distinct dt from partition_timestamp_1 PREHOOK: type: QUERY PREHOOK: Input: default@partition_timestamp_1 -PREHOOK: Input: default@partition_timestamp_1@dt=2000-01-01 01%3A00%3A00.0/region=1 -PREHOOK: Input: default@partition_timestamp_1@dt=2000-01-01 02%3A00%3A00.0/region=2 -PREHOOK: Input: default@partition_timestamp_1@dt=2001-01-01 01%3A00%3A00.0/region=2020-20-20 -PREHOOK: Input: default@partition_timestamp_1@dt=2001-01-01 02%3A00%3A00.0/region=1 -PREHOOK: Input: default@partition_timestamp_1@dt=2001-01-01 03%3A00%3A00.0/region=10 +PREHOOK: Input: default@partition_timestamp_1@dt=2000-01-01 01%3A00%3A00/region=1 +PREHOOK: Input: default@partition_timestamp_1@dt=2000-01-01 02%3A00%3A00/region=2 +PREHOOK: Input: default@partition_timestamp_1@dt=2001-01-01 01%3A00%3A00/region=2020-20-20 +PREHOOK: Input: default@partition_timestamp_1@dt=2001-01-01 02%3A00%3A00/region=1 +PREHOOK: Input: default@partition_timestamp_1@dt=2001-01-01 03%3A00%3A00/region=10 #### A masked pattern was here #### POSTHOOK: query: select distinct dt from partition_timestamp_1 POSTHOOK: type: QUERY POSTHOOK: Input: default@partition_timestamp_1 -POSTHOOK: Input: default@partition_timestamp_1@dt=2000-01-01 01%3A00%3A00.0/region=1 -POSTHOOK: Input: default@partition_timestamp_1@dt=2000-01-01 02%3A00%3A00.0/region=2 -POSTHOOK: Input: default@partition_timestamp_1@dt=2001-01-01 01%3A00%3A00.0/region=2020-20-20 -POSTHOOK: Input: default@partition_timestamp_1@dt=2001-01-01 02%3A00%3A00.0/region=1 -POSTHOOK: Input: default@partition_timestamp_1@dt=2001-01-01 03%3A00%3A00.0/region=10 +POSTHOOK: Input: default@partition_timestamp_1@dt=2000-01-01 01%3A00%3A00/region=1 +POSTHOOK: Input: default@partition_timestamp_1@dt=2000-01-01 02%3A00%3A00/region=2 +POSTHOOK: Input: default@partition_timestamp_1@dt=2001-01-01 01%3A00%3A00/region=2020-20-20 +POSTHOOK: Input: default@partition_timestamp_1@dt=2001-01-01 02%3A00%3A00/region=1 +POSTHOOK: Input: default@partition_timestamp_1@dt=2001-01-01 03%3A00%3A00/region=10 #### A masked pattern was here #### 2000-01-01 01:00:00 2000-01-01 02:00:00