diff --git itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/TestAuthorizationApiAuthorizer.java itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/TestAuthorizationApiAuthorizer.java index e65bf4d..e5182b2 100644 --- itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/TestAuthorizationApiAuthorizer.java +++ itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/TestAuthorizationApiAuthorizer.java @@ -156,7 +156,7 @@ public void testCreateRole() throws Exception { FunctionInvoker invoker = new FunctionInvoker() { @Override public void invoke() throws Exception { - msc.create_role(new Role()); + msc.create_role(new Role("role1", 0, "owner")); } }; testFunction(invoker); diff --git itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/StorageBasedMetastoreTestBase.java itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/StorageBasedMetastoreTestBase.java index b59d2e1..ae853ba 100644 --- itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/StorageBasedMetastoreTestBase.java +++ itests/hive-unit/src/test/java/org/apache/hadoop/hive/ql/security/StorageBasedMetastoreTestBase.java @@ -34,8 +34,10 @@ import org.apache.hadoop.hive.ql.security.authorization.AuthorizationPreEventListener; import org.apache.hadoop.hive.ql.security.authorization.StorageBasedAuthorizationProvider; import org.apache.hadoop.hive.ql.session.SessionState; +import org.apache.hadoop.hive.ql.WindowsPathUtil; import org.apache.hadoop.hive.shims.ShimLoader; import org.apache.hadoop.security.UserGroupInformation; +import org.apache.hadoop.util.Shell; import org.junit.After; import org.junit.Assert; import org.junit.Before; @@ -55,7 +57,11 @@ protected String getAuthorizationProvider(){ } protected HiveConf createHiveConf() throws Exception { - return new HiveConf(this.getClass()); + HiveConf conf = new HiveConf(this.getClass()); + if (Shell.WINDOWS) { + WindowsPathUtil.convertPathsFromWindowsToHdfs(conf); + } + return conf; } @Before diff --git ql/src/test/queries/clientpositive/create_like.q ql/src/test/queries/clientpositive/create_like.q index eb9f191..b1785f5 100644 --- ql/src/test/queries/clientpositive/create_like.q +++ ql/src/test/queries/clientpositive/create_like.q @@ -21,8 +21,10 @@ INSERT OVERWRITE TABLE table2 SELECT key, value FROM src WHERE key = 100; SELECT * FROM table1; SELECT * FROM table2; -CREATE EXTERNAL TABLE table4 (a INT) LOCATION '${system:hive.root}/data/files/ext_test'; -CREATE EXTERNAL TABLE table5 LIKE table4 LOCATION '${system:hive.root}/data/files/ext_test'; +dfs -cp ${system:hive.root}/data/files/ext_test ${system:test.tmp.dir}/ext_test; + +CREATE EXTERNAL TABLE table4 (a INT) LOCATION '${system:test.tmp.dir}/ext_test'; +CREATE EXTERNAL TABLE table5 LIKE table4 LOCATION '${system:test.tmp.dir}/ext_test'; SELECT * FROM table4; SELECT * FROM table5; @@ -31,7 +33,7 @@ DROP TABLE table5; SELECT * FROM table4; DROP TABLE table4; -CREATE EXTERNAL TABLE table4 (a INT) LOCATION '${system:hive.root}/data/files/ext_test'; +CREATE EXTERNAL TABLE table4 (a INT) LOCATION '${system:test.tmp.dir}/ext_test'; SELECT * FROM table4; CREATE TABLE doctors STORED AS AVRO TBLPROPERTIES ('avro.schema.literal'='{ diff --git ql/src/test/queries/clientpositive/stats_noscan_2.q ql/src/test/queries/clientpositive/stats_noscan_2.q index b106b30..8639ab7 100644 --- ql/src/test/queries/clientpositive/stats_noscan_2.q +++ ql/src/test/queries/clientpositive/stats_noscan_2.q @@ -1,7 +1,8 @@ +dfs -cp ${system:hive.root}/data/files/ext_test ${system:test.tmp.dir}/analyze_external; -- test analyze table compute statistiscs [noscan] on external table -- 1 test table -CREATE EXTERNAL TABLE anaylyze_external (a INT) LOCATION '${system:hive.root}/data/files/ext_test'; +CREATE EXTERNAL TABLE anaylyze_external (a INT) LOCATION '${system:test.tmp.dir}/analyze_external'; SELECT * FROM anaylyze_external; analyze table anaylyze_external compute statistics noscan; describe formatted anaylyze_external; diff --git ql/src/test/results/clientpositive/input_part10_win.q.out ql/src/test/results/clientpositive/input_part10_win.q.out index 92df63c..80da57c 100644 --- ql/src/test/results/clientpositive/input_part10_win.q.out +++ ql/src/test/results/clientpositive/input_part10_win.q.out @@ -9,6 +9,8 @@ CREATE TABLE part_special ( ts STRING ) PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@part_special POSTHOOK: query: -- INCLUDE_OS_WINDOWS -- included only on windows because of difference in file name encoding logic @@ -20,6 +22,7 @@ CREATE TABLE part_special ( ts STRING ) POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default POSTHOOK: Output: default@part_special PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE part_special PARTITION(ds='2008 04 08', ts = '10:11:12=455') @@ -29,9 +32,6 @@ POSTHOOK: query: EXPLAIN INSERT OVERWRITE TABLE part_special PARTITION(ds='2008 04 08', ts = '10:11:12=455') SELECT 1, 2 FROM src LIMIT 1 POSTHOOK: type: QUERY -ABSTRACT SYNTAX TREE: - (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME part_special) (TOK_PARTSPEC (TOK_PARTVAL ds '2008 04 08') (TOK_PARTVAL ts '10:11:12=455')))) (TOK_SELECT (TOK_SELEXPR 1) (TOK_SELEXPR 2)) (TOK_LIMIT 1))) - STAGE DEPENDENCIES: Stage-1 is a root stage Stage-0 depends on stages: Stage-1 @@ -40,32 +40,32 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 Map Reduce - Alias -> Map Operator Tree: - src + Map Operator Tree: TableScan alias: src + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE Select Operator - expressions: - expr: 1 - type: int - expr: 2 - type: int + expressions: 1 (type: int), 2 (type: int) outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE Limit + Number of rows: 1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator sort order: - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: int + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: int) Reduce Operator Tree: - Extract + Select Operator + expressions: VALUE._col0 (type: int), VALUE._col1 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE Limit + Number of rows: 1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - GlobalTableId: 1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -88,7 +88,6 @@ STAGE PLANS: Stage: Stage-2 Stats-Aggr Operator - PREHOOK: query: INSERT OVERWRITE TABLE part_special PARTITION(ds='2008 04 08', ts = '10:11:12=455') SELECT 1, 2 FROM src LIMIT 1 PREHOOK: type: QUERY @@ -103,20 +102,20 @@ POSTHOOK: Lineage: part_special PARTITION(ds=2008 04 08,ts=10:11:12=455).a SIMPL POSTHOOK: Lineage: part_special PARTITION(ds=2008 04 08,ts=10:11:12=455).b SIMPLE [] PREHOOK: query: DESCRIBE EXTENDED part_special PARTITION(ds='2008 04 08', ts = '10:11:12=455') PREHOOK: type: DESCTABLE +PREHOOK: Input: default@part_special POSTHOOK: query: DESCRIBE EXTENDED part_special PARTITION(ds='2008 04 08', ts = '10:11:12=455') POSTHOOK: type: DESCTABLE -POSTHOOK: Lineage: part_special PARTITION(ds=2008 04 08,ts=10:11:12=455).a SIMPLE [] -POSTHOOK: Lineage: part_special PARTITION(ds=2008 04 08,ts=10:11:12=455).b SIMPLE [] -a string None -b string None -ds string None -ts string None +POSTHOOK: Input: default@part_special +a string +b string +ds string +ts string # Partition Information # col_name data_type comment -ds string None -ts string None +ds string +ts string #### A masked pattern was here #### PREHOOK: query: SELECT * FROM part_special WHERE ds='2008 04 08' AND ts = '10:11:12=455' @@ -129,6 +128,4 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@part_special POSTHOOK: Input: default@part_special@ds=2008%2004%2008/ts=10%3A11%3A12%3D455 #### A masked pattern was here #### -POSTHOOK: Lineage: part_special PARTITION(ds=2008 04 08,ts=10:11:12=455).a SIMPLE [] -POSTHOOK: Lineage: part_special PARTITION(ds=2008 04 08,ts=10:11:12=455).b SIMPLE [] 1 2 2008 04 08 10:11:12=455 diff --git ql/src/test/results/clientpositive/load_dyn_part14_win.q.out ql/src/test/results/clientpositive/load_dyn_part14_win.q.out index 13b6c0b..e499552 100644 --- ql/src/test/results/clientpositive/load_dyn_part14_win.q.out +++ ql/src/test/results/clientpositive/load_dyn_part14_win.q.out @@ -5,6 +5,8 @@ PREHOOK: query: -- INCLUDE_OS_WINDOWS create table if not exists nzhang_part14 (key string) partitioned by (value string) PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@nzhang_part14 POSTHOOK: query: -- INCLUDE_OS_WINDOWS -- included only on windows because of difference in file name encoding logic @@ -12,18 +14,21 @@ POSTHOOK: query: -- INCLUDE_OS_WINDOWS create table if not exists nzhang_part14 (key string) partitioned by (value string) POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default POSTHOOK: Output: default@nzhang_part14 PREHOOK: query: describe extended nzhang_part14 PREHOOK: type: DESCTABLE +PREHOOK: Input: default@nzhang_part14 POSTHOOK: query: describe extended nzhang_part14 POSTHOOK: type: DESCTABLE -key string None -value string None +POSTHOOK: Input: default@nzhang_part14 +key string +value string # Partition Information # col_name data_type comment -value string None +value string #### A masked pattern was here #### PREHOOK: query: explain @@ -46,9 +51,6 @@ select key, value from ( select 'k3' as key, ' ' as value from src limit 2 ) T POSTHOOK: type: QUERY -ABSTRACT SYNTAX TREE: - (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_UNION (TOK_UNION (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR 'k1' key) (TOK_SELEXPR (TOK_FUNCTION TOK_STRING TOK_NULL) value)) (TOK_LIMIT 2))) (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR 'k2' key) (TOK_SELEXPR '' value)) (TOK_LIMIT 2)))) (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR 'k3' key) (TOK_SELEXPR ' ' value)) (TOK_LIMIT 2)))) T)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME nzhang_part14) (TOK_PARTSPEC (TOK_PARTVAL value)))) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL key)) (TOK_SELEXPR (TOK_TABLE_OR_COL value))))) - STAGE DEPENDENCIES: Stage-1 is a root stage Stage-2 depends on stages: Stage-1, Stage-9, Stage-10 @@ -65,88 +67,79 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 Map Reduce - Alias -> Map Operator Tree: - null-subquery1-subquery2:t-subquery1-subquery2:src + Map Operator Tree: TableScan alias: src + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE Select Operator - expressions: - expr: 'k2' - type: string - expr: '' - type: string + expressions: 'k2' (type: string), '' (type: string) outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 85000 Basic stats: COMPLETE Column stats: COMPLETE Limit + Number of rows: 2 + Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator sort order: - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) Reduce Operator Tree: - Extract + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE Limit + Number of rows: 2 + Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - GlobalTableId: 0 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe Stage: Stage-2 Map Reduce - Alias -> Map Operator Tree: -#### A masked pattern was here #### + Map Operator Tree: TableScan Union + Statistics: Num rows: 6 Data size: 1022 Basic stats: COMPLETE Column stats: COMPLETE Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string + expressions: _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - GlobalTableId: 1 + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part14 -#### A masked pattern was here #### TableScan Union + Statistics: Num rows: 6 Data size: 1022 Basic stats: COMPLETE Column stats: COMPLETE Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string + expressions: _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - GlobalTableId: 1 + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part14 -#### A masked pattern was here #### TableScan Union + Statistics: Num rows: 6 Data size: 1022 Basic stats: COMPLETE Column stats: COMPLETE Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: string + expressions: _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - GlobalTableId: 1 + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -179,11 +172,10 @@ STAGE PLANS: Stage: Stage-4 Map Reduce - Alias -> Map Operator Tree: -#### A masked pattern was here #### + Map Operator Tree: + TableScan File Output Operator compressed: false - GlobalTableId: 0 table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -192,11 +184,10 @@ STAGE PLANS: Stage: Stage-6 Map Reduce - Alias -> Map Operator Tree: -#### A masked pattern was here #### + Map Operator Tree: + TableScan File Output Operator compressed: false - GlobalTableId: 0 table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat @@ -211,68 +202,67 @@ STAGE PLANS: Stage: Stage-9 Map Reduce - Alias -> Map Operator Tree: - null-subquery2:t-subquery2:src + Map Operator Tree: TableScan alias: src + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE Select Operator - expressions: - expr: 'k3' - type: string - expr: ' ' - type: string + expressions: 'k3' (type: string), ' ' (type: string) outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 85500 Basic stats: COMPLETE Column stats: COMPLETE Limit + Number of rows: 2 + Statistics: Num rows: 2 Data size: 342 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator sort order: - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + Statistics: Num rows: 2 Data size: 342 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) Reduce Operator Tree: - Extract + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 342 Basic stats: COMPLETE Column stats: COMPLETE Limit + Number of rows: 2 + Statistics: Num rows: 2 Data size: 342 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - GlobalTableId: 0 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe Stage: Stage-10 Map Reduce - Alias -> Map Operator Tree: - null-subquery1-subquery1:t-subquery1-subquery1:src + Map Operator Tree: TableScan alias: src + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE Select Operator - expressions: - expr: 'k1' - type: string - expr: UDFToString(null) - type: string + expressions: 'k1' (type: string), UDFToString(null) (type: string) outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 85000 Basic stats: COMPLETE Column stats: COMPLETE Limit + Number of rows: 2 + Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator sort order: - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string + Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) Reduce Operator Tree: - Extract + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE Limit + Number of rows: 2 + Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - GlobalTableId: 0 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe PREHOOK: query: insert overwrite table nzhang_part14 partition(value) select key, value from ( @@ -295,37 +285,35 @@ select key, value from ( ) T POSTHOOK: type: QUERY POSTHOOK: Input: default@src -POSTHOOK: Output: default@nzhang_part14@value=%20 +POSTHOOK: Output: default@nzhang_part14@value=%2520 POSTHOOK: Output: default@nzhang_part14@value=__HIVE_DEFAULT_PARTITION__ -POSTHOOK: Lineage: nzhang_part14 PARTITION(value= ).key EXPRESSION [] +POSTHOOK: Lineage: nzhang_part14 PARTITION(value=%20).key EXPRESSION [] POSTHOOK: Lineage: nzhang_part14 PARTITION(value=__HIVE_DEFAULT_PARTITION__).key EXPRESSION [] PREHOOK: query: show partitions nzhang_part14 PREHOOK: type: SHOWPARTITIONS +PREHOOK: Input: default@nzhang_part14 POSTHOOK: query: show partitions nzhang_part14 POSTHOOK: type: SHOWPARTITIONS -POSTHOOK: Lineage: nzhang_part14 PARTITION(value= ).key EXPRESSION [] -POSTHOOK: Lineage: nzhang_part14 PARTITION(value=__HIVE_DEFAULT_PARTITION__).key EXPRESSION [] -value=%20 +POSTHOOK: Input: default@nzhang_part14 +value=%2520 value=__HIVE_DEFAULT_PARTITION__ PREHOOK: query: select * from nzhang_part14 where value <> 'a' order by key, value PREHOOK: type: QUERY PREHOOK: Input: default@nzhang_part14 -PREHOOK: Input: default@nzhang_part14@value=%20 +PREHOOK: Input: default@nzhang_part14@value=%2520 PREHOOK: Input: default@nzhang_part14@value=__HIVE_DEFAULT_PARTITION__ #### A masked pattern was here #### POSTHOOK: query: select * from nzhang_part14 where value <> 'a' order by key, value POSTHOOK: type: QUERY POSTHOOK: Input: default@nzhang_part14 -POSTHOOK: Input: default@nzhang_part14@value=%20 +POSTHOOK: Input: default@nzhang_part14@value=%2520 POSTHOOK: Input: default@nzhang_part14@value=__HIVE_DEFAULT_PARTITION__ #### A masked pattern was here #### -POSTHOOK: Lineage: nzhang_part14 PARTITION(value= ).key EXPRESSION [] -POSTHOOK: Lineage: nzhang_part14 PARTITION(value=__HIVE_DEFAULT_PARTITION__).key EXPRESSION [] k1 __HIVE_DEFAULT_PARTITION__ k1 __HIVE_DEFAULT_PARTITION__ k2 __HIVE_DEFAULT_PARTITION__ k2 __HIVE_DEFAULT_PARTITION__ -k3 -k3 +k3 %20 +k3 %20 diff --git ql/src/test/results/clientpositive/scriptfile1_win.q.out ql/src/test/results/clientpositive/scriptfile1_win.q.out index 1f84bd3..0c9c546 100644 --- ql/src/test/results/clientpositive/scriptfile1_win.q.out +++ ql/src/test/results/clientpositive/scriptfile1_win.q.out @@ -3,6 +3,7 @@ PREHOOK: query: -- INCLUDE_OS_WINDOWS CREATE TABLE dest1(key INT, value STRING) PREHOOK: type: CREATETABLE PREHOOK: Output: database:default +PREHOOK: Output: default@dest1 POSTHOOK: query: -- INCLUDE_OS_WINDOWS CREATE TABLE dest1(key INT, value STRING) @@ -39,8 +40,6 @@ POSTHOOK: query: SELECT dest1.* FROM dest1 POSTHOOK: type: QUERY POSTHOOK: Input: default@dest1 #### A masked pattern was here #### -POSTHOOK: Lineage: dest1.key SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1.value SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ] NULL NULL NULL NULL 10 val_10