Index: hbase-handler/build.xml
===================================================================
--- hbase-handler/build.xml (revision 966856)
+++ hbase-handler/build.xml (working copy)
@@ -34,7 +34,7 @@
-
+
Index: jdbc/build.xml
===================================================================
--- jdbc/build.xml (revision 966856)
+++ jdbc/build.xml (working copy)
@@ -32,7 +32,7 @@
-
+
Index: data/warehouse/src/.gitignore
===================================================================
--- data/warehouse/src/.gitignore (revision 966856)
+++ data/warehouse/src/.gitignore (working copy)
@@ -1 +0,0 @@
-# Dummy file to make Git recognize this empty directory
Index: build-common.xml
===================================================================
--- build-common.xml (revision 966856)
+++ build-common.xml (working copy)
@@ -58,7 +58,7 @@
-
+
@@ -254,11 +254,8 @@
-
-
-
-
-
+
+
@@ -407,11 +404,11 @@
-->
-
+
-
+
Index: hwi/build.xml
===================================================================
--- hwi/build.xml (revision 966856)
+++ hwi/build.xml (working copy)
@@ -34,7 +34,7 @@
-
+
Index: service/build.xml
===================================================================
--- service/build.xml (revision 966856)
+++ service/build.xml (working copy)
@@ -33,7 +33,7 @@
-
+
Index: contrib/build.xml
===================================================================
--- contrib/build.xml (revision 966856)
+++ contrib/build.xml (working copy)
@@ -33,7 +33,7 @@
-
+
Index: odbc/build.xml
===================================================================
--- odbc/build.xml (revision 966856)
+++ odbc/build.xml (working copy)
@@ -32,7 +32,7 @@
-
+
Index: ql/src/test/results/clientnegative/ddltime.q.out
===================================================================
--- ql/src/test/results/clientnegative/ddltime.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/ddltime.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table T2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table T2
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table T2 like srcpart
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table T2 like srcpart
Index: ql/src/test/results/clientnegative/alter_non_native.q.out
===================================================================
--- ql/src/test/results/clientnegative/alter_non_native.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/alter_non_native.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE non_native1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE non_native1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE non_native1(key int, value string)
STORED BY 'org.apache.hadoop.hive.ql.metadata.DefaultStorageHandler'
PREHOOK: type: CREATETABLE
Index: ql/src/test/results/clientnegative/ctas.q.out
===================================================================
--- ql/src/test/results/clientnegative/ctas.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/ctas.q.out (working copy)
@@ -1,5 +1 @@
-PREHOOK: query: drop table nzhang_ctas4
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table nzhang_ctas4
-POSTHOOK: type: DROPTABLE
FAILED: Error in semantic analysis: CREATE-TABLE-AS-SELECT cannot create external table.
Index: ql/src/test/results/clientnegative/load_wrong_fileformat_rc_seq.q.out
===================================================================
--- ql/src/test/results/clientnegative/load_wrong_fileformat_rc_seq.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/load_wrong_fileformat_rc_seq.q.out (working copy)
@@ -1,16 +1,14 @@
PREHOOK: query: -- test for loading into tables with the correct file format
-- test for loading into partitions with the correct file format
-DROP TABLE T1
-PREHOOK: type: DROPTABLE
+
+CREATE TABLE T1(name STRING) STORED AS RCFILE
+PREHOOK: type: CREATETABLE
POSTHOOK: query: -- test for loading into tables with the correct file format
-- test for loading into partitions with the correct file format
-DROP TABLE T1
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE T1(name STRING) STORED AS RCFILE
-PREHOOK: type: CREATETABLE
-POSTHOOK: query: CREATE TABLE T1(name STRING) STORED AS RCFILE
+
+CREATE TABLE T1(name STRING) STORED AS RCFILE
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: default@T1
PREHOOK: query: LOAD DATA LOCAL INPATH '../data/files/kv1.seq' INTO TABLE T1
Index: ql/src/test/results/clientnegative/dyn_part1.q.out
===================================================================
--- ql/src/test/results/clientnegative/dyn_part1.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/dyn_part1.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table dynamic_partition
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table dynamic_partition
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table dynamic_partition (key string) partitioned by (value string)
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table dynamic_partition (key string) partitioned by (value string)
Index: ql/src/test/results/clientnegative/create_view_failure2.q.out
===================================================================
--- ql/src/test/results/clientnegative/create_view_failure2.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/create_view_failure2.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE xxx4
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE xxx4
-POSTHOOK: type: DROPTABLE
PREHOOK: query: DROP VIEW xxx4
PREHOOK: type: DROPVIEW
POSTHOOK: query: DROP VIEW xxx4
@@ -9,11 +5,11 @@
PREHOOK: query: -- views and tables share the same namespace
CREATE VIEW xxx4 AS SELECT key FROM src
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/1327466632/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-17_460_630133670553073559/10000
POSTHOOK: query: -- views and tables share the same namespace
CREATE VIEW xxx4 AS SELECT key FROM src
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/1327466632/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-17_460_630133670553073559/10000
POSTHOOK: Output: default@xxx4
PREHOOK: query: CREATE TABLE xxx4(key int)
PREHOOK: type: CREATETABLE
Index: ql/src/test/results/clientnegative/nopart_insert.q.out
===================================================================
--- ql/src/test/results/clientnegative/nopart_insert.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/nopart_insert.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE nopart_insert
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE nopart_insert
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE nopart_insert(a STRING, b STRING) PARTITIONED BY (ds STRING)
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE nopart_insert(a STRING, b STRING) PARTITIONED BY (ds STRING)
Index: ql/src/test/results/clientnegative/load_non_native.q.out
===================================================================
--- ql/src/test/results/clientnegative/load_non_native.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/load_non_native.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE non_native2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE non_native2
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE non_native2(key int, value string)
STORED BY 'org.apache.hadoop.hive.ql.metadata.DefaultStorageHandler'
PREHOOK: type: CREATETABLE
Index: ql/src/test/results/clientnegative/altern1.q.out
===================================================================
--- ql/src/test/results/clientnegative/altern1.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/altern1.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table altern1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table altern1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table altern1(a int, b int) partitioned by (ds string)
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table altern1(a int, b int) partitioned by (ds string)
Index: ql/src/test/results/clientnegative/invalidate_view1.q.out
===================================================================
--- ql/src/test/results/clientnegative/invalidate_view1.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/invalidate_view1.q.out (working copy)
@@ -6,10 +6,6 @@
PREHOOK: type: DROPVIEW
POSTHOOK: query: DROP VIEW xxx9
POSTHOOK: type: DROPVIEW
-PREHOOK: query: DROP TABLE xxx10
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE xxx10
-POSTHOOK: type: DROPTABLE
PREHOOK: query: -- create two levels of view reference, then invalidate intermediate view
-- by dropping a column from underlying table, and verify that
-- querying outermost view results in full error context
@@ -23,17 +19,17 @@
POSTHOOK: Output: default@xxx10
PREHOOK: query: CREATE VIEW xxx9 AS SELECT * FROM xxx10
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/1720348561/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-31_670_5934534815948379913/10000
POSTHOOK: query: CREATE VIEW xxx9 AS SELECT * FROM xxx10
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/1720348561/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-31_670_5934534815948379913/10000
POSTHOOK: Output: default@xxx9
PREHOOK: query: CREATE VIEW xxx8 AS SELECT * FROM xxx9 xxx
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/1891844844/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-31_706_855643045189903710/10000
POSTHOOK: query: CREATE VIEW xxx8 AS SELECT * FROM xxx9 xxx
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/1891844844/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-31_706_855643045189903710/10000
POSTHOOK: Output: default@xxx8
PREHOOK: query: ALTER TABLE xxx10 REPLACE COLUMNS (key int)
PREHOOK: type: ALTERTABLE_REPLACECOLS
Index: ql/src/test/results/clientnegative/external2.q.out
===================================================================
--- ql/src/test/results/clientnegative/external2.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/external2.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table external2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table external2
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create external table external2(a int, b int) partitioned by (ds string)
PREHOOK: type: CREATETABLE
POSTHOOK: query: create external table external2(a int, b int) partitioned by (ds string)
Index: ql/src/test/results/clientnegative/create_view_failure1.q.out
===================================================================
--- ql/src/test/results/clientnegative/create_view_failure1.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/create_view_failure1.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE xxx12
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE xxx12
-POSTHOOK: type: DROPTABLE
PREHOOK: query: DROP VIEW xxx12
PREHOOK: type: DROPVIEW
POSTHOOK: query: DROP VIEW xxx12
@@ -15,6 +11,6 @@
POSTHOOK: Output: default@xxx12
PREHOOK: query: CREATE VIEW xxx12 AS SELECT key FROM src
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/Users/jsichi/open/hive-trunk/build/ql/tmp/194996627/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-43-17_176_344624644454180304/10000
FAILED: Error in metadata: AlreadyExistsException(message:Table xxx12 already exists)
FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask
Index: ql/src/test/results/clientnegative/smb_bucketmapjoin.q.out
===================================================================
--- ql/src/test/results/clientnegative/smb_bucketmapjoin.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/smb_bucketmapjoin.q.out (working copy)
@@ -1,16 +1,8 @@
-PREHOOK: query: drop table smb_bucket4_1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table smb_bucket4_1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE smb_bucket4_1(key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE smb_bucket4_1(key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: default@smb_bucket4_1
-PREHOOK: query: drop table smb_bucket4_2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table smb_bucket4_2
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE smb_bucket4_2(key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE smb_bucket4_2(key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS
Index: ql/src/test/results/clientnegative/addpart1.q.out
===================================================================
--- ql/src/test/results/clientnegative/addpart1.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/addpart1.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table addpart1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table addpart1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table addpart1 (a int) partitioned by (b string, c string)
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table addpart1 (a int) partitioned by (b string, c string)
Index: ql/src/test/results/clientnegative/union2.q.out
===================================================================
--- ql/src/test/results/clientnegative/union2.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/union2.q.out (working copy)
@@ -1,11 +1,3 @@
-PREHOOK: query: drop table union2_t1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table union2_t1
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: drop table union2_t2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table union2_t2
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table if not exists union2_t1(r string, c string, v string)
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table if not exists union2_t1(r string, c string, v string)
Index: ql/src/test/results/clientnegative/load_wrong_fileformat_txt_seq.q.out
===================================================================
--- ql/src/test/results/clientnegative/load_wrong_fileformat_txt_seq.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/load_wrong_fileformat_txt_seq.q.out (working copy)
@@ -1,16 +1,14 @@
PREHOOK: query: -- test for loading into tables with the correct file format
-- test for loading into partitions with the correct file format
-DROP TABLE T1
-PREHOOK: type: DROPTABLE
+
+CREATE TABLE T1(name STRING) STORED AS TEXTFILE
+PREHOOK: type: CREATETABLE
POSTHOOK: query: -- test for loading into tables with the correct file format
-- test for loading into partitions with the correct file format
-DROP TABLE T1
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE T1(name STRING) STORED AS TEXTFILE
-PREHOOK: type: CREATETABLE
-POSTHOOK: query: CREATE TABLE T1(name STRING) STORED AS TEXTFILE
+
+CREATE TABLE T1(name STRING) STORED AS TEXTFILE
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: default@T1
PREHOOK: query: LOAD DATA LOCAL INPATH '../data/files/kv1.seq' INTO TABLE T1
Index: ql/src/test/results/clientnegative/drop_view_failure1.q.out
===================================================================
--- ql/src/test/results/clientnegative/drop_view_failure1.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/drop_view_failure1.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE xxx1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE xxx1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE xxx1(key int)
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE xxx1(key int)
Index: ql/src/test/results/clientnegative/invalid_create_tbl1.q.out
===================================================================
--- ql/src/test/results/clientnegative/invalid_create_tbl1.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/invalid_create_tbl1.q.out (working copy)
@@ -1,5 +1 @@
-PREHOOK: query: DROP TABLE inv_valid_tbl1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE inv_valid_tbl1
-POSTHOOK: type: DROPTABLE
FAILED: Error in semantic analysis: DATE, DATETIME, and TIMESTAMP types aren't supported yet. Please use STRING instead.
Index: ql/src/test/results/clientnegative/deletejar.q.out
===================================================================
--- ql/src/test/results/clientnegative/deletejar.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/deletejar.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE DELETEJAR
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE DELETEJAR
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE DELETEJAR(KEY STRING, VALUE STRING) ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.TestSerDe' STORED AS TEXTFILE
PREHOOK: type: CREATETABLE
FAILED: Error in metadata: Cannot validate serde: org.apache.hadoop.hive.serde2.TestSerDe
Index: ql/src/test/results/clientnegative/external1.q.out
===================================================================
--- ql/src/test/results/clientnegative/external1.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/external1.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table external1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table external1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create external table external1(a int, b int) location 'invalidscheme://data.s3ndemo.hive/kv'
PREHOOK: type: CREATETABLE
FAILED: Error in metadata: MetaException(message:Got exception: java.io.IOException No FileSystem for scheme: invalidscheme)
Index: ql/src/test/results/clientnegative/load_wrong_fileformat.q.out
===================================================================
--- ql/src/test/results/clientnegative/load_wrong_fileformat.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/load_wrong_fileformat.q.out (working copy)
@@ -1,16 +1,14 @@
PREHOOK: query: -- test for loading into tables with the correct file format
-- test for loading into partitions with the correct file format
-DROP TABLE load_wrong_fileformat_T1
-PREHOOK: type: DROPTABLE
+
+CREATE TABLE load_wrong_fileformat_T1(name STRING) STORED AS SEQUENCEFILE
+PREHOOK: type: CREATETABLE
POSTHOOK: query: -- test for loading into tables with the correct file format
-- test for loading into partitions with the correct file format
-DROP TABLE load_wrong_fileformat_T1
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE load_wrong_fileformat_T1(name STRING) STORED AS SEQUENCEFILE
-PREHOOK: type: CREATETABLE
-POSTHOOK: query: CREATE TABLE load_wrong_fileformat_T1(name STRING) STORED AS SEQUENCEFILE
+
+CREATE TABLE load_wrong_fileformat_T1(name STRING) STORED AS SEQUENCEFILE
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: default@load_wrong_fileformat_T1
PREHOOK: query: LOAD DATA LOCAL INPATH '../data/files/kv1.txt' INTO TABLE load_wrong_fileformat_T1
Index: ql/src/test/results/clientnegative/create_insert_outputformat.q.out
===================================================================
--- ql/src/test/results/clientnegative/create_insert_outputformat.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/create_insert_outputformat.q.out (working copy)
@@ -1,5 +1 @@
-PREHOOK: query: DROP TABLE table_test_output_fomat
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE table_test_output_fomat
-POSTHOOK: type: DROPTABLE
FAILED: Error in semantic analysis: Output Format must implement HiveOutputFormat, otherwise it should be either IgnoreKeyTextOutputFormat or SequenceFileOutputFormat
Index: ql/src/test/results/clientnegative/nopart_load.q.out
===================================================================
--- ql/src/test/results/clientnegative/nopart_load.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/nopart_load.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE nopart_load
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE nopart_load
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE nopart_load(a STRING, b STRING) PARTITIONED BY (ds STRING)
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE nopart_load(a STRING, b STRING) PARTITIONED BY (ds STRING)
Index: ql/src/test/results/clientnegative/dyn_part2.q.out
===================================================================
--- ql/src/test/results/clientnegative/dyn_part2.q.out (revision 966856)
+++ ql/src/test/results/clientnegative/dyn_part2.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table nzhang_part1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table nzhang_part1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table nzhang_part1 (key string, value string) partitioned by (ds string, hr string)
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table nzhang_part1 (key string, value string) partitioned by (ds string, hr string)
Index: ql/src/test/results/clientpositive/inputddl8.q.out
===================================================================
--- ql/src/test/results/clientpositive/inputddl8.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/inputddl8.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE INPUTDDL8
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE INPUTDDL8
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE INPUTDDL8 COMMENT 'This is a thrift based table'
PARTITIONED BY(ds STRING, country STRING)
CLUSTERED BY(aint) SORTED BY(lint) INTO 32 BUCKETS
@@ -32,9 +28,4 @@
ds string
country string
-Detailed Table Information Table(tableName:inputddl8, dbName:default, owner:pyang, createTime:1264209638, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[], location:file:/data/users/pyang/task2/trunk/VENDOR.hive/trunk/build/ql/test/data/warehouse/inputddl8, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:32, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.thrift.ThriftDeserializer, parameters:{serialization.class=org.apache.hadoop.hive.serde2.thrift.test.Complex,serialization.format=com.facebook.thrift.protocol.TBinaryProtocol}), bucketCols:[aint], sortCols:[Order(col:lint, order:1)], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null), FieldSchema(name:country, type:string, comment:null)], parameters:{transient_lastDdlTime=1264209638,comment=This is a thrift based table}, viewOriginalText:null, viewExpandedText:null)
-PREHOOK: query: DROP TABLE INPUTDDL8
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE INPUTDDL8
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@inputddl8
+Detailed Table Information Table(tableName:inputddl8, dbName:default, owner:jssarma, createTime:1279737072, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[], location:file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/inputddl8, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:32, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.thrift.ThriftDeserializer, parameters:{serialization.class=org.apache.hadoop.hive.serde2.thrift.test.Complex, serialization.format=com.facebook.thrift.protocol.TBinaryProtocol}), bucketCols:[aint], sortCols:[Order(col:lint, order:1)], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null), FieldSchema(name:country, type:string, comment:null)], parameters:{transient_lastDdlTime=1279737072, comment=This is a thrift based table}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)
Index: ql/src/test/results/clientpositive/union19.q.out
===================================================================
--- ql/src/test/results/clientpositive/union19.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/union19.q.out (working copy)
@@ -1,11 +1,3 @@
-PREHOOK: query: drop table DEST1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table DEST1
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: drop table DEST2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table DEST2
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE DEST1(key STRING, value STRING) STORED AS TEXTFILE
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE DEST1(key STRING, value STRING) STORED AS TEXTFILE
@@ -88,7 +80,7 @@
Stage: Stage-3
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-37-10_956_8209072529967031935/10004
+ file:/tmp/jssarma/hive_2010-07-21_13-40-25_638_6209859465363547113/10004
Union
Select Operator
expressions:
@@ -135,7 +127,7 @@
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest2
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-37-10_956_8209072529967031935/10005
+ file:/tmp/jssarma/hive_2010-07-21_13-40-25_638_6209859465363547113/10005
Union
Select Operator
expressions:
@@ -275,11 +267,11 @@
PREHOOK: query: SELECT DEST1.* FROM DEST1 SORT BY DEST1.key, DEST1.value
PREHOOK: type: QUERY
PREHOOK: Input: default@dest1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-37-21_983_2907581159747186721/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-40-33_529_5679096718849362367/10000
POSTHOOK: query: SELECT DEST1.* FROM DEST1 SORT BY DEST1.key, DEST1.value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-37-21_983_2907581159747186721/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-40-33_529_5679096718849362367/10000
POSTHOOK: Lineage: dest1.key EXPRESSION [(src)s2.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.value EXPRESSION [(src)s1.null, (src)s2.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest2.key EXPRESSION [(src)s2.FieldSchema(name:key, type:string, comment:default), ]
@@ -598,11 +590,11 @@
PREHOOK: query: SELECT DEST2.* FROM DEST2 SORT BY DEST2.key, DEST2.val1, DEST2.val2
PREHOOK: type: QUERY
PREHOOK: Input: default@dest2
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-37-25_452_88250716044597875/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-40-36_079_8511341470250811242/10000
POSTHOOK: query: SELECT DEST2.* FROM DEST2 SORT BY DEST2.key, DEST2.val1, DEST2.val2
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest2
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-37-25_452_88250716044597875/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-40-36_079_8511341470250811242/10000
POSTHOOK: Lineage: dest1.key EXPRESSION [(src)s2.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.value EXPRESSION [(src)s1.null, (src)s2.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest2.key EXPRESSION [(src)s2.FieldSchema(name:key, type:string, comment:default), ]
@@ -1109,23 +1101,3 @@
98 val_98 val_98
98 val_98 val_98
tst1 500 500
-PREHOOK: query: drop table DEST1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table DEST1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@dest1
-POSTHOOK: Lineage: dest1.key EXPRESSION [(src)s2.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.value EXPRESSION [(src)s1.null, (src)s2.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest2.key EXPRESSION [(src)s2.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: dest2.val1 EXPRESSION [(src)s1.null, (src)s2.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest2.val2 EXPRESSION [(src)s1.null, (src)s2.FieldSchema(name:value, type:string, comment:default), ]
-PREHOOK: query: drop table DEST2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table DEST2
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@dest2
-POSTHOOK: Lineage: dest1.key EXPRESSION [(src)s2.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.value EXPRESSION [(src)s1.null, (src)s2.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest2.key EXPRESSION [(src)s2.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: dest2.val1 EXPRESSION [(src)s1.null, (src)s2.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest2.val2 EXPRESSION [(src)s1.null, (src)s2.FieldSchema(name:value, type:string, comment:default), ]
Index: ql/src/test/results/clientpositive/input32.q.out
===================================================================
--- ql/src/test/results/clientpositive/input32.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/input32.q.out (working copy)
@@ -1,11 +1,3 @@
-PREHOOK: query: drop table tst_dest32
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table tst_dest32
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: drop table dest32
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table dest32
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table dest32(a int)
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table dest32(a int)
@@ -102,22 +94,10 @@
PREHOOK: query: select * from tst_dest32
PREHOOK: type: QUERY
PREHOOK: Input: default@tst_dest32
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-12-44_362_8153903898810014347/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-27-54_520_2654638815572437304/10000
POSTHOOK: query: select * from tst_dest32
POSTHOOK: type: QUERY
POSTHOOK: Input: default@tst_dest32
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-12-44_362_8153903898810014347/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-27-54_520_2654638815572437304/10000
POSTHOOK: Lineage: tst_dest32.a EXPRESSION [(srcbucket)srcbucket.null, ]
1000
-PREHOOK: query: drop table tst_dest32
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table tst_dest32
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@tst_dest32
-POSTHOOK: Lineage: tst_dest32.a EXPRESSION [(srcbucket)srcbucket.null, ]
-PREHOOK: query: drop table dest32
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table dest32
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@dest32
-POSTHOOK: Lineage: tst_dest32.a EXPRESSION [(srcbucket)srcbucket.null, ]
Index: ql/src/test/results/clientpositive/union4.q.out
===================================================================
--- ql/src/test/results/clientpositive/union4.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/union4.q.out (working copy)
@@ -1,14 +1,12 @@
PREHOOK: query: -- union case: both subqueries are map-reduce jobs on same input, followed by filesink
-drop table tmptable
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: -- union case: both subqueries are map-reduce jobs on same input, followed by filesink
-drop table tmptable
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: create table tmptable(key string, value int)
+create table tmptable(key string, value int)
PREHOOK: type: CREATETABLE
-POSTHOOK: query: create table tmptable(key string, value int)
+POSTHOOK: query: -- union case: both subqueries are map-reduce jobs on same input, followed by filesink
+
+
+create table tmptable(key string, value int)
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: default@tmptable
PREHOOK: query: explain
@@ -79,7 +77,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-43-02_028_7405143269986564685/10002
+ file:/tmp/jssarma/hive_2010-07-21_13-41-40_681_7920974237145003540/10002
Union
Select Operator
expressions:
@@ -103,7 +101,7 @@
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: tmptable
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-43-02_028_7405143269986564685/10004
+ file:/tmp/jssarma/hive_2010-07-21_13-41-40_681_7920974237145003540/10004
Union
Select Operator
expressions:
@@ -135,7 +133,7 @@
Move Operator
files:
hdfs directory: true
- destination: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-43-02_028_7405143269986564685/10000
+ destination: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_13-41-40_681_7920974237145003540/10000
Stage: Stage-0
Move Operator
@@ -150,7 +148,7 @@
Stage: Stage-3
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-43-02_028_7405143269986564685/10003
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_13-41-40_681_7920974237145003540/10003
Reduce Output Operator
sort order:
Map-reduce partition columns:
@@ -233,19 +231,12 @@
PREHOOK: query: select * from tmptable x sort by x.key
PREHOOK: type: QUERY
PREHOOK: Input: default@tmptable
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-43-14_790_3647521945659456151/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-41-50_925_942224749030960902/10000
POSTHOOK: query: select * from tmptable x sort by x.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@tmptable
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-43-14_790_3647521945659456151/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_13-41-50_925_942224749030960902/10000
POSTHOOK: Lineage: tmptable.key EXPRESSION []
POSTHOOK: Lineage: tmptable.value EXPRESSION [(src)s1.null, (src)s2.null, ]
tst1 500
tst2 500
-PREHOOK: query: drop table tmptable
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table tmptable
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@tmptable
-POSTHOOK: Lineage: tmptable.key EXPRESSION []
-POSTHOOK: Lineage: tmptable.value EXPRESSION [(src)s1.null, (src)s2.null, ]
Index: ql/src/test/results/clientpositive/groupby9.q.out
===================================================================
--- ql/src/test/results/clientpositive/groupby9.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/groupby9.q.out (working copy)
@@ -1,11 +1,3 @@
-PREHOOK: query: drop table DEST1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table DEST1
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: drop table DEST2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table DEST2
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE DEST1(key INT, value STRING) STORED AS TEXTFILE
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE DEST1(key INT, value STRING) STORED AS TEXTFILE
@@ -95,7 +87,7 @@
Stage: Stage-3
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-09-55_073_4811751815642872862/10004
+ file:/tmp/jssarma/hive_2010-07-21_11-25-41_473_5912399545173757449/10004
Reduce Output Operator
key expressions:
expr: _col0
@@ -154,7 +146,7 @@
Stage: Stage-4
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-09-55_073_4811751815642872862/10005
+ file:/tmp/jssarma/hive_2010-07-21_11-25-41_473_5912399545173757449/10005
Reduce Output Operator
key expressions:
expr: _col0
@@ -243,11 +235,11 @@
PREHOOK: query: SELECT DEST1.* FROM DEST1
PREHOOK: type: QUERY
PREHOOK: Input: default@dest1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-10-08_620_486578954348947231/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-25-49_248_6786053973649103964/10000
POSTHOOK: query: SELECT DEST1.* FROM DEST1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-10-08_620_486578954348947231/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-25-49_248_6786053973649103964/10000
POSTHOOK: Lineage: dest1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.value EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest2.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -565,11 +557,11 @@
PREHOOK: query: SELECT DEST2.* FROM DEST2
PREHOOK: type: QUERY
PREHOOK: Input: default@dest2
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-10-08_680_6077462527451103689/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-25-49_307_6015579282653689221/10000
POSTHOOK: query: SELECT DEST2.* FROM DEST2
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest2
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-10-08_680_6077462527451103689/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-25-49_307_6015579282653689221/10000
POSTHOOK: Lineage: dest1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.value EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest2.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -884,23 +876,3 @@
96 val_96 1
97 val_97 1
98 val_98 1
-PREHOOK: query: drop table DEST1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table DEST1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@dest1
-POSTHOOK: Lineage: dest1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.value EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest2.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: dest2.val1 SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest2.val2 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-PREHOOK: query: drop table DEST2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table DEST2
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@dest2
-POSTHOOK: Lineage: dest1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.value EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest2.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: dest2.val1 SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest2.val2 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
Index: ql/src/test/results/clientpositive/insert1.q.out
===================================================================
--- ql/src/test/results/clientpositive/insert1.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/insert1.q.out (working copy)
@@ -1,11 +1,3 @@
-PREHOOK: query: drop table insert1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table insert1
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: drop table insert2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table insert2
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table insert1(key int, value string) stored as textfile
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table insert1(key int, value string) stored as textfile
@@ -26,17 +18,3 @@
POSTHOOK: Output: default@insert1
POSTHOOK: Lineage: insert1.key SIMPLE [(insert2)a.FieldSchema(name:key, type:int, comment:null), ]
POSTHOOK: Lineage: insert1.value SIMPLE [(insert2)a.FieldSchema(name:value, type:string, comment:null), ]
-PREHOOK: query: drop table insert1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table insert1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@insert1
-POSTHOOK: Lineage: insert1.key SIMPLE [(insert2)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: insert1.value SIMPLE [(insert2)a.FieldSchema(name:value, type:string, comment:null), ]
-PREHOOK: query: drop table insert2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table insert2
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@insert2
-POSTHOOK: Lineage: insert1.key SIMPLE [(insert2)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: insert1.value SIMPLE [(insert2)a.FieldSchema(name:value, type:string, comment:null), ]
Index: ql/src/test/results/clientpositive/input41.q.out
===================================================================
--- ql/src/test/results/clientpositive/input41.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/input41.q.out (working copy)
@@ -25,17 +25,11 @@
PREHOOK: query: select * from dest_sp x order by x.cnt limit 2
PREHOOK: type: QUERY
PREHOOK: Input: default@dest_sp
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-15-40_207_9006053612393059372/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-28-55_506_7975559281496695000/10000
POSTHOOK: query: select * from dest_sp x order by x.cnt limit 2
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest_sp
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-15-40_207_9006053612393059372/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-28-55_506_7975559281496695000/10000
POSTHOOK: Lineage: dest_sp.cnt EXPRESSION [(src)src.null, (srcpart)srcpart.null, ]
0
500
-PREHOOK: query: drop table dest_sp
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table dest_sp
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@dest_sp
-POSTHOOK: Lineage: dest_sp.cnt EXPRESSION [(src)src.null, (srcpart)srcpart.null, ]
Index: ql/src/test/results/clientpositive/load_dyn_part7.q.out
===================================================================
--- ql/src/test/results/clientpositive/load_dyn_part7.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/load_dyn_part7.q.out (working copy)
@@ -6,10 +6,6 @@
ds=2008-04-08/hr=12
ds=2008-04-09/hr=11
ds=2008-04-09/hr=12
-PREHOOK: query: drop table nzhang_part7
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table nzhang_part7
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table if not exists nzhang_part7 like srcpart
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table if not exists nzhang_part7 like srcpart
@@ -24,7 +20,7 @@
ds string
hr string
-Detailed Table Information Table(tableName:nzhang_part7, dbName:default, owner:null, createTime:1271267706, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:default), FieldSchema(name:value, type:string, comment:default)], location:file:/data/users/nzhang/work/876/apache-hive/build/ql/test/data/warehouse/nzhang_part7, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null), FieldSchema(name:hr, type:string, comment:null)], parameters:{EXTERNAL=FALSE,transient_lastDdlTime=1271267706}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)
+Detailed Table Information Table(tableName:nzhang_part7, dbName:default, owner:null, createTime:1279737594, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:default), FieldSchema(name:value, type:string, comment:default)], location:file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/nzhang_part7, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null), FieldSchema(name:hr, type:string, comment:null)], parameters:{EXTERNAL=FALSE, transient_lastDdlTime=1279737594}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)
PREHOOK: query: insert overwrite table nzhang_part7 partition (ds='2010-03-03', hr='12') select key, value from srcpart where ds = '2008-04-08' and hr = '12'
PREHOOK: type: QUERY
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
@@ -45,11 +41,11 @@
PREHOOK: query: select * from nzhang_part7 where ds is not null and hr is not null
PREHOOK: type: QUERY
PREHOOK: Input: default@nzhang_part7@ds=2010-03-03/hr=12
-PREHOOK: Output: file:/data/users/nzhang/work/876/apache-hive/build/ql/scratchdir/hive_2010-04-14_10-55-10_491_3925213404431690499/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-39-57_928_4887604849335253138/10000
POSTHOOK: query: select * from nzhang_part7 where ds is not null and hr is not null
POSTHOOK: type: QUERY
POSTHOOK: Input: default@nzhang_part7@ds=2010-03-03/hr=12
-POSTHOOK: Output: file:/data/users/nzhang/work/876/apache-hive/build/ql/scratchdir/hive_2010-04-14_10-55-10_491_3925213404431690499/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-39-57_928_4887604849335253138/10000
POSTHOOK: Lineage: nzhang_part7 PARTITION(ds=2010-03-03,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ]
POSTHOOK: Lineage: nzhang_part7 PARTITION(ds=2010-03-03,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:hr, type:string, comment:null), ]
238 val_238 2010-03-03 12
@@ -552,10 +548,3 @@
400 val_400 2010-03-03 12
200 val_200 2010-03-03 12
97 val_97 2010-03-03 12
-PREHOOK: query: drop table nzhang_part7
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table nzhang_part7
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@nzhang_part7
-POSTHOOK: Lineage: nzhang_part7 PARTITION(ds=2010-03-03,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ]
-POSTHOOK: Lineage: nzhang_part7 PARTITION(ds=2010-03-03,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:hr, type:string, comment:null), ]
Index: ql/src/test/results/clientpositive/inputddl3.q.out
===================================================================
--- ql/src/test/results/clientpositive/inputddl3.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/inputddl3.q.out (working copy)
@@ -35,8 +35,3 @@
POSTHOOK: type: DESCTABLE
key int
value string
-PREHOOK: query: DROP TABLE INPUTDDL3
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE INPUTDDL3
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@inputddl3
Index: ql/src/test/results/clientpositive/join29.q.out
===================================================================
--- ql/src/test/results/clientpositive/join29.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/join29.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: drop TABLE dest_j1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop TABLE dest_j1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE dest_j1(key STRING, cnt1 INT, cnt2 INT)
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE dest_j1(key STRING, cnt1 INT, cnt2 INT)
@@ -91,7 +87,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-15-39_938_6824990071701609235/10002
+ file:/tmp/jssarma/hive_2010-07-21_11-32-59_660_3940646832112023058/10002
Common Join Operator
condition map:
Inner Join 0 to 1
@@ -142,11 +138,11 @@
Local Work:
Map Reduce Local Work
Alias -> Map Local Tables:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-15-39_938_6824990071701609235/10004
+ file:/tmp/jssarma/hive_2010-07-21_11-32-59_660_3940646832112023058/10004
Fetch Operator
limit: -1
Alias -> Map Local Operator Tree:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-15-39_938_6824990071701609235/10004
+ file:/tmp/jssarma/hive_2010-07-21_11-32-59_660_3940646832112023058/10004
Common Join Operator
condition map:
Inner Join 0 to 1
@@ -202,7 +198,7 @@
Move Operator
files:
hdfs directory: true
- destination: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-15-39_938_6824990071701609235/10000
+ destination: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-32-59_660_3940646832112023058/10000
Stage: Stage-0
Move Operator
@@ -217,7 +213,7 @@
Stage: Stage-3
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-15-39_938_6824990071701609235/10003
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-32-59_660_3940646832112023058/10003
Reduce Output Operator
sort order:
Map-reduce partition columns:
@@ -321,11 +317,11 @@
PREHOOK: query: select * from dest_j1 x order by x.key
PREHOOK: type: QUERY
PREHOOK: Input: default@dest_j1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-15-53_522_4492151849840756172/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-33-08_454_5846290866810953063/10000
POSTHOOK: query: select * from dest_j1 x order by x.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest_j1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-15-53_522_4492151849840756172/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-33-08_454_5846290866810953063/10000
POSTHOOK: Lineage: dest_j1.cnt1 EXPRESSION [(src1)x.null, ]
POSTHOOK: Lineage: dest_j1.cnt2 EXPRESSION [(src)y.null, ]
POSTHOOK: Lineage: dest_j1.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ]
@@ -344,11 +340,3 @@
406 1 4
66 1 1
98 1 2
-PREHOOK: query: drop TABLE dest_j1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop TABLE dest_j1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@dest_j1
-POSTHOOK: Lineage: dest_j1.cnt1 EXPRESSION [(src1)x.null, ]
-POSTHOOK: Lineage: dest_j1.cnt2 EXPRESSION [(src)y.null, ]
-POSTHOOK: Lineage: dest_j1.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ]
Index: ql/src/test/results/clientpositive/load_dyn_part13.q.out
===================================================================
--- ql/src/test/results/clientpositive/load_dyn_part13.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/load_dyn_part13.q.out (working copy)
@@ -6,10 +6,6 @@
ds=2008-04-08/hr=12
ds=2008-04-09/hr=11
ds=2008-04-09/hr=12
-PREHOOK: query: drop table nzhang_part13
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table nzhang_part13
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table if not exists nzhang_part13 like srcpart
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table if not exists nzhang_part13 like srcpart
@@ -24,7 +20,7 @@
ds string
hr string
-Detailed Table Information Table(tableName:nzhang_part13, dbName:default, owner:null, createTime:1271267587, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:default), FieldSchema(name:value, type:string, comment:default)], location:file:/data/users/nzhang/work/876/apache-hive/build/ql/test/data/warehouse/nzhang_part13, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null), FieldSchema(name:hr, type:string, comment:null)], parameters:{EXTERNAL=FALSE,transient_lastDdlTime=1271267587}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)
+Detailed Table Information Table(tableName:nzhang_part13, dbName:default, owner:null, createTime:1279737513, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:default), FieldSchema(name:value, type:string, comment:default)], location:file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/nzhang_part13, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null), FieldSchema(name:hr, type:string, comment:null)], parameters:{EXTERNAL=FALSE, transient_lastDdlTime=1279737513}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)
PREHOOK: query: explain
insert overwrite table nzhang_part13 partition (ds="2010-03-03", hr)
select * from (
@@ -191,12 +187,12 @@
PREHOOK: type: QUERY
PREHOOK: Input: default@nzhang_part13@ds=2010-03-03/hr=22
PREHOOK: Input: default@nzhang_part13@ds=2010-03-03/hr=33
-PREHOOK: Output: file:/data/users/nzhang/work/876/apache-hive/build/ql/scratchdir/hive_2010-04-14_10-53-12_256_3141269474960312939/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-38-36_782_8524700985752178497/10000
POSTHOOK: query: select * from nzhang_part13 where ds is not null and hr is not null
POSTHOOK: type: QUERY
POSTHOOK: Input: default@nzhang_part13@ds=2010-03-03/hr=22
POSTHOOK: Input: default@nzhang_part13@ds=2010-03-03/hr=33
-POSTHOOK: Output: file:/data/users/nzhang/work/876/apache-hive/build/ql/scratchdir/hive_2010-04-14_10-53-12_256_3141269474960312939/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-38-36_782_8524700985752178497/10000
POSTHOOK: Lineage: nzhang_part13 PARTITION(ds=2010-03-03,hr=22).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: nzhang_part13 PARTITION(ds=2010-03-03,hr=22).value EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: nzhang_part13 PARTITION(ds=2010-03-03,hr=33).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -235,12 +231,3 @@
26 val_26 2010-03-03 33
28 val_28 2010-03-03 33
37 val_37 2010-03-03 33
-PREHOOK: query: drop table nzhang_part13
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table nzhang_part13
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@nzhang_part13
-POSTHOOK: Lineage: nzhang_part13 PARTITION(ds=2010-03-03,hr=22).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: nzhang_part13 PARTITION(ds=2010-03-03,hr=22).value EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: nzhang_part13 PARTITION(ds=2010-03-03,hr=33).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: nzhang_part13 PARTITION(ds=2010-03-03,hr=33).value EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
Index: ql/src/test/results/clientpositive/loadpart1.q.out
===================================================================
--- ql/src/test/results/clientpositive/loadpart1.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/loadpart1.q.out (working copy)
@@ -1,11 +1,3 @@
-PREHOOK: query: drop table hive_test_src
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table hive_test_src
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: drop table hive_test_dst
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table hive_test_dst
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table hive_test_src ( col1 string ) stored as textfile
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table hive_test_src ( col1 string ) stored as textfile
@@ -33,11 +25,11 @@
PREHOOK: query: select * from hive_test_dst where pcol1='test_part' and pcol2='test_Part'
PREHOOK: type: QUERY
PREHOOK: Input: default@hive_test_dst@pcol1=test_part/pcol2=test_Part
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-18-27_161_4058867181423672724/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-40-08_582_5840765576374175576/10000
POSTHOOK: query: select * from hive_test_dst where pcol1='test_part' and pcol2='test_Part'
POSTHOOK: type: QUERY
POSTHOOK: Input: default@hive_test_dst@pcol1=test_part/pcol2=test_Part
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-18-27_161_4058867181423672724/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-40-08_582_5840765576374175576/10000
POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
1 test_part test_Part
2 test_part test_Part
@@ -57,20 +49,20 @@
POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
PREHOOK: query: select * from hive_test_dst where pcol1='test_part' and pcol2='test_part'
PREHOOK: type: QUERY
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-18-32_005_7262990256756311796/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-40-11_329_4921545927597783630/10000
POSTHOOK: query: select * from hive_test_dst where pcol1='test_part' and pcol2='test_part'
POSTHOOK: type: QUERY
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-18-32_005_7262990256756311796/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-40-11_329_4921545927597783630/10000
POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
PREHOOK: query: select * from hive_test_dst where pcol1='test_part'
PREHOOK: type: QUERY
PREHOOK: Input: default@hive_test_dst@pcol1=test_part/pcol2=test_Part
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-18-32_055_1756029915176080993/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-40-11_375_7311986483705850817/10000
POSTHOOK: query: select * from hive_test_dst where pcol1='test_part'
POSTHOOK: type: QUERY
POSTHOOK: Input: default@hive_test_dst@pcol1=test_part/pcol2=test_Part
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-18-32_055_1756029915176080993/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-40-11_375_7311986483705850817/10000
POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
1 test_part test_Part
@@ -81,31 +73,17 @@
6 test_part test_Part
PREHOOK: query: select * from hive_test_dst where pcol1='test_part' and pcol2='test_part'
PREHOOK: type: QUERY
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-18-32_130_4753679815952473777/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-40-11_440_6906366422294316277/10000
POSTHOOK: query: select * from hive_test_dst where pcol1='test_part' and pcol2='test_part'
POSTHOOK: type: QUERY
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-18-32_130_4753679815952473777/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-40-11_440_6906366422294316277/10000
POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
PREHOOK: query: select * from hive_test_dst where pcol1='test_Part'
PREHOOK: type: QUERY
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-18-32_181_3514948387985518789/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-40-11_474_301584563410557410/10000
POSTHOOK: query: select * from hive_test_dst where pcol1='test_Part'
POSTHOOK: type: QUERY
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-18-32_181_3514948387985518789/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-40-11_474_301584563410557410/10000
POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
-PREHOOK: query: drop table hive_test_src
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table hive_test_src
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@hive_test_src
-POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
-POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
-PREHOOK: query: drop table hive_test_dst
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table hive_test_dst
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@hive_test_dst
-POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
-POSTHOOK: Lineage: hive_test_dst PARTITION(pcol1=test_part,pcol2=test_Part).col1 SIMPLE [(hive_test_src)hive_test_src.FieldSchema(name:col1, type:string, comment:null), ]
Index: ql/src/test/results/clientpositive/binarysortable_1.q.out
===================================================================
--- ql/src/test/results/clientpositive/binarysortable_1.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/binarysortable_1.q.out (working copy)
@@ -116,7 +116,7 @@
) a
PREHOOK: type: QUERY
PREHOOK: Input: default@mytable
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1914197516/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-03_011_4547908398997924956/10000
POSTHOOK: query: SELECT REGEXP_REPLACE(REGEXP_REPLACE(REGEXP_REPLACE(key, '\001', '^A'), '\0', '^@'), '\002', '^B'), value
FROM (
SELECT key, sum(value) as value
@@ -125,7 +125,7 @@
) a
POSTHOOK: type: QUERY
POSTHOOK: Input: default@mytable
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1914197516/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-03_011_4547908398997924956/10000
^@^@^@ 7.0
^@^A^@ 9.0
^@test^@ 2.0
@@ -136,8 +136,3 @@
test^@^@^A^Atest 6.0
test^@test 4.0
test^Atest 5.0
-PREHOOK: query: DROP TABLE mytable
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE mytable
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@mytable
Index: ql/src/test/results/clientpositive/show_tables.q.out
===================================================================
--- ql/src/test/results/clientpositive/show_tables.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/show_tables.q.out (working copy)
@@ -68,13 +68,3 @@
POSTHOOK: type: SHOWTABLES
shtb_test1
shtb_test2
-PREHOOK: query: DROP TABLE shtb_test1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE shtb_test1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@shtb_test1
-PREHOOK: query: DROP TABLE shtb_test2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE shtb_test2
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@shtb_test2
Index: ql/src/test/results/clientpositive/join38.q.out
===================================================================
--- ql/src/test/results/clientpositive/join38.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/join38.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table tmp
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table tmp
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table tmp(col0 string, col1 string,col2 string,col3 string,col4 string,col5 string,col6 string,col7 string,col8 string,col9 string,col10 string,col11 string)
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table tmp(col0 string, col1 string,col2 string,col3 string,col4 string,col5 string,col6 string,col7 string,col8 string,col9 string,col10 string,col11 string)
@@ -30,11 +26,11 @@
PREHOOK: query: select * from tmp
PREHOOK: type: QUERY
PREHOOK: Input: default@tmp
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-16-41_890_1216587155753076086/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-34-29_174_1856986099030897576/10000
POSTHOOK: query: select * from tmp
POSTHOOK: type: QUERY
POSTHOOK: Input: default@tmp
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-16-41_890_1216587155753076086/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-34-29_174_1856986099030897576/10000
POSTHOOK: Lineage: tmp.col0 SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: tmp.col1 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: tmp.col10 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -142,7 +138,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-16-41_968_6961135931629257781/10002
+ file:/tmp/jssarma/hive_2010-07-21_11-34-29_228_6035667266137509751/10002
Select Operator
expressions:
expr: _col1
@@ -230,7 +226,7 @@
PREHOOK: type: QUERY
PREHOOK: Input: default@tmp
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-16-42_398_5553003257697465955/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-34-29_328_1884952246433236749/10000
POSTHOOK: query: FROM src a JOIN tmp b ON (a.key = b.col11)
SELECT /*+ MAPJOIN(a) */ a.value, b.col5, count(1) as count
where b.col11 = 111
@@ -238,7 +234,7 @@
POSTHOOK: type: QUERY
POSTHOOK: Input: default@tmp
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-16-42_398_5553003257697465955/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-34-29_328_1884952246433236749/10000
POSTHOOK: Lineage: tmp.col0 SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: tmp.col1 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: tmp.col10 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
@@ -252,20 +248,3 @@
POSTHOOK: Lineage: tmp.col8 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: tmp.col9 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
val_111 105 2
-PREHOOK: query: drop table tmp
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table tmp
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@tmp
-POSTHOOK: Lineage: tmp.col0 SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: tmp.col1 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: tmp.col10 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: tmp.col11 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: tmp.col2 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: tmp.col3 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: tmp.col4 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: tmp.col5 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: tmp.col6 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: tmp.col7 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: tmp.col8 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: tmp.col9 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
Index: ql/src/test/results/clientpositive/bucketmapjoin5.q.out
===================================================================
--- ql/src/test/results/clientpositive/bucketmapjoin5.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/bucketmapjoin5.q.out (working copy)
@@ -1,27 +1,3 @@
-PREHOOK: query: drop table bucketmapjoin_hash_result_2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table bucketmapjoin_hash_result_2
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: drop table bucketmapjoin_hash_result_1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table bucketmapjoin_hash_result_1
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: drop table bucketmapjoin_tmp_result
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table bucketmapjoin_tmp_result
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: drop table srcbucket_mapjoin
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table srcbucket_mapjoin
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: drop table srcbucket_mapjoin_part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table srcbucket_mapjoin_part
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: drop table srcbucket_mapjoin_part_2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table srcbucket_mapjoin_part_2
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE srcbucket_mapjoin(key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE srcbucket_mapjoin(key int, value string) CLUSTERED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE
@@ -184,7 +160,7 @@
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-14-40_283_968548719978651164/10002
+ directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-27_050_3859053918333083391/10002
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -195,12 +171,12 @@
columns.types string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
name bucketmapjoin_tmp_result
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584880
+ transient_lastDdlTime 1279736127
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucketmapjoin_tmp_result
TotalFiles: 1
@@ -248,7 +224,7 @@
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-14-40_283_968548719978651164/10002
+ directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-27_050_3859053918333083391/10002
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -259,12 +235,12 @@
columns.types string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
name bucketmapjoin_tmp_result
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584880
+ transient_lastDdlTime 1279736127
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucketmapjoin_tmp_result
TotalFiles: 1
@@ -273,22 +249,22 @@
Alias Bucket Base File Name Mapping:
a {srcbucket20.txt=[srcbucket20.txt], srcbucket21.txt=[srcbucket21.txt], srcbucket22.txt=[srcbucket20.txt], srcbucket23.txt=[srcbucket21.txt], ds=2008-04-09/srcbucket20.txt=[srcbucket20.txt], ds=2008-04-09/srcbucket21.txt=[srcbucket21.txt], ds=2008-04-09/srcbucket22.txt=[srcbucket20.txt], ds=2008-04-09/srcbucket23.txt=[srcbucket21.txt]}
Alias Bucket File Name Mapping:
- a {file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08/srcbucket20.txt=[file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket20.txt], file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08/srcbucket21.txt=[file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket21.txt], file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08/srcbucket22.txt=[file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket20.txt], file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08/srcbucket23.txt=[file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket21.txt], file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09/srcbucket20.txt=[file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket20.txt], file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09/srcbucket21.txt=[file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket21.txt], file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09/srcbucket22.txt=[file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket20.txt], file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09/srcbucket23.txt=[file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket21.txt]}
+ a {file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08/srcbucket20.txt=[file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket20.txt], file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08/srcbucket21.txt=[file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket21.txt], file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08/srcbucket22.txt=[file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket20.txt], file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08/srcbucket23.txt=[file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket21.txt], file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09/srcbucket20.txt=[file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket20.txt], file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09/srcbucket21.txt=[file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket21.txt], file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09/srcbucket22.txt=[file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket20.txt], file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09/srcbucket23.txt=[file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket21.txt]}
Alias Bucket Output File Name Mapping:
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08/srcbucket20.txt 0
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08/srcbucket21.txt 1
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08/srcbucket22.txt 2
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08/srcbucket23.txt 3
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09/srcbucket20.txt 0
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09/srcbucket21.txt 1
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09/srcbucket22.txt 2
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09/srcbucket23.txt 3
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08/srcbucket20.txt 0
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08/srcbucket21.txt 1
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08/srcbucket22.txt 2
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08/srcbucket23.txt 3
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09/srcbucket20.txt 0
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09/srcbucket21.txt 1
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09/srcbucket22.txt 2
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09/srcbucket23.txt 3
Needs Tagging: false
Path -> Alias:
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08 [b]
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09 [b]
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08 [b]
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09 [b]
Path -> Partition:
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-08
Partition
base file name: ds=2008-04-08
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -302,13 +278,13 @@
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part
name srcbucket_mapjoin_part
partition_columns ds
serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584877
+ transient_lastDdlTime 1279736124
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -320,17 +296,17 @@
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part
name srcbucket_mapjoin_part
partition_columns ds
serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584877
+ transient_lastDdlTime 1279736124
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcbucket_mapjoin_part
name: srcbucket_mapjoin_part
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part/ds=2008-04-09
Partition
base file name: ds=2008-04-09
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -344,13 +320,13 @@
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part
name srcbucket_mapjoin_part
partition_columns ds
serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584877
+ transient_lastDdlTime 1279736124
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -362,13 +338,13 @@
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part
name srcbucket_mapjoin_part
partition_columns ds
serialization.ddl struct srcbucket_mapjoin_part { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584877
+ transient_lastDdlTime 1279736124
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcbucket_mapjoin_part
name: srcbucket_mapjoin_part
@@ -380,14 +356,14 @@
Move Operator
files:
hdfs directory: true
- source: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-14-40_283_968548719978651164/10002
- destination: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-14-40_283_968548719978651164/10000
+ source: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-27_050_3859053918333083391/10002
+ destination: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-27_050_3859053918333083391/10000
Stage: Stage-0
Move Operator
tables:
replace: true
- source: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-14-40_283_968548719978651164/10000
+ source: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-27_050_3859053918333083391/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -397,20 +373,20 @@
columns.types string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
name bucketmapjoin_tmp_result
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584880
+ transient_lastDdlTime 1279736127
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucketmapjoin_tmp_result
- tmp directory: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-14-40_283_968548719978651164/10001
+ tmp directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-27_050_3859053918333083391/10001
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-14-40_283_968548719978651164/10002
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-27_050_3859053918333083391/10002
Reduce Output Operator
sort order:
Map-reduce partition columns:
@@ -426,9 +402,9 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-14-40_283_968548719978651164/10002 [file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-14-40_283_968548719978651164/10002]
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-27_050_3859053918333083391/10002 [file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-27_050_3859053918333083391/10002]
Path -> Partition:
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-14-40_283_968548719978651164/10002
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-27_050_3859053918333083391/10002
Partition
base file name: 10002
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -439,12 +415,12 @@
columns.types string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
name bucketmapjoin_tmp_result
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584880
+ transient_lastDdlTime 1279736127
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -455,12 +431,12 @@
columns.types string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
name bucketmapjoin_tmp_result
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584880
+ transient_lastDdlTime 1279736127
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucketmapjoin_tmp_result
name: bucketmapjoin_tmp_result
@@ -469,7 +445,7 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-14-40_283_968548719978651164/10000
+ directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-27_050_3859053918333083391/10000
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -480,12 +456,12 @@
columns.types string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
name bucketmapjoin_tmp_result
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584880
+ transient_lastDdlTime 1279736127
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucketmapjoin_tmp_result
TotalFiles: 1
@@ -516,11 +492,11 @@
PREHOOK: query: select count(1) from bucketmapjoin_tmp_result
PREHOOK: type: QUERY
PREHOOK: Input: default@bucketmapjoin_tmp_result
-PREHOOK: Output: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-14-52_156_4855266710517903794/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-15-34_483_4663724029568697547/10000
POSTHOOK: query: select count(1) from bucketmapjoin_tmp_result
POSTHOOK: type: QUERY
POSTHOOK: Input: default@bucketmapjoin_tmp_result
-POSTHOOK: Output: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-14-52_156_4855266710517903794/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-15-34_483_4663724029568697547/10000
POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part)b.FieldSchema(name:key, type:int, comment:null), ]
@@ -571,11 +547,11 @@
PREHOOK: query: select count(1) from bucketmapjoin_tmp_result
PREHOOK: type: QUERY
PREHOOK: Input: default@bucketmapjoin_tmp_result
-PREHOOK: Output: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-09_738_8775012863049281501/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-15-47_270_4545636427863902378/10000
POSTHOOK: query: select count(1) from bucketmapjoin_tmp_result
POSTHOOK: type: QUERY
POSTHOOK: Input: default@bucketmapjoin_tmp_result
-POSTHOOK: Output: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-09_738_8775012863049281501/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-15-47_270_4545636427863902378/10000
POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
@@ -614,14 +590,14 @@
PREHOOK: type: QUERY
PREHOOK: Input: default@bucketmapjoin_hash_result_2
PREHOOK: Input: default@bucketmapjoin_hash_result_1
-PREHOOK: Output: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-17_170_2507838828904708527/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-15-52_543_2406511845804161610/10000
POSTHOOK: query: select a.key-b.key, a.value1-b.value1, a.value2-b.value2
from bucketmapjoin_hash_result_1 a left outer join bucketmapjoin_hash_result_2 b
on a.key = b.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@bucketmapjoin_hash_result_2
POSTHOOK: Input: default@bucketmapjoin_hash_result_1
-POSTHOOK: Output: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-17_170_2507838828904708527/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-15-52_543_2406511845804161610/10000
POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
@@ -709,7 +685,7 @@
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-21_973_5804199576184173549/10002
+ directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-55_337_7214106352789078536/10002
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -720,12 +696,12 @@
columns.types string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
name bucketmapjoin_tmp_result
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584880
+ transient_lastDdlTime 1279736147
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucketmapjoin_tmp_result
TotalFiles: 1
@@ -773,7 +749,7 @@
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-21_973_5804199576184173549/10002
+ directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-55_337_7214106352789078536/10002
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -784,12 +760,12 @@
columns.types string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
name bucketmapjoin_tmp_result
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584880
+ transient_lastDdlTime 1279736147
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucketmapjoin_tmp_result
TotalFiles: 1
@@ -798,18 +774,18 @@
Alias Bucket Base File Name Mapping:
a {srcbucket22.txt=[srcbucket20.txt], srcbucket23.txt=[srcbucket21.txt], ds=2008-04-09/srcbucket22.txt=[srcbucket20.txt], ds=2008-04-09/srcbucket23.txt=[srcbucket21.txt]}
Alias Bucket File Name Mapping:
- a {file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-08/srcbucket22.txt=[file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket20.txt], file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-08/srcbucket23.txt=[file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket21.txt], file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-09/srcbucket22.txt=[file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket20.txt], file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-09/srcbucket23.txt=[file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket21.txt]}
+ a {file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-08/srcbucket22.txt=[file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket20.txt], file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-08/srcbucket23.txt=[file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket21.txt], file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-09/srcbucket22.txt=[file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket20.txt], file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-09/srcbucket23.txt=[file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin/srcbucket21.txt]}
Alias Bucket Output File Name Mapping:
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-08/srcbucket22.txt 0
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-08/srcbucket23.txt 1
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-09/srcbucket22.txt 0
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-09/srcbucket23.txt 1
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-08/srcbucket22.txt 0
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-08/srcbucket23.txt 1
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-09/srcbucket22.txt 0
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-09/srcbucket23.txt 1
Needs Tagging: false
Path -> Alias:
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-08 [b]
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-09 [b]
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-08 [b]
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-09 [b]
Path -> Partition:
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-08
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-08
Partition
base file name: ds=2008-04-08
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -823,13 +799,13 @@
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2
name srcbucket_mapjoin_part_2
partition_columns ds
serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584879
+ transient_lastDdlTime 1279736126
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -841,17 +817,17 @@
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2
name srcbucket_mapjoin_part_2
partition_columns ds
serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584879
+ transient_lastDdlTime 1279736126
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcbucket_mapjoin_part_2
name: srcbucket_mapjoin_part_2
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-09
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2/ds=2008-04-09
Partition
base file name: ds=2008-04-09
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -865,13 +841,13 @@
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2
name srcbucket_mapjoin_part_2
partition_columns ds
serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584879
+ transient_lastDdlTime 1279736126
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -883,13 +859,13 @@
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcbucket_mapjoin_part_2
name srcbucket_mapjoin_part_2
partition_columns ds
serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584879
+ transient_lastDdlTime 1279736126
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcbucket_mapjoin_part_2
name: srcbucket_mapjoin_part_2
@@ -901,14 +877,14 @@
Move Operator
files:
hdfs directory: true
- source: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-21_973_5804199576184173549/10002
- destination: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-21_973_5804199576184173549/10000
+ source: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-55_337_7214106352789078536/10002
+ destination: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-55_337_7214106352789078536/10000
Stage: Stage-0
Move Operator
tables:
replace: true
- source: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-21_973_5804199576184173549/10000
+ source: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-55_337_7214106352789078536/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -918,20 +894,20 @@
columns.types string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
name bucketmapjoin_tmp_result
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584880
+ transient_lastDdlTime 1279736147
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucketmapjoin_tmp_result
- tmp directory: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-21_973_5804199576184173549/10001
+ tmp directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-55_337_7214106352789078536/10001
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-21_973_5804199576184173549/10002
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-55_337_7214106352789078536/10002
Reduce Output Operator
sort order:
Map-reduce partition columns:
@@ -947,9 +923,9 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-21_973_5804199576184173549/10002 [file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-21_973_5804199576184173549/10002]
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-55_337_7214106352789078536/10002 [file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-55_337_7214106352789078536/10002]
Path -> Partition:
- file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-21_973_5804199576184173549/10002
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-55_337_7214106352789078536/10002
Partition
base file name: 10002
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -960,12 +936,12 @@
columns.types string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
name bucketmapjoin_tmp_result
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584880
+ transient_lastDdlTime 1279736147
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -976,12 +952,12 @@
columns.types string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
name bucketmapjoin_tmp_result
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584880
+ transient_lastDdlTime 1279736147
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucketmapjoin_tmp_result
name: bucketmapjoin_tmp_result
@@ -990,7 +966,7 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-21_973_5804199576184173549/10000
+ directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-15-55_337_7214106352789078536/10000
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -1001,12 +977,12 @@
columns.types string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/njain/hive_commit1/hive_commit1/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucketmapjoin_tmp_result
name bucketmapjoin_tmp_result
serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270584880
+ transient_lastDdlTime 1279736147
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucketmapjoin_tmp_result
TotalFiles: 1
@@ -1049,11 +1025,11 @@
PREHOOK: query: select count(1) from bucketmapjoin_tmp_result
PREHOOK: type: QUERY
PREHOOK: Input: default@bucketmapjoin_tmp_result
-PREHOOK: Output: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-31_179_7313595341297279681/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-16-00_681_6039313017560460235/10000
POSTHOOK: query: select count(1) from bucketmapjoin_tmp_result
POSTHOOK: type: QUERY
POSTHOOK: Input: default@bucketmapjoin_tmp_result
-POSTHOOK: Output: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-31_179_7313595341297279681/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-16-00_681_6039313017560460235/10000
POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
@@ -1140,11 +1116,11 @@
PREHOOK: query: select count(1) from bucketmapjoin_tmp_result
PREHOOK: type: QUERY
PREHOOK: Input: default@bucketmapjoin_tmp_result
-PREHOOK: Output: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-46_464_3385157090510772728/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-16-12_250_4593300394239034330/10000
POSTHOOK: query: select count(1) from bucketmapjoin_tmp_result
POSTHOOK: type: QUERY
POSTHOOK: Input: default@bucketmapjoin_tmp_result
-POSTHOOK: Output: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-46_464_3385157090510772728/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-16-12_250_4593300394239034330/10000
POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
@@ -1207,14 +1183,14 @@
PREHOOK: type: QUERY
PREHOOK: Input: default@bucketmapjoin_hash_result_2
PREHOOK: Input: default@bucketmapjoin_hash_result_1
-PREHOOK: Output: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-54_066_2106291188917250080/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-16-17_491_993666857276206775/10000
POSTHOOK: query: select a.key-b.key, a.value1-b.value1, a.value2-b.value2
from bucketmapjoin_hash_result_1 a left outer join bucketmapjoin_hash_result_2 b
on a.key = b.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@bucketmapjoin_hash_result_2
POSTHOOK: Input: default@bucketmapjoin_hash_result_1
-POSTHOOK: Output: file:/data/users/njain/hive_commit1/hive_commit1/build/ql/scratchdir/hive_2010-04-06_13-15-54_066_2106291188917250080/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-16-17_491_993666857276206775/10000
POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
@@ -1240,177 +1216,3 @@
POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part_2)b.FieldSchema(name:key, type:int, comment:null), ]
POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part_2)b.FieldSchema(name:key, type:int, comment:null), ]
NULL NULL NULL
-PREHOOK: query: drop table bucketmapjoin_hash_result_2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table bucketmapjoin_hash_result_2
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@bucketmapjoin_hash_result_2
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part)b.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part)b.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part_2)b.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part_2)b.FieldSchema(name:key, type:int, comment:null), ]
-PREHOOK: query: drop table bucketmapjoin_hash_result_1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table bucketmapjoin_hash_result_1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@bucketmapjoin_hash_result_1
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part)b.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part)b.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part_2)b.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part_2)b.FieldSchema(name:key, type:int, comment:null), ]
-PREHOOK: query: drop table bucketmapjoin_tmp_result
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table bucketmapjoin_tmp_result
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@bucketmapjoin_tmp_result
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part)b.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part)b.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part_2)b.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part_2)b.FieldSchema(name:key, type:int, comment:null), ]
-PREHOOK: query: drop table srcbucket_mapjoin
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table srcbucket_mapjoin
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@srcbucket_mapjoin
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part)b.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part)b.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part_2)b.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part_2)b.FieldSchema(name:key, type:int, comment:null), ]
-PREHOOK: query: drop table srcbucket_mapjoin_part
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table srcbucket_mapjoin_part
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@srcbucket_mapjoin_part
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part)b.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part)b.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part_2)b.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part_2)b.FieldSchema(name:key, type:int, comment:null), ]
-PREHOOK: query: drop table srcbucket_mapjoin_part_2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table srcbucket_mapjoin_part_2
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@srcbucket_mapjoin_part_2
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_1.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.key EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:key, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value1 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value1, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_hash_result_2.value2 EXPRESSION [(bucketmapjoin_tmp_result)bucketmapjoin_tmp_result.FieldSchema(name:value2, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.key SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value1 SIMPLE [(srcbucket_mapjoin)a.FieldSchema(name:value, type:string, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part)b.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part)b.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part_2)b.FieldSchema(name:key, type:int, comment:null), ]
-POSTHOOK: Lineage: bucketmapjoin_tmp_result.value2 SIMPLE [(srcbucket_mapjoin_part_2)b.FieldSchema(name:key, type:int, comment:null), ]
Index: ql/src/test/results/clientpositive/nullinput.q.out
===================================================================
--- ql/src/test/results/clientpositive/nullinput.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/nullinput.q.out (working copy)
@@ -6,21 +6,16 @@
PREHOOK: query: select x.* from tstnullinut x
PREHOOK: type: QUERY
PREHOOK: Input: default@tstnullinut
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1976338873/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-44-42_572_4519312906185121533/10000
POSTHOOK: query: select x.* from tstnullinut x
POSTHOOK: type: QUERY
POSTHOOK: Input: default@tstnullinut
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1976338873/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-44-42_572_4519312906185121533/10000
PREHOOK: query: select x.a, count(1) from tstnullinut x group by x.a
PREHOOK: type: QUERY
PREHOOK: Input: default@tstnullinut
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1008263015/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-44-42_620_1570615621828082286/10000
POSTHOOK: query: select x.a, count(1) from tstnullinut x group by x.a
POSTHOOK: type: QUERY
POSTHOOK: Input: default@tstnullinut
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/1008263015/10000
-PREHOOK: query: drop table tstnullinut
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table tstnullinut
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@tstnullinut
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-44-42_620_1570615621828082286/10000
Index: ql/src/test/results/clientpositive/load_dyn_part2.q.out
===================================================================
--- ql/src/test/results/clientpositive/load_dyn_part2.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/load_dyn_part2.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table nzhang_part_bucket
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table nzhang_part_bucket
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table if not exists nzhang_part_bucket (key string, value string)
partitioned by (ds string, hr string)
clustered by (key) into 10 buckets
@@ -20,7 +16,7 @@
ds string
hr string
-Detailed Table Information Table(tableName:nzhang_part_bucket, dbName:default, owner:nzhang, createTime:1271267614, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/data/users/nzhang/work/876/apache-hive/build/ql/test/data/warehouse/nzhang_part_bucket, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:10, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[key], sortCols:[], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null), FieldSchema(name:hr, type:string, comment:null)], parameters:{transient_lastDdlTime=1271267614}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)
+Detailed Table Information Table(tableName:nzhang_part_bucket, dbName:default, owner:jssarma, createTime:1279737530, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/nzhang_part_bucket, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:10, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[key], sortCols:[], parameters:{}), partitionKeys:[FieldSchema(name:ds, type:string, comment:null), FieldSchema(name:hr, type:string, comment:null)], parameters:{transient_lastDdlTime=1279737530}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)
PREHOOK: query: explain
insert overwrite table nzhang_part_bucket partition (ds='2010-03-23', hr) select key, value, hr from srcpart where ds is not null and hr is not null
PREHOOK: type: QUERY
@@ -127,11 +123,11 @@
PREHOOK: query: select * from nzhang_part_bucket where ds='2010-03-23' and hr='11' order by key
PREHOOK: type: QUERY
PREHOOK: Input: default@nzhang_part_bucket@ds=2010-03-23/hr=11
-PREHOOK: Output: file:/data/users/nzhang/work/876/apache-hive/build/ql/scratchdir/hive_2010-04-14_10-53-42_252_5131661556862540128/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-38-55_332_7977928498639952021/10000
POSTHOOK: query: select * from nzhang_part_bucket where ds='2010-03-23' and hr='11' order by key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@nzhang_part_bucket@ds=2010-03-23/hr=11
-POSTHOOK: Output: file:/data/users/nzhang/work/876/apache-hive/build/ql/scratchdir/hive_2010-04-14_10-53-42_252_5131661556862540128/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-38-55_332_7977928498639952021/10000
POSTHOOK: Lineage: nzhang_part_bucket PARTITION(ds=2010-03-23,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ]
POSTHOOK: Lineage: nzhang_part_bucket PARTITION(ds=2010-03-23,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:hr, type:string, comment:null), ]
POSTHOOK: Lineage: nzhang_part_bucket PARTITION(ds=2010-03-23,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ]
@@ -1139,11 +1135,11 @@
PREHOOK: query: select * from nzhang_part_bucket where ds='2010-03-23' and hr='12' order by key
PREHOOK: type: QUERY
PREHOOK: Input: default@nzhang_part_bucket@ds=2010-03-23/hr=12
-PREHOOK: Output: file:/data/users/nzhang/work/876/apache-hive/build/ql/scratchdir/hive_2010-04-14_10-53-49_136_4048550724530357597/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-39-00_003_5219381574950480620/10000
POSTHOOK: query: select * from nzhang_part_bucket where ds='2010-03-23' and hr='12' order by key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@nzhang_part_bucket@ds=2010-03-23/hr=12
-POSTHOOK: Output: file:/data/users/nzhang/work/876/apache-hive/build/ql/scratchdir/hive_2010-04-14_10-53-49_136_4048550724530357597/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-39-00_003_5219381574950480620/10000
POSTHOOK: Lineage: nzhang_part_bucket PARTITION(ds=2010-03-23,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ]
POSTHOOK: Lineage: nzhang_part_bucket PARTITION(ds=2010-03-23,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:hr, type:string, comment:null), ]
POSTHOOK: Lineage: nzhang_part_bucket PARTITION(ds=2010-03-23,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ]
@@ -2148,12 +2144,3 @@
98 val_98 2010-03-23 12
98 val_98 2010-03-23 12
98 val_98 2010-03-23 12
-PREHOOK: query: drop table nzhang_part_bucket
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table nzhang_part_bucket
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@nzhang_part_bucket
-POSTHOOK: Lineage: nzhang_part_bucket PARTITION(ds=2010-03-23,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ]
-POSTHOOK: Lineage: nzhang_part_bucket PARTITION(ds=2010-03-23,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:hr, type:string, comment:null), ]
-POSTHOOK: Lineage: nzhang_part_bucket PARTITION(ds=2010-03-23,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ]
-POSTHOOK: Lineage: nzhang_part_bucket PARTITION(ds=2010-03-23,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:hr, type:string, comment:null), ]
Index: ql/src/test/results/clientpositive/join_reorder3.q.out
===================================================================
--- ql/src/test/results/clientpositive/join_reorder3.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/join_reorder3.q.out (working copy)
@@ -1,19 +1,3 @@
-PREHOOK: query: DROP TABLE T1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE T1
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: DROP TABLE T2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE T2
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: DROP TABLE T3
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE T3
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: DROP TABLE T4
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE T4
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE T1(key STRING, val STRING) STORED AS TEXTFILE
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE T1(key STRING, val STRING) STORED AS TEXTFILE
@@ -198,7 +182,7 @@
PREHOOK: Input: default@t2
PREHOOK: Input: default@t3
PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-34_907_1935969450376975821/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-37-00_490_2765894359578078501/10000
POSTHOOK: query: SELECT /*+ STREAMTABLE(a,c) */ *
FROM T1 a JOIN T2 b ON a.key = b.key
JOIN T3 c ON b.key = c.key
@@ -208,7 +192,7 @@
POSTHOOK: Input: default@t2
POSTHOOK: Input: default@t3
POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-34_907_1935969450376975821/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-37-00_490_2765894359578078501/10000
2 12 2 22 2 12 2 12
PREHOOK: query: EXPLAIN
SELECT /*+ STREAMTABLE(a,c) */ *
@@ -431,7 +415,7 @@
PREHOOK: Input: default@t2
PREHOOK: Input: default@t3
PREHOOK: Input: default@t1
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-41_586_5996486802122601404/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-37-03_448_3407530204384318986/10000
POSTHOOK: query: SELECT /*+ STREAMTABLE(a,c) */ *
FROM T1 a JOIN T2 b ON a.key = b.key
JOIN T3 c ON a.val = c.val
@@ -441,25 +425,5 @@
POSTHOOK: Input: default@t2
POSTHOOK: Input: default@t3
POSTHOOK: Input: default@t1
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_0/build/ql/scratchdir/hive_2010-02-12_22-33-41_586_5996486802122601404/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-37-03_448_3407530204384318986/10000
2 22 2 12 2 12 2 12
-PREHOOK: query: DROP TABLE T1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE T1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@t1
-PREHOOK: query: DROP TABLE T2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE T2
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@t2
-PREHOOK: query: DROP TABLE T3
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE T3
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@t3
-PREHOOK: query: DROP TABLE T4
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE T4
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@t4
Index: ql/src/test/results/clientpositive/join24.q.out
===================================================================
--- ql/src/test/results/clientpositive/join24.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/join24.q.out (working copy)
@@ -18,18 +18,11 @@
PREHOOK: query: SELECT sum(a.cnt) FROM tst1 a JOIN tst1 b ON a.key = b.key
PREHOOK: type: QUERY
PREHOOK: Input: default@tst1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-18-43_917_116815481322601521/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-32-31_747_8781349854837521584/10000
POSTHOOK: query: SELECT sum(a.cnt) FROM tst1 a JOIN tst1 b ON a.key = b.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@tst1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-18-43_917_116815481322601521/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-32-31_747_8781349854837521584/10000
POSTHOOK: Lineage: tst1.cnt EXPRESSION [(src)a.null, ]
POSTHOOK: Lineage: tst1.key SIMPLE [(src)a.FieldSchema(name:key, type:string, comment:default), ]
500
-PREHOOK: query: drop table tst1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table tst1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@tst1
-POSTHOOK: Lineage: tst1.cnt EXPRESSION [(src)a.null, ]
-POSTHOOK: Lineage: tst1.key SIMPLE [(src)a.FieldSchema(name:key, type:string, comment:default), ]
Index: ql/src/test/results/clientpositive/input1.q.out
===================================================================
--- ql/src/test/results/clientpositive/input1.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/input1.q.out (working copy)
@@ -33,8 +33,3 @@
POSTHOOK: type: DESCTABLE
a int
b double
-PREHOOK: query: DROP TABLE TEST1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE TEST1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@test1
Index: ql/src/test/results/clientpositive/smb_mapjoin_5.q.out
===================================================================
--- ql/src/test/results/clientpositive/smb_mapjoin_5.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/smb_mapjoin_5.q.out (working copy)
@@ -1,15 +1,3 @@
-PREHOOK: query: drop table smb_bucket_3
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table smb_bucket_3
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: drop table smb_bucket_2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table smb_bucket_2
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: drop table smb_bucket_1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table smb_bucket_1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table smb_bucket_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table smb_bucket_1(key int, value string) CLUSTERED BY (key) SORTED BY (key) INTO 1 BUCKETS STORED AS RCFILE
@@ -122,13 +110,13 @@
PREHOOK: Input: default@smb_bucket_2
PREHOOK: Input: default@smb_bucket_3
PREHOOK: Input: default@smb_bucket_1
-PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-36-22_865_4487686662397360125/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-00-41_524_6364697581539120113/10000
POSTHOOK: query: select /*+mapjoin(a,c)*/ * from smb_bucket_1 a join smb_bucket_2 b on a.key = b.key join smb_bucket_3 c on b.key=c.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@smb_bucket_2
POSTHOOK: Input: default@smb_bucket_3
POSTHOOK: Input: default@smb_bucket_1
-POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-36-22_865_4487686662397360125/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-00-41_524_6364697581539120113/10000
PREHOOK: query: explain
select /*+mapjoin(a,c)*/ * from smb_bucket_1 a left outer join smb_bucket_2 b on a.key = b.key join smb_bucket_3 c on b.key=c.key
PREHOOK: type: QUERY
@@ -211,13 +199,13 @@
PREHOOK: Input: default@smb_bucket_2
PREHOOK: Input: default@smb_bucket_3
PREHOOK: Input: default@smb_bucket_1
-PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-36-27_181_2689448944799681233/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-00-44_285_5430902360454396405/10000
POSTHOOK: query: select /*+mapjoin(a,c)*/ * from smb_bucket_1 a left outer join smb_bucket_2 b on a.key = b.key join smb_bucket_3 c on b.key=c.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@smb_bucket_2
POSTHOOK: Input: default@smb_bucket_3
POSTHOOK: Input: default@smb_bucket_1
-POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-36-27_181_2689448944799681233/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-00-44_285_5430902360454396405/10000
PREHOOK: query: explain
select /*+mapjoin(a,c)*/ * from smb_bucket_1 a left outer join smb_bucket_2 b on a.key = b.key left outer join smb_bucket_3 c on b.key=c.key
PREHOOK: type: QUERY
@@ -300,13 +288,13 @@
PREHOOK: Input: default@smb_bucket_2
PREHOOK: Input: default@smb_bucket_3
PREHOOK: Input: default@smb_bucket_1
-PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-36-30_958_2862069044325639274/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-00-47_123_2017928206991935005/10000
POSTHOOK: query: select /*+mapjoin(a,c)*/ * from smb_bucket_1 a left outer join smb_bucket_2 b on a.key = b.key left outer join smb_bucket_3 c on b.key=c.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@smb_bucket_2
POSTHOOK: Input: default@smb_bucket_3
POSTHOOK: Input: default@smb_bucket_1
-POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-36-30_958_2862069044325639274/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-00-47_123_2017928206991935005/10000
1 val_1 NULL NULL NULL NULL
3 val_3 NULL NULL NULL NULL
4 val_4 NULL NULL NULL NULL
@@ -394,13 +382,13 @@
PREHOOK: Input: default@smb_bucket_2
PREHOOK: Input: default@smb_bucket_3
PREHOOK: Input: default@smb_bucket_1
-PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-36-35_461_5426046396711146559/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-00-49_934_7244598924132392571/10000
POSTHOOK: query: select /*+mapjoin(a,c)*/ * from smb_bucket_1 a left outer join smb_bucket_2 b on a.key = b.key right outer join smb_bucket_3 c on b.key=c.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@smb_bucket_2
POSTHOOK: Input: default@smb_bucket_3
POSTHOOK: Input: default@smb_bucket_1
-POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-36-35_461_5426046396711146559/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-00-49_934_7244598924132392571/10000
NULL NULL NULL NULL 4 val_4
NULL NULL NULL NULL 10 val_10
NULL NULL NULL NULL 17 val_17
@@ -489,13 +477,13 @@
PREHOOK: Input: default@smb_bucket_2
PREHOOK: Input: default@smb_bucket_3
PREHOOK: Input: default@smb_bucket_1
-PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-36-39_715_1766943698430352843/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-00-52_668_3083976054304451937/10000
POSTHOOK: query: select /*+mapjoin(a,c)*/ * from smb_bucket_1 a left outer join smb_bucket_2 b on a.key = b.key full outer join smb_bucket_3 c on b.key=c.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@smb_bucket_2
POSTHOOK: Input: default@smb_bucket_3
POSTHOOK: Input: default@smb_bucket_1
-POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-36-39_715_1766943698430352843/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-00-52_668_3083976054304451937/10000
1 val_1 NULL NULL NULL NULL
3 val_3 NULL NULL NULL NULL
4 val_4 NULL NULL NULL NULL
@@ -589,13 +577,13 @@
PREHOOK: Input: default@smb_bucket_2
PREHOOK: Input: default@smb_bucket_3
PREHOOK: Input: default@smb_bucket_1
-PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-36-43_973_8806162188078780005/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-00-55_443_5909131157897439506/10000
POSTHOOK: query: select /*+mapjoin(a,c)*/ * from smb_bucket_1 a right outer join smb_bucket_2 b on a.key = b.key join smb_bucket_3 c on b.key=c.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@smb_bucket_2
POSTHOOK: Input: default@smb_bucket_3
POSTHOOK: Input: default@smb_bucket_1
-POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-36-43_973_8806162188078780005/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-00-55_443_5909131157897439506/10000
NULL NULL 20 val_20 20 val_20
NULL NULL 23 val_23 23 val_23
PREHOOK: query: explain
@@ -680,13 +668,13 @@
PREHOOK: Input: default@smb_bucket_2
PREHOOK: Input: default@smb_bucket_3
PREHOOK: Input: default@smb_bucket_1
-PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-36-48_619_2725384740684408102/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-00-58_148_4762824573436827785/10000
POSTHOOK: query: select /*+mapjoin(a,c)*/ * from smb_bucket_1 a right outer join smb_bucket_2 b on a.key = b.key left outer join smb_bucket_3 c on b.key=c.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@smb_bucket_2
POSTHOOK: Input: default@smb_bucket_3
POSTHOOK: Input: default@smb_bucket_1
-POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-36-48_619_2725384740684408102/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-00-58_148_4762824573436827785/10000
NULL NULL 20 val_20 20 val_20
NULL NULL 23 val_23 23 val_23
NULL NULL 25 val_25 NULL NULL
@@ -773,13 +761,13 @@
PREHOOK: Input: default@smb_bucket_2
PREHOOK: Input: default@smb_bucket_3
PREHOOK: Input: default@smb_bucket_1
-PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-36-53_557_849175655418581108/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-01-00_906_6750483002470270430/10000
POSTHOOK: query: select /*+mapjoin(a,c)*/ * from smb_bucket_1 a right outer join smb_bucket_2 b on a.key = b.key right outer join smb_bucket_3 c on b.key=c.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@smb_bucket_2
POSTHOOK: Input: default@smb_bucket_3
POSTHOOK: Input: default@smb_bucket_1
-POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-36-53_557_849175655418581108/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-01-00_906_6750483002470270430/10000
NULL NULL NULL NULL 4 val_4
NULL NULL NULL NULL 10 val_10
NULL NULL NULL NULL 17 val_17
@@ -868,13 +856,13 @@
PREHOOK: Input: default@smb_bucket_2
PREHOOK: Input: default@smb_bucket_3
PREHOOK: Input: default@smb_bucket_1
-PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-36-58_093_7448714045296241567/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-01-03_722_2114361761413441528/10000
POSTHOOK: query: select /*+mapjoin(a,c)*/ * from smb_bucket_1 a right outer join smb_bucket_2 b on a.key = b.key full outer join smb_bucket_3 c on b.key=c.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@smb_bucket_2
POSTHOOK: Input: default@smb_bucket_3
POSTHOOK: Input: default@smb_bucket_1
-POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-36-58_093_7448714045296241567/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-01-03_722_2114361761413441528/10000
NULL NULL NULL NULL 4 val_4
NULL NULL NULL NULL 10 val_10
NULL NULL NULL NULL 17 val_17
@@ -965,13 +953,13 @@
PREHOOK: Input: default@smb_bucket_2
PREHOOK: Input: default@smb_bucket_3
PREHOOK: Input: default@smb_bucket_1
-PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-37-02_568_4148492212234051327/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-01-06_447_2677961588327359326/10000
POSTHOOK: query: select /*+mapjoin(a,c)*/ * from smb_bucket_1 a full outer join smb_bucket_2 b on a.key = b.key join smb_bucket_3 c on b.key=c.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@smb_bucket_2
POSTHOOK: Input: default@smb_bucket_3
POSTHOOK: Input: default@smb_bucket_1
-POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-37-02_568_4148492212234051327/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-01-06_447_2677961588327359326/10000
NULL NULL 20 val_20 20 val_20
NULL NULL 23 val_23 23 val_23
PREHOOK: query: explain
@@ -1056,13 +1044,13 @@
PREHOOK: Input: default@smb_bucket_2
PREHOOK: Input: default@smb_bucket_3
PREHOOK: Input: default@smb_bucket_1
-PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-37-06_625_1217755236567536175/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-01-09_243_6046930209513279147/10000
POSTHOOK: query: select /*+mapjoin(a,c)*/ * from smb_bucket_1 a full outer join smb_bucket_2 b on a.key = b.key left outer join smb_bucket_3 c on b.key=c.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@smb_bucket_2
POSTHOOK: Input: default@smb_bucket_3
POSTHOOK: Input: default@smb_bucket_1
-POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-37-06_625_1217755236567536175/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-01-09_243_6046930209513279147/10000
1 val_1 NULL NULL NULL NULL
3 val_3 NULL NULL NULL NULL
4 val_4 NULL NULL NULL NULL
@@ -1154,13 +1142,13 @@
PREHOOK: Input: default@smb_bucket_2
PREHOOK: Input: default@smb_bucket_3
PREHOOK: Input: default@smb_bucket_1
-PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-37-11_033_6560347088229684794/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-01-12_084_8336441249717148411/10000
POSTHOOK: query: select /*+mapjoin(a,c)*/ * from smb_bucket_1 a full outer join smb_bucket_2 b on a.key = b.key right outer join smb_bucket_3 c on b.key=c.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@smb_bucket_2
POSTHOOK: Input: default@smb_bucket_3
POSTHOOK: Input: default@smb_bucket_1
-POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-37-11_033_6560347088229684794/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-01-12_084_8336441249717148411/10000
NULL NULL NULL NULL 4 val_4
NULL NULL NULL NULL 10 val_10
NULL NULL NULL NULL 17 val_17
@@ -1249,13 +1237,13 @@
PREHOOK: Input: default@smb_bucket_2
PREHOOK: Input: default@smb_bucket_3
PREHOOK: Input: default@smb_bucket_1
-PREHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-37-15_593_2231203586411799562/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-01-14_792_2190079985184779302/10000
POSTHOOK: query: select /*+mapjoin(a,c)*/ * from smb_bucket_1 a full outer join smb_bucket_2 b on a.key = b.key full outer join smb_bucket_3 c on b.key=c.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@smb_bucket_2
POSTHOOK: Input: default@smb_bucket_3
POSTHOOK: Input: default@smb_bucket_1
-POSTHOOK: Output: file:/var/folders/6g/6grtCwPMEf4sqHUPpy6xQG9ByHg/-Tmp-/heyongqiang/hive_2010-06-25_22-37-15_593_2231203586411799562/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_12-01-14_792_2190079985184779302/10000
1 val_1 NULL NULL NULL NULL
3 val_3 NULL NULL NULL NULL
4 val_4 NULL NULL NULL NULL
@@ -1269,18 +1257,3 @@
NULL NULL 23 val_23 23 val_23
NULL NULL 25 val_25 NULL NULL
NULL NULL 30 val_30 NULL NULL
-PREHOOK: query: drop table smb_bucket_3
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table smb_bucket_3
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@smb_bucket_3
-PREHOOK: query: drop table smb_bucket_2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table smb_bucket_2
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@smb_bucket_2
-PREHOOK: query: drop table smb_bucket_1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table smb_bucket_1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@smb_bucket_1
Index: ql/src/test/results/clientpositive/join33.q.out
===================================================================
--- ql/src/test/results/clientpositive/join33.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/join33.q.out (working copy)
@@ -45,7 +45,7 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/tmp/jssarma/hive_2010-06-10_16-21-45_664_3974168394039456921/10002
+ directory: file:/tmp/jssarma/hive_2010-07-21_11-33-43_308_1055825397797894449/10002
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -81,7 +81,7 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/tmp/jssarma/hive_2010-06-10_16-21-45_664_3974168394039456921/10002
+ directory: file:/tmp/jssarma/hive_2010-07-21_11-33-43_308_1055825397797894449/10002
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -112,7 +112,7 @@
serialization.ddl struct src { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1276212105
+ transient_lastDdlTime 1279735685
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -128,7 +128,7 @@
serialization.ddl struct src { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1276212105
+ transient_lastDdlTime 1279735685
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: src
name: src
@@ -136,7 +136,7 @@
Stage: Stage-1
Map Reduce
Alias -> Map Operator Tree:
- file:/tmp/jssarma/hive_2010-06-10_16-21-45_664_3974168394039456921/10002
+ file:/tmp/jssarma/hive_2010-07-21_11-33-43_308_1055825397797894449/10002
Select Operator
expressions:
expr: _col0
@@ -193,7 +193,7 @@
Needs Tagging: true
Path -> Alias:
file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11 [z]
- file:/tmp/jssarma/hive_2010-06-10_16-21-45_664_3974168394039456921/10002 [file:/tmp/jssarma/hive_2010-06-10_16-21-45_664_3974168394039456921/10002]
+ file:/tmp/jssarma/hive_2010-07-21_11-33-43_308_1055825397797894449/10002 [file:/tmp/jssarma/hive_2010-07-21_11-33-43_308_1055825397797894449/10002]
Path -> Partition:
file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=11
Partition
@@ -215,7 +215,7 @@
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1276212103
+ transient_lastDdlTime 1279735681
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -232,11 +232,11 @@
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1276212103
+ transient_lastDdlTime 1279735681
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
name: srcpart
- file:/tmp/jssarma/hive_2010-06-10_16-21-45_664_3974168394039456921/10002
+ file:/tmp/jssarma/hive_2010-07-21_11-33-43_308_1055825397797894449/10002
Partition
base file name: 10002
input format: org.apache.hadoop.mapred.SequenceFileInputFormat
@@ -273,7 +273,7 @@
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-06-10_16-21-45_664_3974168394039456921/10000
+ directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-33-43_308_1055825397797894449/10000
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -289,7 +289,7 @@
serialization.ddl struct dest_j1 { string key, string value, string val2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1276212105
+ transient_lastDdlTime 1279737223
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest_j1
TotalFiles: 1
@@ -299,7 +299,7 @@
Move Operator
tables:
replace: true
- source: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-06-10_16-21-45_664_3974168394039456921/10000
+ source: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-33-43_308_1055825397797894449/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -314,10 +314,10 @@
serialization.ddl struct dest_j1 { string key, string value, string val2}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1276212105
+ transient_lastDdlTime 1279737223
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest_j1
- tmp directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-06-10_16-21-45_664_3974168394039456921/10001
+ tmp directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-33-43_308_1055825397797894449/10001
PREHOOK: query: INSERT OVERWRITE TABLE dest_j1
@@ -344,11 +344,11 @@
PREHOOK: query: select * from dest_j1 x order by x.key
PREHOOK: type: QUERY
PREHOOK: Input: default@dest_j1
-PREHOOK: Output: file:/tmp/jssarma/hive_2010-06-10_16-21-51_616_8853310441674539967/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-33-48_544_7786473350800117933/10000
POSTHOOK: query: select * from dest_j1 x order by x.key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest_j1
-POSTHOOK: Output: file:/tmp/jssarma/hive_2010-06-10_16-21-51_616_8853310441674539967/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-33-48_544_7786473350800117933/10000
POSTHOOK: Lineage: dest_j1.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: dest_j1.val2 EXPRESSION [(src)y.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest_j1.value SIMPLE [(srcpart)z.FieldSchema(name:hr, type:string, comment:null), ]
@@ -437,11 +437,3 @@
98 val_98 val_98
98 val_98 val_98
98 val_98 val_98
-PREHOOK: query: drop table dest_j1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table dest_j1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@dest_j1
-POSTHOOK: Lineage: dest_j1.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: dest_j1.val2 EXPRESSION [(src)y.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest_j1.value SIMPLE [(srcpart)z.FieldSchema(name:hr, type:string, comment:null), ]
Index: ql/src/test/results/clientpositive/input15.q.out
===================================================================
--- ql/src/test/results/clientpositive/input15.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/input15.q.out (working copy)
@@ -35,8 +35,3 @@
POSTHOOK: type: DESCTABLE
key int
value string
-PREHOOK: query: DROP TABLE TEST15
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE TEST15
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@test15
Index: ql/src/test/results/clientpositive/input_part2.q.out
===================================================================
--- ql/src/test/results/clientpositive/input_part2.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/input_part2.q.out (working copy)
@@ -69,7 +69,7 @@
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10004
+ directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10004
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -80,12 +80,12 @@
columns.types int:string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/dest1
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/dest1
name dest1
serialization.ddl struct dest1 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270516376
+ transient_lastDdlTime 1279737005
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest1
TotalFiles: 1
@@ -120,7 +120,7 @@
File Output Operator
compressed: false
GlobalTableId: 2
- directory: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10005
+ directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10005
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -131,22 +131,22 @@
columns.types int:string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/dest2
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/dest2
name dest2
serialization.ddl struct dest2 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270516376
+ transient_lastDdlTime 1279737005
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest2
TotalFiles: 1
MultiFileSpray: false
Needs Tagging: false
Path -> Alias:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [srcpart]
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12 [srcpart]
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12 [srcpart]
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12 [srcpart]
Path -> Partition:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-08/hr=12
Partition
base file name: hr=12
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -160,13 +160,13 @@
columns.types string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/srcpart
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcpart
name srcpart
partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270516374
+ transient_lastDdlTime 1279735681
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -177,17 +177,17 @@
columns.types string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/srcpart
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcpart
name srcpart
partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270516374
+ transient_lastDdlTime 1279735681
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
name: srcpart
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcpart/ds=2008-04-09/hr=12
Partition
base file name: hr=12
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -201,13 +201,13 @@
columns.types string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/srcpart
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcpart
name srcpart
partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270516374
+ transient_lastDdlTime 1279735681
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -218,13 +218,13 @@
columns.types string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/srcpart
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/srcpart
name srcpart
partition_columns ds/hr
serialization.ddl struct srcpart { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270516374
+ transient_lastDdlTime 1279735681
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: srcpart
name: srcpart
@@ -236,14 +236,14 @@
Move Operator
files:
hdfs directory: true
- source: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10004
- destination: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10000
+ source: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10004
+ destination: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10000
Stage: Stage-0
Move Operator
tables:
replace: true
- source: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10000
+ source: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -253,20 +253,20 @@
columns.types int:string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/dest1
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/dest1
name dest1
serialization.ddl struct dest1 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270516376
+ transient_lastDdlTime 1279737005
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest1
- tmp directory: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10001
+ tmp directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10001
Stage: Stage-3
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10004
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10004
Reduce Output Operator
sort order:
Map-reduce partition columns:
@@ -284,9 +284,9 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10004 [file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10004]
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10004 [file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10004]
Path -> Partition:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10004
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10004
Partition
base file name: 10004
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -297,12 +297,12 @@
columns.types int:string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/dest1
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/dest1
name dest1
serialization.ddl struct dest1 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270516376
+ transient_lastDdlTime 1279737005
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -313,12 +313,12 @@
columns.types int:string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/dest1
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/dest1
name dest1
serialization.ddl struct dest1 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270516376
+ transient_lastDdlTime 1279737005
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest1
name: dest1
@@ -327,7 +327,7 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10000
+ directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10000
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -338,12 +338,12 @@
columns.types int:string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/dest1
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/dest1
name dest1
serialization.ddl struct dest1 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270516376
+ transient_lastDdlTime 1279737005
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest1
TotalFiles: 1
@@ -356,14 +356,14 @@
Move Operator
files:
hdfs directory: true
- source: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10005
- destination: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10002
+ source: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10005
+ destination: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10002
Stage: Stage-1
Move Operator
tables:
replace: true
- source: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10002
+ source: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10002
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -373,20 +373,20 @@
columns.types int:string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/dest2
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/dest2
name dest2
serialization.ddl struct dest2 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270516376
+ transient_lastDdlTime 1279737005
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest2
- tmp directory: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10003
+ tmp directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10003
Stage: Stage-6
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10005
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10005
Reduce Output Operator
sort order:
Map-reduce partition columns:
@@ -404,9 +404,9 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10005 [file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10005]
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10005 [file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10005]
Path -> Partition:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10005
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10005
Partition
base file name: 10005
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -417,12 +417,12 @@
columns.types int:string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/dest2
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/dest2
name dest2
serialization.ddl struct dest2 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270516376
+ transient_lastDdlTime 1279737005
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -433,12 +433,12 @@
columns.types int:string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/dest2
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/dest2
name dest2
serialization.ddl struct dest2 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270516376
+ transient_lastDdlTime 1279737005
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest2
name: dest2
@@ -447,7 +447,7 @@
File Output Operator
compressed: false
GlobalTableId: 0
- directory: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-12-56_344_4511101399410096223/10002
+ directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-30-05_098_8127089578831075147/10002
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -458,12 +458,12 @@
columns.types int:string:string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/dest2
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/dest2
name dest2
serialization.ddl struct dest2 { i32 key, string value, string hr, string ds}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270516376
+ transient_lastDdlTime 1279737005
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: dest2
TotalFiles: 1
@@ -497,11 +497,11 @@
PREHOOK: query: SELECT dest1.* FROM dest1 sort by key,value,ds,hr
PREHOOK: type: QUERY
PREHOOK: Input: default@dest1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-13-14_720_9039761864198236527/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-30-12_828_6857029233330905446/10000
POSTHOOK: query: SELECT dest1.* FROM dest1 sort by key,value,ds,hr
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-13-14_720_9039761864198236527/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-30-12_828_6857029233330905446/10000
POSTHOOK: Lineage: dest1.ds SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.hr SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.key EXPRESSION [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ]
@@ -597,11 +597,11 @@
PREHOOK: query: SELECT dest2.* FROM dest2 sort by key,value,ds,hr
PREHOOK: type: QUERY
PREHOOK: Input: default@dest2
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-13-21_979_5910588511102117456/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-30-15_409_5600979597397235313/10000
POSTHOOK: query: SELECT dest2.* FROM dest2 sort by key,value,ds,hr
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest2
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-13-21_979_5910588511102117456/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-30-15_409_5600979597397235313/10000
POSTHOOK: Lineage: dest1.ds SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.hr SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.key EXPRESSION [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ]
@@ -694,16 +694,3 @@
97 val_97 12 2008-04-09
98 val_98 12 2008-04-09
98 val_98 12 2008-04-09
-PREHOOK: query: drop table dest2
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table dest2
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@dest2
-POSTHOOK: Lineage: dest1.ds SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.hr SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.key EXPRESSION [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ]
-POSTHOOK: Lineage: dest1.value SIMPLE [(srcpart)srcpart.FieldSchema(name:hr, type:string, comment:null), ]
-POSTHOOK: Lineage: dest2.ds SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: dest2.hr SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest2.key EXPRESSION [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ]
-POSTHOOK: Lineage: dest2.value SIMPLE [(srcpart)srcpart.FieldSchema(name:hr, type:string, comment:null), ]
Index: ql/src/test/results/clientpositive/disable_file_format_check.q.out
===================================================================
--- ql/src/test/results/clientpositive/disable_file_format_check.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/disable_file_format_check.q.out (working copy)
@@ -18,13 +18,3 @@
POSTHOOK: query: load data local inpath '../data/files/kv1.txt' overwrite into table kv_fileformat_check_seq
POSTHOOK: type: LOAD
POSTHOOK: Output: default@kv_fileformat_check_seq
-PREHOOK: query: drop table kv_fileformat_check_seq
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table kv_fileformat_check_seq
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@kv_fileformat_check_seq
-PREHOOK: query: drop table kv_fileformat_check_txt
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table kv_fileformat_check_txt
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@kv_fileformat_check_txt
Index: ql/src/test/results/clientpositive/input38.q.out
===================================================================
--- ql/src/test/results/clientpositive/input38.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/input38.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table dest1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table dest1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE dest1(key STRING, value STRING) STORED AS TEXTFILE
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE dest1(key STRING, value STRING) STORED AS TEXTFILE
@@ -79,7 +75,7 @@
Move Operator
files:
hdfs directory: true
- destination: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-12-53_095_8197621741613421666/10000
+ destination: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-28-12_445_5695557680495519967/10000
Stage: Stage-0
Move Operator
@@ -94,7 +90,7 @@
Stage: Stage-2
Map Reduce
Alias -> Map Operator Tree:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-12-53_095_8197621741613421666/10002
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-28-12_445_5695557680495519967/10002
Reduce Output Operator
sort order:
Map-reduce partition columns:
@@ -141,11 +137,11 @@
PREHOOK: query: SELECT dest1.* FROM dest1
PREHOOK: type: QUERY
PREHOOK: Input: default@dest1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-12-57_512_6617436684028270134/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-28-15_024_5142138661614792399/10000
POSTHOOK: query: SELECT dest1.* FROM dest1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-12-57_512_6617436684028270134/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-28-15_024_5142138661614792399/10000
POSTHOOK: Lineage: dest1.key SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.value SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
238 val_238 3 7
@@ -648,10 +644,3 @@
400 val_400 3 7
200 val_200 3 7
97 val_97 3 7
-PREHOOK: query: drop table dest1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table dest1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@dest1
-POSTHOOK: Lineage: dest1.key SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.value SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
Index: ql/src/test/results/clientpositive/create_view.q.out
===================================================================
--- ql/src/test/results/clientpositive/create_view.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/create_view.q.out (working copy)
@@ -74,71 +74,67 @@
PREHOOK: type: DROPFUNCTION
POSTHOOK: query: DROP TEMPORARY FUNCTION test_explode
POSTHOOK: type: DROPFUNCTION
-PREHOOK: query: DROP TABLE table1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE table1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: SELECT * FROM src WHERE key=86
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-07-51_249_8922015804755915765/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-18-57_921_7591165424312533742/10000
POSTHOOK: query: SELECT * FROM src WHERE key=86
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-07-51_249_8922015804755915765/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-18-57_921_7591165424312533742/10000
86 val_86
PREHOOK: query: CREATE VIEW view1 AS SELECT value FROM src WHERE key=86
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-07-55_984_3430099555842364804/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-00_421_2109283399258862165/10000
POSTHOOK: query: CREATE VIEW view1 AS SELECT value FROM src WHERE key=86
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-07-55_984_3430099555842364804/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-00_421_2109283399258862165/10000
POSTHOOK: Output: default@view1
PREHOOK: query: CREATE VIEW view2 AS SELECT * FROM src
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-07-56_249_108391416081868894/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-00_454_3605012159234387182/10000
POSTHOOK: query: CREATE VIEW view2 AS SELECT * FROM src
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-07-56_249_108391416081868894/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-00_454_3605012159234387182/10000
POSTHOOK: Output: default@view2
PREHOOK: query: CREATE VIEW view3(valoo)
TBLPROPERTIES ("fear" = "factor")
AS SELECT upper(value) FROM src WHERE key=86
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-07-56_437_8091674377507713014/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-00_480_4023689453233021298/10000
POSTHOOK: query: CREATE VIEW view3(valoo)
TBLPROPERTIES ("fear" = "factor")
AS SELECT upper(value) FROM src WHERE key=86
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-07-56_437_8091674377507713014/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-00_480_4023689453233021298/10000
POSTHOOK: Output: default@view3
PREHOOK: query: SELECT * from view1
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-07-56_507_8760093099562298778/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-00_511_3538329728957553516/10000
POSTHOOK: query: SELECT * from view1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-07-56_507_8760093099562298778/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-00_511_3538329728957553516/10000
val_86
PREHOOK: query: SELECT * from view2 where key=18
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-00_929_2355503113072117248/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-03_015_2577597731172760421/10000
POSTHOOK: query: SELECT * from view2 where key=18
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-00_929_2355503113072117248/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-03_015_2577597731172760421/10000
18 val_18
18 val_18
PREHOOK: query: SELECT * from view3
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-04_967_4103779794841926239/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-05_549_1542870472733242854/10000
POSTHOOK: query: SELECT * from view3
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-04_967_4103779794841926239/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-05_549_1542870472733242854/10000
VAL_86
PREHOOK: query: -- test EXPLAIN output for CREATE VIEW
EXPLAIN
@@ -239,7 +235,7 @@
POSTHOOK: type: DESCTABLE
value string
-Detailed Table Information Table(tableName:view1, dbName:default, owner:athusoo, createTime:1270516076, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:value, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1270516076}, viewOriginalText:SELECT value FROM src WHERE key=86, viewExpandedText:SELECT `src`.`value` FROM `src` WHERE `src`.`key`=86, tableType:VIRTUAL_VIEW)
+Detailed Table Information Table(tableName:view1, dbName:default, owner:jssarma, createTime:1279736340, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:value, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1279736340}, viewOriginalText:SELECT value FROM src WHERE key=86, viewExpandedText:SELECT `src`.`value` FROM `src` WHERE `src`.`key`=86, tableType:VIRTUAL_VIEW)
PREHOOK: query: DESCRIBE view2
PREHOOK: type: DESCTABLE
POSTHOOK: query: DESCRIBE view2
@@ -253,7 +249,7 @@
key string
value string
-Detailed Table Information Table(tableName:view2, dbName:default, owner:athusoo, createTime:1270516076, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1270516076}, viewOriginalText:SELECT * FROM src, viewExpandedText:SELECT `src`.`key`, `src`.`value` FROM `src`, tableType:VIRTUAL_VIEW)
+Detailed Table Information Table(tableName:view2, dbName:default, owner:jssarma, createTime:1279736340, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1279736340}, viewOriginalText:SELECT * FROM src, viewExpandedText:SELECT `src`.`key`, `src`.`value` FROM `src`, tableType:VIRTUAL_VIEW)
PREHOOK: query: DESCRIBE view3
PREHOOK: type: DESCTABLE
POSTHOOK: query: DESCRIBE view3
@@ -265,7 +261,7 @@
POSTHOOK: type: DESCTABLE
valoo string
-Detailed Table Information Table(tableName:view3, dbName:default, owner:athusoo, createTime:1270516076, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:valoo, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1270516076,fear=factor}, viewOriginalText:SELECT upper(value) FROM src WHERE key=86, viewExpandedText:SELECT `_c0` AS `valoo` FROM (SELECT upper(`src`.`value`) FROM `src` WHERE `src`.`key`=86) `view3`, tableType:VIRTUAL_VIEW)
+Detailed Table Information Table(tableName:view3, dbName:default, owner:jssarma, createTime:1279736340, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:valoo, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1279736340, fear=factor}, viewOriginalText:SELECT upper(value) FROM src WHERE key=86, viewExpandedText:SELECT `_c0` AS `valoo` FROM (SELECT upper(`src`.`value`) FROM `src` WHERE `src`.`key`=86) `view3`, tableType:VIRTUAL_VIEW)
PREHOOK: query: ALTER VIEW view3 SET TBLPROPERTIES ("biggest" = "loser")
PREHOOK: type: ALTERVIEW_PROPERTIES
POSTHOOK: query: ALTER VIEW view3 SET TBLPROPERTIES ("biggest" = "loser")
@@ -278,7 +274,7 @@
POSTHOOK: type: DESCTABLE
valoo string
-Detailed Table Information Table(tableName:view3, dbName:default, owner:athusoo, createTime:1270516076, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:valoo, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{last_modified_by=athusoo,last_modified_time=1270516090,biggest=loser,transient_lastDdlTime=1270516090,fear=factor}, viewOriginalText:SELECT upper(value) FROM src WHERE key=86, viewExpandedText:SELECT `_c0` AS `valoo` FROM (SELECT upper(`src`.`value`) FROM `src` WHERE `src`.`key`=86) `view3`, tableType:VIRTUAL_VIEW)
+Detailed Table Information Table(tableName:view3, dbName:default, owner:jssarma, createTime:1279736340, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:valoo, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{last_modified_by=jssarma, last_modified_time=1279736348, biggest=loser, transient_lastDdlTime=1279736348, fear=factor}, viewOriginalText:SELECT upper(value) FROM src WHERE key=86, viewExpandedText:SELECT `_c0` AS `valoo` FROM (SELECT upper(`src`.`value`) FROM `src` WHERE `src`.`key`=86) `view3`, tableType:VIRTUAL_VIEW)
PREHOOK: query: CREATE TABLE table1 (key int)
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE table1 (key int)
@@ -294,7 +290,7 @@
POSTHOOK: type: DESCTABLE
key int
-Detailed Table Information Table(tableName:table1, dbName:default, owner:athusoo, createTime:1270516090, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:int, comment:null)], location:file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/test/data/warehouse/table1, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1270516090}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)
+Detailed Table Information Table(tableName:table1, dbName:default, owner:jssarma, createTime:1279736349, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:int, comment:null)], location:file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/table1, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1279736349}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)
PREHOOK: query: DESCRIBE EXTENDED src1
PREHOOK: type: DESCTABLE
POSTHOOK: query: DESCRIBE EXTENDED src1
@@ -302,7 +298,7 @@
key string default
value string default
-Detailed Table Information Table(tableName:src1, dbName:default, owner:null, createTime:1270516069, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:default), FieldSchema(name:value, type:string, comment:default)], location:file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/test/data/warehouse/src1, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1270516069}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)
+Detailed Table Information Table(tableName:src1, dbName:default, owner:null, createTime:1279735684, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:default), FieldSchema(name:value, type:string, comment:default)], location:file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/src1, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1279735685}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)
PREHOOK: query: -- use DESCRIBE EXTENDED on a base table as a point of comparison for
-- view descriptions
DESCRIBE EXTENDED table1
@@ -313,7 +309,7 @@
POSTHOOK: type: DESCTABLE
key int
-Detailed Table Information Table(tableName:table1, dbName:default, owner:athusoo, createTime:1270516090, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:int, comment:null)], location:file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/test/data/warehouse/table1, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1270516090}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)
+Detailed Table Information Table(tableName:table1, dbName:default, owner:jssarma, createTime:1279736349, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:int, comment:null)], location:file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/table1, inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1279736349}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE)
PREHOOK: query: INSERT OVERWRITE TABLE table1 SELECT key FROM src WHERE key = 86
PREHOOK: type: QUERY
PREHOOK: Input: default@src
@@ -326,29 +322,29 @@
PREHOOK: query: SELECT * FROM table1
PREHOOK: type: QUERY
PREHOOK: Input: default@table1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-14_929_2065389972532664204/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-12_054_6185622979764502547/10000
POSTHOOK: query: SELECT * FROM table1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@table1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-14_929_2065389972532664204/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-12_054_6185622979764502547/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
86
PREHOOK: query: CREATE VIEW view4 AS SELECT * FROM table1
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-14_977_3015839202999327586/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-12_107_373418218143741165/10000
POSTHOOK: query: CREATE VIEW view4 AS SELECT * FROM table1
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-14_977_3015839202999327586/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-12_107_373418218143741165/10000
POSTHOOK: Output: default@view4
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
PREHOOK: query: SELECT * FROM view4
PREHOOK: type: QUERY
PREHOOK: Input: default@table1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-15_021_9184931574613208716/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-12_159_5736721309726450669/10000
POSTHOOK: query: SELECT * FROM view4
POSTHOOK: type: QUERY
POSTHOOK: Input: default@table1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-15_021_9184931574613208716/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-12_159_5736721309726450669/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
86
PREHOOK: query: DESCRIBE view4
@@ -367,21 +363,21 @@
PREHOOK: query: SELECT * FROM table1
PREHOOK: type: QUERY
PREHOOK: Input: default@table1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-19_105_5319518691601572932/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-14_884_6210545047734052322/10000
POSTHOOK: query: SELECT * FROM table1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@table1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-19_105_5319518691601572932/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-14_884_6210545047734052322/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
86 NULL
PREHOOK: query: SELECT * FROM view4
PREHOOK: type: QUERY
PREHOOK: Input: default@table1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-19_169_7551758402210110903/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-14_937_5926817710279534745/10000
POSTHOOK: query: SELECT * FROM view4
POSTHOOK: type: QUERY
POSTHOOK: Input: default@table1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-19_169_7551758402210110903/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-14_937_5926817710279534745/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
86
PREHOOK: query: DESCRIBE table1
@@ -400,21 +396,21 @@
PREHOOK: query: CREATE VIEW view5 AS SELECT v1.key as key1, v2.key as key2
FROM view4 v1 join view4 v2
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-23_879_1296898702625082815/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-17_650_1581979006666486814/10000
POSTHOOK: query: CREATE VIEW view5 AS SELECT v1.key as key1, v2.key as key2
FROM view4 v1 join view4 v2
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-23_879_1296898702625082815/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-17_650_1581979006666486814/10000
POSTHOOK: Output: default@view5
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
PREHOOK: query: SELECT * FROM view5
PREHOOK: type: QUERY
PREHOOK: Input: default@table1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-24_036_4857028717949657394/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-18_112_5156849292712269194/10000
POSTHOOK: query: SELECT * FROM view5
POSTHOOK: type: QUERY
POSTHOOK: Input: default@table1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-24_036_4857028717949657394/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-18_112_5156849292712269194/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
86 86
PREHOOK: query: DESCRIBE view5
@@ -429,13 +425,13 @@
CREATE VIEW view6(valoo COMMENT 'I cannot spell') AS
SELECT upper(value) as blarg FROM src WHERE key=86
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-29_046_710120165826100548/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-20_848_5361966232039317942/10000
POSTHOOK: query: -- verify that column name and comment in DDL portion
-- overrides column alias in SELECT
CREATE VIEW view6(valoo COMMENT 'I cannot spell') AS
SELECT upper(value) as blarg FROM src WHERE key=86
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-29_046_710120165826100548/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-20_848_5361966232039317942/10000
POSTHOOK: Output: default@view6
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
PREHOOK: query: DESCRIBE view6
@@ -451,7 +447,7 @@
ORDER BY key, value
LIMIT 10
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-29_265_1329913665876875445/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-21_026_524519195839168130/10000
POSTHOOK: query: -- verify that ORDER BY and LIMIT are both supported in view def
CREATE VIEW view7 AS
SELECT * FROM src
@@ -459,17 +455,17 @@
ORDER BY key, value
LIMIT 10
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-29_265_1329913665876875445/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-21_026_524519195839168130/10000
POSTHOOK: Output: default@view7
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
PREHOOK: query: SELECT * FROM view7
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-29_421_7101883282785856237/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-21_142_2584542483416509596/10000
POSTHOOK: query: SELECT * FROM view7
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-29_421_7101883282785856237/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-21_142_2584542483416509596/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
82 val_82
83 val_83
@@ -487,14 +483,14 @@
SELECT * FROM view7 ORDER BY key DESC, value
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-34_345_8066508064336105515/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-23_826_7709192776984057918/10000
POSTHOOK: query: -- top-level ORDER BY should override the one inside the view
-- (however, the inside ORDER BY should still influence the evaluation
-- of the limit)
SELECT * FROM view7 ORDER BY key DESC, value
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-34_345_8066508064336105515/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-23_826_7709192776984057918/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
90 val_90
90 val_90
@@ -510,12 +506,12 @@
SELECT * FROM view7 LIMIT 5
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-43_427_4325024040225755725/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-29_094_7201546132722765129/10000
POSTHOOK: query: -- top-level LIMIT should override if lower
SELECT * FROM view7 LIMIT 5
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-43_427_4325024040225755725/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-29_094_7201546132722765129/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
82 val_82
83 val_83
@@ -526,12 +522,12 @@
SELECT * FROM view7 LIMIT 20
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-49_324_8898744127113466306/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-31_669_4716567495375910798/10000
POSTHOOK: query: -- but not if higher
SELECT * FROM view7 LIMIT 20
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-49_324_8898744127113466306/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-31_669_4716567495375910798/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
82 val_82
83 val_83
@@ -556,12 +552,12 @@
SELECT test_translate('abc', 'a', 'b')
FROM table1
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-53_840_2964907692712697676/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-34_300_8840699164640193963/10000
POSTHOOK: query: CREATE VIEW view8(c) AS
SELECT test_translate('abc', 'a', 'b')
FROM table1
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-53_840_2964907692712697676/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-34_300_8840699164640193963/10000
POSTHOOK: Output: default@view8
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
PREHOOK: query: DESCRIBE EXTENDED view8
@@ -571,17 +567,17 @@
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
c string
-Detailed Table Information Table(tableName:view8, dbName:default, owner:athusoo, createTime:1270516133, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:c, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1270516133}, viewOriginalText:SELECT test_translate('abc', 'a', 'b')
+Detailed Table Information Table(tableName:view8, dbName:default, owner:jssarma, createTime:1279736374, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:c, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1279736374}, viewOriginalText:SELECT test_translate('abc', 'a', 'b')
FROM table1, viewExpandedText:SELECT `_c0` AS `c` FROM (SELECT `test_translate`('abc', 'a', 'b')
FROM `table1`) `view8`, tableType:VIRTUAL_VIEW)
PREHOOK: query: SELECT * FROM view8
PREHOOK: type: QUERY
PREHOOK: Input: default@table1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-53_924_7826463408574432406/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-34_450_3256858177096010496/10000
POSTHOOK: query: SELECT * FROM view8
POSTHOOK: type: QUERY
POSTHOOK: Input: default@table1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-53_924_7826463408574432406/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-34_450_3256858177096010496/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
bbc
PREHOOK: query: -- test usage of a UDAF within a view
@@ -597,12 +593,12 @@
SELECT test_max(length(value))
FROM src
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-58_747_2163708409319152021/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-36_956_3482633729704300290/10000
POSTHOOK: query: CREATE VIEW view9(m) AS
SELECT test_max(length(value))
FROM src
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-58_747_2163708409319152021/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-36_956_3482633729704300290/10000
POSTHOOK: Output: default@view9
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
PREHOOK: query: DESCRIBE EXTENDED view9
@@ -612,29 +608,29 @@
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
m int
-Detailed Table Information Table(tableName:view9, dbName:default, owner:athusoo, createTime:1270516138, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:m, type:int, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1270516138}, viewOriginalText:SELECT test_max(length(value))
+Detailed Table Information Table(tableName:view9, dbName:default, owner:jssarma, createTime:1279736376, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:m, type:int, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1279736376}, viewOriginalText:SELECT test_max(length(value))
FROM src, viewExpandedText:SELECT `_c0` AS `m` FROM (SELECT `test_max`(length(`src`.`value`))
FROM `src`) `view9`, tableType:VIRTUAL_VIEW)
PREHOOK: query: SELECT * FROM view9
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-59_002_4960552290114258251/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-37_098_8931758755865355991/10000
POSTHOOK: query: SELECT * FROM view9
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-08-59_002_4960552290114258251/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-37_098_8931758755865355991/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
7
PREHOOK: query: -- test usage of a subselect within a view
CREATE VIEW view10 AS
SELECT slurp.* FROM (SELECT * FROM src WHERE key=86) slurp
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-03_445_2038640404964168331/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-39_728_8538567105732339323/10000
POSTHOOK: query: -- test usage of a subselect within a view
CREATE VIEW view10 AS
SELECT slurp.* FROM (SELECT * FROM src WHERE key=86) slurp
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-03_445_2038640404964168331/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-39_728_8538567105732339323/10000
POSTHOOK: Output: default@view10
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
PREHOOK: query: DESCRIBE EXTENDED view10
@@ -645,15 +641,15 @@
key string
value string
-Detailed Table Information Table(tableName:view10, dbName:default, owner:athusoo, createTime:1270516143, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1270516143}, viewOriginalText:SELECT slurp.* FROM (SELECT * FROM src WHERE key=86) slurp, viewExpandedText:SELECT `slurp`.`key`, `slurp`.`value` FROM (SELECT `src`.`key`, `src`.`value` FROM `src` WHERE `src`.`key`=86) `slurp`, tableType:VIRTUAL_VIEW)
+Detailed Table Information Table(tableName:view10, dbName:default, owner:jssarma, createTime:1279736379, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1279736379}, viewOriginalText:SELECT slurp.* FROM (SELECT * FROM src WHERE key=86) slurp, viewExpandedText:SELECT `slurp`.`key`, `slurp`.`value` FROM (SELECT `src`.`key`, `src`.`value` FROM `src` WHERE `src`.`key`=86) `slurp`, tableType:VIRTUAL_VIEW)
PREHOOK: query: SELECT * FROM view10
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-03_534_4513987569168219590/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-39_954_121140881764531324/10000
POSTHOOK: query: SELECT * FROM view10
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-03_534_4513987569168219590/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-39_954_121140881764531324/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
86 val_86
PREHOOK: query: -- test usage of a UDTF within a view
@@ -669,12 +665,12 @@
SELECT test_explode(array(1,2,3)) AS (boom)
FROM table1
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-08_172_6937271646873245957/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-42_507_941005864769279541/10000
POSTHOOK: query: CREATE VIEW view11 AS
SELECT test_explode(array(1,2,3)) AS (boom)
FROM table1
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-08_172_6937271646873245957/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-42_507_941005864769279541/10000
POSTHOOK: Output: default@view11
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
PREHOOK: query: DESCRIBE EXTENDED view11
@@ -684,17 +680,17 @@
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
boom int
-Detailed Table Information Table(tableName:view11, dbName:default, owner:athusoo, createTime:1270516148, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:boom, type:int, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1270516148}, viewOriginalText:SELECT test_explode(array(1,2,3)) AS (boom)
+Detailed Table Information Table(tableName:view11, dbName:default, owner:jssarma, createTime:1279736382, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:boom, type:int, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1279736382}, viewOriginalText:SELECT test_explode(array(1,2,3)) AS (boom)
FROM table1, viewExpandedText:SELECT `test_explode`(array(1,2,3)) AS (`boom`)
FROM `table1`, tableType:VIRTUAL_VIEW)
PREHOOK: query: SELECT * FROM view11
PREHOOK: type: QUERY
PREHOOK: Input: default@table1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-08_313_4664224396935808807/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-42_621_147481859612435905/10000
POSTHOOK: query: SELECT * FROM view11
POSTHOOK: type: QUERY
POSTHOOK: Input: default@table1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-08_313_4664224396935808807/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-42_621_147481859612435905/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
1
2
@@ -703,12 +699,12 @@
CREATE VIEW view12 AS
SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-12_286_8427548702068975898/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-45_155_5582073945596947542/10000
POSTHOOK: query: -- test usage of LATERAL within a view
CREATE VIEW view12 AS
SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-12_286_8427548702068975898/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-45_155_5582073945596947542/10000
POSTHOOK: Output: default@view12
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
PREHOOK: query: DESCRIBE EXTENDED view12
@@ -720,17 +716,17 @@
value string
mycol int
-Detailed Table Information Table(tableName:view12, dbName:default, owner:athusoo, createTime:1270516152, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null), FieldSchema(name:mycol, type:int, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1270516152}, viewOriginalText:SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol, viewExpandedText:SELECT `src`.`key`, `src`.`value`, `mytable`.`mycol` FROM `src` LATERAL VIEW explode(array(1,2,3)) `myTable` AS `myCol`, tableType:VIRTUAL_VIEW)
+Detailed Table Information Table(tableName:view12, dbName:default, owner:jssarma, createTime:1279736385, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value, type:string, comment:null), FieldSchema(name:mycol, type:int, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1279736385}, viewOriginalText:SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol, viewExpandedText:SELECT `src`.`key`, `src`.`value`, `mytable`.`mycol` FROM `src` LATERAL VIEW explode(array(1,2,3)) `myTable` AS `myCol`, tableType:VIRTUAL_VIEW)
PREHOOK: query: SELECT * FROM view12
ORDER BY key ASC, myCol ASC LIMIT 1
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-12_406_2588118547562629364/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-45_288_6471519897685463611/10000
POSTHOOK: query: SELECT * FROM view12
ORDER BY key ASC, myCol ASC LIMIT 1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-12_406_2588118547562629364/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-45_288_6471519897685463611/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
0 val_0 1
PREHOOK: query: -- test usage of LATERAL with a view as the LHS
@@ -738,13 +734,13 @@
ORDER BY key ASC, myCol ASC LIMIT 1
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-17_007_2569567271321396526/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-47_967_4711091040062160152/10000
POSTHOOK: query: -- test usage of LATERAL with a view as the LHS
SELECT * FROM view2 LATERAL VIEW explode(array(1,2,3)) myTable AS myCol
ORDER BY key ASC, myCol ASC LIMIT 1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-17_007_2569567271321396526/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-47_967_4711091040062160152/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
0 val_0 1
PREHOOK: query: -- test usage of TABLESAMPLE within a view
@@ -752,13 +748,13 @@
SELECT s.key
FROM srcbucket TABLESAMPLE (BUCKET 1 OUT OF 5 ON key) s
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-21_328_3318797810129325812/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-50_665_3879547564260377848/10000
POSTHOOK: query: -- test usage of TABLESAMPLE within a view
CREATE VIEW view13 AS
SELECT s.key
FROM srcbucket TABLESAMPLE (BUCKET 1 OUT OF 5 ON key) s
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-21_328_3318797810129325812/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-50_665_3879547564260377848/10000
POSTHOOK: Output: default@view13
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
PREHOOK: query: DESCRIBE EXTENDED view13
@@ -768,19 +764,19 @@
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
key int
-Detailed Table Information Table(tableName:view13, dbName:default, owner:athusoo, createTime:1270516161, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:int, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1270516161}, viewOriginalText:SELECT s.key
+Detailed Table Information Table(tableName:view13, dbName:default, owner:jssarma, createTime:1279736390, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:int, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1279736390}, viewOriginalText:SELECT s.key
FROM srcbucket TABLESAMPLE (BUCKET 1 OUT OF 5 ON key) s, viewExpandedText:SELECT `s`.`key`
FROM `srcbucket` TABLESAMPLE (BUCKET 1 OUT OF 5 ON `key`) `s`, tableType:VIRTUAL_VIEW)
PREHOOK: query: SELECT * FROM view13
ORDER BY key LIMIT 12
PREHOOK: type: QUERY
PREHOOK: Input: default@srcbucket
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-21_411_3195453265987665126/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-50_808_1357112183697677370/10000
POSTHOOK: query: SELECT * FROM view13
ORDER BY key LIMIT 12
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcbucket
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-21_411_3195453265987665126/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-50_808_1357112183697677370/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
0
0
@@ -807,7 +803,7 @@
select s4.key as key, s4.value as value from src s4 where s4.key < 10) unionsrc2
ON (unionsrc1.key = unionsrc2.key)
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-25_913_4823304300133994589/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-53_382_1837567325290565612/10000
POSTHOOK: query: -- test usage of JOIN+UNION+AGG all within same view
CREATE VIEW view14 AS
SELECT unionsrc1.key as k1, unionsrc1.value as v1,
@@ -821,7 +817,7 @@
select s4.key as key, s4.value as value from src s4 where s4.key < 10) unionsrc2
ON (unionsrc1.key = unionsrc2.key)
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-25_913_4823304300133994589/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-53_382_1837567325290565612/10000
POSTHOOK: Output: default@view14
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
PREHOOK: query: DESCRIBE EXTENDED view14
@@ -834,7 +830,7 @@
k2 string
v2 string
-Detailed Table Information Table(tableName:view14, dbName:default, owner:athusoo, createTime:1270516166, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:k1, type:string, comment:null), FieldSchema(name:v1, type:string, comment:null), FieldSchema(name:k2, type:string, comment:null), FieldSchema(name:v2, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1270516166}, viewOriginalText:SELECT unionsrc1.key as k1, unionsrc1.value as v1,
+Detailed Table Information Table(tableName:view14, dbName:default, owner:jssarma, createTime:1279736393, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:k1, type:string, comment:null), FieldSchema(name:v1, type:string, comment:null), FieldSchema(name:k2, type:string, comment:null), FieldSchema(name:v2, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1279736393}, viewOriginalText:SELECT unionsrc1.key as k1, unionsrc1.value as v1,
unionsrc2.key as k2, unionsrc2.value as v2
FROM (select 'tst1' as key, cast(count(1) as string) as value from src s1
UNION ALL
@@ -857,12 +853,12 @@
ORDER BY k1
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-26_119_2761258753248431304/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-53_542_3266426952139641562/10000
POSTHOOK: query: SELECT * FROM view14
ORDER BY k1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-26_119_2761258753248431304/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-19-53_542_3266426952139641562/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
0 val_0 0 val_0
0 val_0 0 val_0
@@ -893,14 +889,14 @@
FROM src
GROUP BY key
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-58_798_7555941970013621980/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-20-12_519_296683891105463355/10000
POSTHOOK: query: -- test usage of GROUP BY within view
CREATE VIEW view15 AS
SELECT key,COUNT(value) AS value_count
FROM src
GROUP BY key
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-58_798_7555941970013621980/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-20-12_519_296683891105463355/10000
POSTHOOK: Output: default@view15
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
PREHOOK: query: DESCRIBE EXTENDED view15
@@ -911,7 +907,7 @@
key string
value_count bigint
-Detailed Table Information Table(tableName:view15, dbName:default, owner:athusoo, createTime:1270516198, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value_count, type:bigint, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1270516198}, viewOriginalText:SELECT key,COUNT(value) AS value_count
+Detailed Table Information Table(tableName:view15, dbName:default, owner:jssarma, createTime:1279736412, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:key, type:string, comment:null), FieldSchema(name:value_count, type:bigint, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1279736412}, viewOriginalText:SELECT key,COUNT(value) AS value_count
FROM src
GROUP BY key, viewExpandedText:SELECT `src`.`key`,COUNT(`src`.`value`) AS `value_count`
FROM `src`
@@ -921,13 +917,13 @@
LIMIT 10
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-58_914_3308702425437265515/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-20-12_632_2632827347997497908/10000
POSTHOOK: query: SELECT * FROM view15
ORDER BY value_count DESC, key
LIMIT 10
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-09-58_914_3308702425437265515/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-20-12_632_2632827347997497908/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
230 5
348 5
@@ -944,13 +940,13 @@
SELECT DISTINCT value
FROM src
PREHOOK: type: CREATEVIEW
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-10-07_365_8609706981588395246/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-20-17_744_4594166188705296834/10000
POSTHOOK: query: -- test usage of DISTINCT within view
CREATE VIEW view16 AS
SELECT DISTINCT value
FROM src
POSTHOOK: type: CREATEVIEW
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-10-07_365_8609706981588395246/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-20-17_744_4594166188705296834/10000
POSTHOOK: Output: default@view16
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
PREHOOK: query: DESCRIBE EXTENDED view16
@@ -960,7 +956,7 @@
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
value string
-Detailed Table Information Table(tableName:view16, dbName:default, owner:athusoo, createTime:1270516207, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:value, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1270516207}, viewOriginalText:SELECT DISTINCT value
+Detailed Table Information Table(tableName:view16, dbName:default, owner:jssarma, createTime:1279736417, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:value, type:string, comment:null)], location:null, inputFormat:org.apache.hadoop.mapred.SequenceFileInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat, compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:null, parameters:{}), bucketCols:[], sortCols:[], parameters:{}), partitionKeys:[], parameters:{transient_lastDdlTime=1279736417}, viewOriginalText:SELECT DISTINCT value
FROM src, viewExpandedText:SELECT DISTINCT `src`.`value`
FROM `src`, tableType:VIRTUAL_VIEW)
PREHOOK: query: SELECT * FROM view16
@@ -968,13 +964,13 @@
LIMIT 10
PREHOOK: type: QUERY
PREHOOK: Input: default@src
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-10-07_470_4237294142974167361/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-20-17_856_3789669241902298524/10000
POSTHOOK: query: SELECT * FROM view16
ORDER BY value
LIMIT 10
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_1/build/ql/scratchdir/hive_2010-04-05_18-10-07_470_4237294142974167361/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-20-17_856_3789669241902298524/10000
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
val_0
val_10
@@ -988,17 +984,15 @@
val_114
PREHOOK: query: -- this should work since currently we don't track view->table
-- dependencies for implementing RESTRICT
-DROP TABLE table1
-PREHOOK: type: DROPTABLE
+
+
+DROP VIEW view1
+PREHOOK: type: DROPVIEW
POSTHOOK: query: -- this should work since currently we don't track view->table
-- dependencies for implementing RESTRICT
-DROP TABLE table1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@table1
-POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-PREHOOK: query: DROP VIEW view1
-PREHOOK: type: DROPVIEW
-POSTHOOK: query: DROP VIEW view1
+
+
+DROP VIEW view1
POSTHOOK: type: DROPVIEW
POSTHOOK: Output: default@view1
POSTHOOK: Lineage: table1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
Index: ql/src/test/results/clientpositive/groupby3_noskew.q.out
===================================================================
--- ql/src/test/results/clientpositive/groupby3_noskew.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/groupby3_noskew.q.out (working copy)
@@ -171,11 +171,11 @@
PREHOOK: query: SELECT dest1.* FROM dest1
PREHOOK: type: QUERY
PREHOOK: Input: default@dest1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-09-38_039_6030452822453768062/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-23-47_426_1452039016187035932/10000
POSTHOOK: query: SELECT dest1.* FROM dest1
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_2/build/ql/scratchdir/hive_2010-04-05_18-09-38_039_6030452822453768062/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-23-47_426_1452039016187035932/10000
POSTHOOK: Lineage: dest1.c1 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.c2 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.c3 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
@@ -186,17 +186,3 @@
POSTHOOK: Lineage: dest1.c8 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
POSTHOOK: Lineage: dest1.c9 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
130091.0 260.182 256.10355987055016 98.0 0.0 142.92680950752379 143.06995106518903 20428.07287599999 20469.010897795582
-PREHOOK: query: DROP TABLE dest1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE dest1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@dest1
-POSTHOOK: Lineage: dest1.c1 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.c2 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.c3 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.c4 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.c5 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.c6 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.c7 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.c8 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: dest1.c9 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ]
Index: ql/src/test/results/clientpositive/implicit_cast1.q.out
===================================================================
--- ql/src/test/results/clientpositive/implicit_cast1.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/implicit_cast1.q.out (working copy)
@@ -59,15 +59,10 @@
WHERE implicit_test1.a <> 0
PREHOOK: type: QUERY
PREHOOK: Input: default@implicit_test1
-PREHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/439003851/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-26-24_976_2073073634228618169/10000
POSTHOOK: query: SELECT implicit_test1.*
FROM implicit_test1
WHERE implicit_test1.a <> 0
POSTHOOK: type: QUERY
POSTHOOK: Input: default@implicit_test1
-POSTHOOK: Output: file:/data/users/njain/hive5/hive5/build/ql/tmp/439003851/10000
-PREHOOK: query: DROP TABLE implicit_test1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE implicit_test1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@implicit_test1
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-26-24_976_2073073634228618169/10000
Index: ql/src/test/results/clientpositive/input24.q.out
===================================================================
--- ql/src/test/results/clientpositive/input24.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/input24.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table tst
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table tst
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table tst(a int, b int) partitioned by (d string)
PREHOOK: type: CREATETABLE
POSTHOOK: query: create table tst(a int, b int) partitioned by (d string)
@@ -80,14 +76,9 @@
PREHOOK: query: select count(1) from tst x where x.d='2009-01-01'
PREHOOK: type: QUERY
PREHOOK: Input: default@tst@d=2009-01-01
-PREHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1844534662/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-27-18_997_2954899292364767462/10000
POSTHOOK: query: select count(1) from tst x where x.d='2009-01-01'
POSTHOOK: type: QUERY
POSTHOOK: Input: default@tst@d=2009-01-01
-POSTHOOK: Output: file:/data/users/heyongqiang/hive-trunk-commit/build/ql/tmp/1844534662/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-27-18_997_2954899292364767462/10000
0
-PREHOOK: query: drop table tst
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table tst
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@tst
Index: ql/src/test/results/clientpositive/join_hive_626.q.out
===================================================================
--- ql/src/test/results/clientpositive/join_hive_626.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/join_hive_626.q.out (working copy)
@@ -1,15 +1,3 @@
-PREHOOK: query: drop table hive_foo
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table hive_foo
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: drop table hive_bar
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table hive_bar
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: drop table hive_count
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table hive_count
-POSTHOOK: type: DROPTABLE
PREHOOK: query: create table hive_foo (foo_id int, foo_name string, foo_a string, foo_b string,
foo_c string, foo_d string) row format delimited fields terminated by ','
stored as textfile
@@ -187,27 +175,12 @@
PREHOOK: Input: default@hive_foo
PREHOOK: Input: default@hive_count
PREHOOK: Input: default@hive_bar
-PREHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-27-47_249_881439767445940507/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-35-37_392_1859121185876168833/10000
POSTHOOK: query: select hive_foo.foo_name, hive_bar.bar_name, n from hive_foo join hive_bar on hive_foo.foo_id =
hive_bar.foo_id join hive_count on hive_count.bar_id = hive_bar.bar_id
POSTHOOK: type: QUERY
POSTHOOK: Input: default@hive_foo
POSTHOOK: Input: default@hive_count
POSTHOOK: Input: default@hive_bar
-POSTHOOK: Output: file:/data/users/zshao/hadoop_hive_trunk2/.ptest_2/build/ql/scratchdir/hive_2010-02-12_22-27-47_249_881439767445940507/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-35-37_392_1859121185876168833/10000
foo1 bar10 2
-PREHOOK: query: drop table hive_foo
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table hive_foo
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@hive_foo
-PREHOOK: query: drop table hive_bar
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table hive_bar
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@hive_bar
-PREHOOK: query: drop table hive_count
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table hive_count
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@hive_count
Index: ql/src/test/results/clientpositive/rcfile_lazydecompress.q.out
===================================================================
--- ql/src/test/results/clientpositive/rcfile_lazydecompress.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/rcfile_lazydecompress.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE rcfileTableLazyDecompress
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE rcfileTableLazyDecompress
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE table rcfileTableLazyDecompress (key STRING, value STRING) STORED AS RCFile
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE table rcfileTableLazyDecompress (key STRING, value STRING) STORED AS RCFile
@@ -22,11 +18,11 @@
PREHOOK: query: SELECT key, value FROM rcfileTableLazyDecompress where key > 238
PREHOOK: type: QUERY
PREHOOK: Input: default@rcfiletablelazydecompress
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-25-08_126_1758618048562072306/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-48-26_308_5035766417306203353/10000
POSTHOOK: query: SELECT key, value FROM rcfileTableLazyDecompress where key > 238
POSTHOOK: type: QUERY
POSTHOOK: Input: default@rcfiletablelazydecompress
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-25-08_126_1758618048562072306/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-48-26_308_5035766417306203353/10000
POSTHOOK: Lineage: rcfiletablelazydecompress.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: rcfiletablelazydecompress.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
311 val_311
@@ -37,11 +33,11 @@
PREHOOK: query: SELECT key, value FROM rcfileTableLazyDecompress where key > 238 and key < 400
PREHOOK: type: QUERY
PREHOOK: Input: default@rcfiletablelazydecompress
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-25-12_756_7047055078902721683/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-48-28_826_2664555248442666518/10000
POSTHOOK: query: SELECT key, value FROM rcfileTableLazyDecompress where key > 238 and key < 400
POSTHOOK: type: QUERY
POSTHOOK: Input: default@rcfiletablelazydecompress
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-25-12_756_7047055078902721683/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-48-28_826_2664555248442666518/10000
POSTHOOK: Lineage: rcfiletablelazydecompress.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: rcfiletablelazydecompress.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
311 val_311
@@ -50,11 +46,11 @@
PREHOOK: query: SELECT key, count(1) FROM rcfileTableLazyDecompress where key > 238 group by key
PREHOOK: type: QUERY
PREHOOK: Input: default@rcfiletablelazydecompress
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-25-16_962_6661530342559954352/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-48-31_347_4980542755724589562/10000
POSTHOOK: query: SELECT key, count(1) FROM rcfileTableLazyDecompress where key > 238 group by key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@rcfiletablelazydecompress
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-25-16_962_6661530342559954352/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-48-31_347_4980542755724589562/10000
POSTHOOK: Lineage: rcfiletablelazydecompress.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: rcfiletablelazydecompress.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
255 1
@@ -79,11 +75,11 @@
PREHOOK: query: SELECT key, value FROM rcfileTableLazyDecompress where key > 238
PREHOOK: type: QUERY
PREHOOK: Input: default@rcfiletablelazydecompress
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-25-24_793_7750074014093688381/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-48-36_558_7043959496739147343/10000
POSTHOOK: query: SELECT key, value FROM rcfileTableLazyDecompress where key > 238
POSTHOOK: type: QUERY
POSTHOOK: Input: default@rcfiletablelazydecompress
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-25-24_793_7750074014093688381/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-48-36_558_7043959496739147343/10000
POSTHOOK: Lineage: rcfiletablelazydecompress.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: rcfiletablelazydecompress.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: rcfiletablelazydecompress.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
@@ -96,11 +92,11 @@
PREHOOK: query: SELECT key, value FROM rcfileTableLazyDecompress where key > 238 and key < 400
PREHOOK: type: QUERY
PREHOOK: Input: default@rcfiletablelazydecompress
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-25-28_589_3116278472072779381/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-48-39_022_510504722430321903/10000
POSTHOOK: query: SELECT key, value FROM rcfileTableLazyDecompress where key > 238 and key < 400
POSTHOOK: type: QUERY
POSTHOOK: Input: default@rcfiletablelazydecompress
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-25-28_589_3116278472072779381/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-48-39_022_510504722430321903/10000
POSTHOOK: Lineage: rcfiletablelazydecompress.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: rcfiletablelazydecompress.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: rcfiletablelazydecompress.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
@@ -111,11 +107,11 @@
PREHOOK: query: SELECT key, count(1) FROM rcfileTableLazyDecompress where key > 238 group by key
PREHOOK: type: QUERY
PREHOOK: Input: default@rcfiletablelazydecompress
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-25-33_415_4146404104589934880/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-48-41_564_865533805813240026/10000
POSTHOOK: query: SELECT key, count(1) FROM rcfileTableLazyDecompress where key > 238 group by key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@rcfiletablelazydecompress
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-25-33_415_4146404104589934880/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-48-41_564_865533805813240026/10000
POSTHOOK: Lineage: rcfiletablelazydecompress.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: rcfiletablelazydecompress.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: rcfiletablelazydecompress.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
@@ -125,12 +121,3 @@
311 1
409 1
484 1
-PREHOOK: query: DROP TABLE rcfileTableLazyDecompress
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE rcfileTableLazyDecompress
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@rcfiletablelazydecompress
-POSTHOOK: Lineage: rcfiletablelazydecompress.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: rcfiletablelazydecompress.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: rcfiletablelazydecompress.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
-POSTHOOK: Lineage: rcfiletablelazydecompress.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
Index: ql/src/test/results/clientpositive/input10.q.out
===================================================================
--- ql/src/test/results/clientpositive/input10.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/input10.q.out (working copy)
@@ -35,8 +35,3 @@
value string
ds string
hr string
-PREHOOK: query: DROP TABLE TEST10
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE TEST10
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@test10
Index: ql/src/test/results/clientpositive/archive.q.out
===================================================================
--- ql/src/test/results/clientpositive/archive.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/archive.q.out (working copy)
@@ -5,7 +5,7 @@
PREHOOK: type: QUERY
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-PREHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-29-12_801_8718664231713136788/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-19_808_5920190331482198474/10000
POSTHOOK: query: -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
SELECT SUM(hash(col)) FROM (SELECT transform(*) using 'tr "\t" "_"' AS col
@@ -13,7 +13,7 @@
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-POSTHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-29-12_801_8718664231713136788/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-19_808_5920190331482198474/10000
48479881068
PREHOOK: query: ALTER TABLE srcpart ARCHIVE PARTITION (ds='2008-04-08', hr='12')
PREHOOK: type: ALTERTABLE_ARCHIVE
@@ -24,35 +24,35 @@
PREHOOK: type: QUERY
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-PREHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-29-20_510_5269010142014944519/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-24_163_645957732896284359/10000
POSTHOOK: query: SELECT SUM(hash(col)) FROM (SELECT transform(*) using 'tr "\t" "_"' AS col
FROM (SELECT * FROM srcpart WHERE ds='2008-04-08') subq1) subq2
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-POSTHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-29-20_510_5269010142014944519/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-24_163_645957732896284359/10000
48479881068
PREHOOK: query: SELECT key, count(1) FROM srcpart WHERE ds='2008-04-08' AND hr='12' AND key='0' GROUP BY key
PREHOOK: type: QUERY
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-PREHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-29-26_238_1201801305984652550/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-27_091_7903410270275371868/10000
POSTHOOK: query: SELECT key, count(1) FROM srcpart WHERE ds='2008-04-08' AND hr='12' AND key='0' GROUP BY key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-POSTHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-29-26_238_1201801305984652550/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-27_091_7903410270275371868/10000
0 3
PREHOOK: query: SELECT * FROM srcpart a JOIN src b ON a.key=b.key
WHERE a.ds='2008-04-08' AND a.hr='12' AND a.key='0'
PREHOOK: type: QUERY
PREHOOK: Input: default@src
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-PREHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-29-32_413_8808816186480793926/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-30_052_8857134521946599131/10000
POSTHOOK: query: SELECT * FROM srcpart a JOIN src b ON a.key=b.key
WHERE a.ds='2008-04-08' AND a.hr='12' AND a.key='0'
POSTHOOK: type: QUERY
POSTHOOK: Input: default@src
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-POSTHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-29-32_413_8808816186480793926/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-30_052_8857134521946599131/10000
0 val_0 2008-04-08 12 0 val_0
0 val_0 2008-04-08 12 0 val_0
0 val_0 2008-04-08 12 0 val_0
@@ -71,13 +71,13 @@
PREHOOK: type: QUERY
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-PREHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-29-37_857_7662280812791374354/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-33_403_2031008368522834404/10000
POSTHOOK: query: SELECT SUM(hash(col)) FROM (SELECT transform(*) using 'tr "\t" "_"' AS col
FROM (SELECT * FROM srcpart WHERE ds='2008-04-08') subq1) subq2
POSTHOOK: type: QUERY
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11
POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
-POSTHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-29-37_857_7662280812791374354/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-33_403_2031008368522834404/10000
48479881068
PREHOOK: query: CREATE TABLE harbucket(key INT)
PARTITIONED by (ds STRING)
@@ -100,11 +100,11 @@
PREHOOK: query: SELECT key FROM harbucket TABLESAMPLE(BUCKET 1 OUT OF 10) SORT BY key
PREHOOK: type: QUERY
PREHOOK: Input: default@harbucket@ds=1
-PREHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-29-47_247_5412318794268628077/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-39_158_8278614490047409485/10000
POSTHOOK: query: SELECT key FROM harbucket TABLESAMPLE(BUCKET 1 OUT OF 10) SORT BY key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@harbucket@ds=1
-POSTHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-29-47_247_5412318794268628077/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-39_158_8278614490047409485/10000
POSTHOOK: Lineage: harbucket PARTITION(ds=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
0
0
@@ -120,11 +120,11 @@
PREHOOK: query: SELECT key FROM harbucket TABLESAMPLE(BUCKET 1 OUT OF 10) SORT BY key
PREHOOK: type: QUERY
PREHOOK: Input: default@harbucket@ds=1
-PREHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-29-52_696_6661366062442712305/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-43_343_1319156393955499510/10000
POSTHOOK: query: SELECT key FROM harbucket TABLESAMPLE(BUCKET 1 OUT OF 10) SORT BY key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@harbucket@ds=1
-POSTHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-29-52_696_6661366062442712305/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-43_343_1319156393955499510/10000
POSTHOOK: Lineage: harbucket PARTITION(ds=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
0
0
@@ -140,11 +140,11 @@
PREHOOK: query: SELECT key FROM harbucket TABLESAMPLE(BUCKET 1 OUT OF 10) SORT BY key
PREHOOK: type: QUERY
PREHOOK: Input: default@harbucket@ds=1
-PREHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-29-56_920_7660869602739278397/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-46_534_5753932776517458176/10000
POSTHOOK: query: SELECT key FROM harbucket TABLESAMPLE(BUCKET 1 OUT OF 10) SORT BY key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@harbucket@ds=1
-POSTHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-29-56_920_7660869602739278397/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-46_534_5753932776517458176/10000
POSTHOOK: Lineage: harbucket PARTITION(ds=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
0
0
@@ -152,12 +152,6 @@
10
20
30
-PREHOOK: query: DROP TABLE harbucket
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE harbucket
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@harbucket
-POSTHOOK: Lineage: harbucket PARTITION(ds=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
PREHOOK: query: CREATE TABLE old_name(key INT)
PARTITIONED by (ds STRING)
PREHOOK: type: CREATETABLE
@@ -186,12 +180,12 @@
FROM (SELECT * FROM old_name WHERE ds='1') subq1) subq2
PREHOOK: type: QUERY
PREHOOK: Input: default@old_name@ds=1
-PREHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-30-06_143_8274193601305228676/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-53_532_5887928415090778760/10000
POSTHOOK: query: SELECT SUM(hash(col)) FROM (SELECT transform(*) using 'tr "\t" "_"' AS col
FROM (SELECT * FROM old_name WHERE ds='1') subq1) subq2
POSTHOOK: type: QUERY
POSTHOOK: Input: default@old_name@ds=1
-POSTHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-30-06_143_8274193601305228676/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-53_532_5887928415090778760/10000
POSTHOOK: Lineage: harbucket PARTITION(ds=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: old_name PARTITION(ds=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
48656137
@@ -207,19 +201,12 @@
FROM (SELECT * FROM new_name WHERE ds='1') subq1) subq2
PREHOOK: type: QUERY
PREHOOK: Input: default@new_name@ds=1
-PREHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-30-10_661_5999329953207292038/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-56_372_476351158155800671/10000
POSTHOOK: query: SELECT SUM(hash(col)) FROM (SELECT transform(*) using 'tr "\t" "_"' AS col
FROM (SELECT * FROM new_name WHERE ds='1') subq1) subq2
POSTHOOK: type: QUERY
POSTHOOK: Input: default@new_name@ds=1
-POSTHOOK: Output: file:/data/users/pyang/task/trunk/VENDOR.hive/trunk/build/ql/scratchdir/hive_2010-06-21_17-30-10_661_5999329953207292038/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-08-56_372_476351158155800671/10000
POSTHOOK: Lineage: harbucket PARTITION(ds=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: old_name PARTITION(ds=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
NULL
-PREHOOK: query: DROP TABLE new_name
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE new_name
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@new_name
-POSTHOOK: Lineage: harbucket PARTITION(ds=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: old_name PARTITION(ds=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
Index: ql/src/test/results/clientpositive/repair.q.out
===================================================================
--- ql/src/test/results/clientpositive/repair.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/repair.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE repairtable
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE repairtable
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING)
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE repairtable(col STRING) PARTITIONED BY (p1 STRING, p2 STRING)
@@ -27,8 +23,3 @@
PREHOOK: type: MSCK
POSTHOOK: query: MSCK TABLE repairtable
POSTHOOK: type: MSCK
-PREHOOK: query: DROP TABLE repairtable
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE repairtable
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@repairtable
Index: ql/src/test/results/clientpositive/bucket1.q.out
===================================================================
--- ql/src/test/results/clientpositive/bucket1.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/bucket1.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: drop table bucket1_1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table bucket1_1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE bucket1_1(key int, value string) CLUSTERED BY (key) INTO 100 BUCKETS
PREHOOK: type: CREATETABLE
POSTHOOK: query: CREATE TABLE bucket1_1(key int, value string) CLUSTERED BY (key) INTO 100 BUCKETS
@@ -49,9 +45,9 @@
type: string
Needs Tagging: false
Path -> Alias:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/src [src]
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/src [src]
Path -> Partition:
- file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/src
+ file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/src
Partition
base file name: src
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -62,12 +58,12 @@
columns.types string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/src
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/src
name src
serialization.ddl struct src { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270515753
+ transient_lastDdlTime 1279735685
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -78,12 +74,12 @@
columns.types string:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/src
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/src
name src
serialization.ddl struct src { string key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270515753
+ transient_lastDdlTime 1279735685
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: src
name: src
@@ -99,7 +95,7 @@
File Output Operator
compressed: false
GlobalTableId: 1
- directory: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-02-34_525_873235169631577914/10000
+ directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-09-06_147_3859020501578469948/10000
NumFilesPerFileSink: 1
table:
input format: org.apache.hadoop.mapred.TextInputFormat
@@ -111,12 +107,12 @@
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/bucket1_1
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucket1_1
name bucket1_1
serialization.ddl struct bucket1_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270515754
+ transient_lastDdlTime 1279735746
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucket1_1
TotalFiles: 1
@@ -126,7 +122,7 @@
Move Operator
tables:
replace: true
- source: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-02-34_525_873235169631577914/10000
+ source: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-09-06_147_3859020501578469948/10000
table:
input format: org.apache.hadoop.mapred.TextInputFormat
output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
@@ -137,15 +133,15 @@
columns.types int:string
file.inputformat org.apache.hadoop.mapred.TextInputFormat
file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
- location file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/test/data/warehouse/bucket1_1
+ location file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/test/data/warehouse/bucket1_1
name bucket1_1
serialization.ddl struct bucket1_1 { i32 key, string value}
serialization.format 1
serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
- transient_lastDdlTime 1270515754
+ transient_lastDdlTime 1279735746
serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
name: bucket1_1
- tmp directory: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-02-34_525_873235169631577914/10001
+ tmp directory: file:/mnt/vol/devrs004.snc1/jssarma/projects/hive_trunk/build/ql/scratchdir/hive_2010-07-21_11-09-06_147_3859020501578469948/10001
PREHOOK: query: insert overwrite table bucket1_1
@@ -163,11 +159,11 @@
PREHOOK: query: select * from bucket1_1 order by key
PREHOOK: type: QUERY
PREHOOK: Input: default@bucket1_1
-PREHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-02-39_954_6102377035167270592/10000
+PREHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-08_981_1141988062127111870/10000
POSTHOOK: query: select * from bucket1_1 order by key
POSTHOOK: type: QUERY
POSTHOOK: Input: default@bucket1_1
-POSTHOOK: Output: file:/data/users/athusoo/apache_workspaces/hive_trunk_ws1/.ptest_0/build/ql/scratchdir/hive_2010-04-05_18-02-39_954_6102377035167270592/10000
+POSTHOOK: Output: file:/tmp/jssarma/hive_2010-07-21_11-09-08_981_1141988062127111870/10000
POSTHOOK: Lineage: bucket1_1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
POSTHOOK: Lineage: bucket1_1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
0 val_0
@@ -670,10 +666,3 @@
498 val_498
498 val_498
498 val_498
-PREHOOK: query: drop table bucket1_1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: drop table bucket1_1
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Output: default@bucket1_1
-POSTHOOK: Lineage: bucket1_1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ]
-POSTHOOK: Lineage: bucket1_1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ]
Index: ql/src/test/results/clientpositive/create_nested_type.q.out
===================================================================
--- ql/src/test/results/clientpositive/create_nested_type.q.out (revision 966856)
+++ ql/src/test/results/clientpositive/create_nested_type.q.out (working copy)
@@ -1,7 +1,3 @@
-PREHOOK: query: DROP TABLE table1
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE table1
-POSTHOOK: type: DROPTABLE
PREHOOK: query: CREATE TABLE table1 (
a STRING,
b ARRAY,
@@ -34,7 +30,7 @@
c array