Index: hbase-handler/src/test/results/hbase_bulk.m.out =================================================================== --- hbase-handler/src/test/results/hbase_bulk.m.out (revision 1239976) +++ hbase-handler/src/test/results/hbase_bulk.m.out (working copy) @@ -1,132 +0,0 @@ -PREHOOK: query: drop table hbsort -PREHOOK: type: DROPTABLE -POSTHOOK: query: drop table hbsort -POSTHOOK: type: DROPTABLE -PREHOOK: query: drop table hbpartition -PREHOOK: type: DROPTABLE -POSTHOOK: query: drop table hbpartition -POSTHOOK: type: DROPTABLE -PREHOOK: query: -- this is a dummy table used for controlling how the HFiles are --- created -create table hbsort(key string, val string, val2 string) -stored as -INPUTFORMAT 'org.apache.hadoop.mapred.TextInputFormat' -OUTPUTFORMAT 'org.apache.hadoop.hive.hbase.HiveHFileOutputFormat' -#### A masked pattern was here #### -PREHOOK: type: CREATETABLE -POSTHOOK: query: -- this is a dummy table used for controlling how the HFiles are --- created -create table hbsort(key string, val string, val2 string) -stored as -INPUTFORMAT 'org.apache.hadoop.mapred.TextInputFormat' -OUTPUTFORMAT 'org.apache.hadoop.hive.hbase.HiveHFileOutputFormat' -#### A masked pattern was here #### -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@hbsort -PREHOOK: query: -- this is a dummy table used for controlling how the input file --- for TotalOrderPartitioner is created -create table hbpartition(part_break string) -row format serde -'org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe' -stored as -inputformat -'org.apache.hadoop.mapred.TextInputFormat' -outputformat -'org.apache.hadoop.hive.ql.io.HiveNullValueSequenceFileOutputFormat' -PREHOOK: type: CREATETABLE -POSTHOOK: query: -- this is a dummy table used for controlling how the input file --- for TotalOrderPartitioner is created -create table hbpartition(part_break string) -row format serde -'org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe' -stored as -inputformat -'org.apache.hadoop.mapred.TextInputFormat' -outputformat -'org.apache.hadoop.hive.ql.io.HiveNullValueSequenceFileOutputFormat' -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@hbpartition -PREHOOK: query: -- this should produce one file, but we do not --- know what it will be called, so we will copy it to a well known -#### A masked pattern was here #### -insert overwrite table hbpartition -select distinct value -from src -where value='val_100' or value='val_200' -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@hbpartition -POSTHOOK: query: -- this should produce one file, but we do not --- know what it will be called, so we will copy it to a well known -#### A masked pattern was here #### -insert overwrite table hbpartition -select distinct value -from src -where value='val_100' or value='val_200' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@hbpartition -POSTHOOK: Lineage: hbpartition.part_break SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -#### A masked pattern was here #### --- include some trailing blanks and nulls to make sure we handle them correctly -insert overwrite table hbsort -select distinct value, - case when key=103 then cast(null as string) else key end, - case when key=103 then '' - else cast(key+1 as string) end -from src -cluster by value -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@hbsort -#### A masked pattern was here #### --- include some trailing blanks and nulls to make sure we handle them correctly -insert overwrite table hbsort -select distinct value, - case when key=103 then cast(null as string) else key end, - case when key=103 then '' - else cast(key+1 as string) end -from src -cluster by value -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@hbsort -POSTHOOK: Lineage: hbpartition.part_break SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: hbsort.key SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: hbsort.val EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: hbsort.val2 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -#### A masked pattern was here #### -PREHOOK: query: -- To get the files out to your local filesystem for loading into -#### A masked pattern was here #### --- semicolon-terminate the line below before running this test: -#### A masked pattern was here #### - -drop table hbsort -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@hbsort -PREHOOK: Output: default@hbsort -POSTHOOK: query: -- To get the files out to your local filesystem for loading into -#### A masked pattern was here #### --- semicolon-terminate the line below before running this test: -#### A masked pattern was here #### - -drop table hbsort -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@hbsort -POSTHOOK: Output: default@hbsort -POSTHOOK: Lineage: hbpartition.part_break SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: hbsort.key SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: hbsort.val EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: hbsort.val2 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -PREHOOK: query: drop table hbpartition -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@hbpartition -PREHOOK: Output: default@hbpartition -POSTHOOK: query: drop table hbpartition -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@hbpartition -POSTHOOK: Output: default@hbpartition -POSTHOOK: Lineage: hbpartition.part_break SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: hbsort.key SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: hbsort.val EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: hbsort.val2 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] Index: hbase-handler/src/test/results/positive/table_create.q.out =================================================================== --- hbase-handler/src/test/results/positive/table_create.q.out (revision 0) +++ hbase-handler/src/test/results/positive/table_create.q.out (revision 0) @@ -0,0 +1,15 @@ +PREHOOK: query: DROP TABLE hbase_table_1 +PREHOOK: type: DROPTABLE +POSTHOOK: query: DROP TABLE hbase_table_1 +POSTHOOK: type: DROPTABLE +PREHOOK: query: CREATE TABLE hbase_table_1(key int, value string) +STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' +WITH SERDEPROPERTIES ("hbase.columns.mapping" = "cf:string") +TBLPROPERTIES ("hbase.table.name" = "hbase_table_0") +PREHOOK: type: CREATETABLE +POSTHOOK: query: CREATE TABLE hbase_table_1(key int, value string) +STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' +WITH SERDEPROPERTIES ("hbase.columns.mapping" = "cf:string") +TBLPROPERTIES ("hbase.table.name" = "hbase_table_0") +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: default@hbase_table_1 Index: hbase-handler/src/test/results/negative/cascade_dbdrop.q.out =================================================================== --- hbase-handler/src/test/results/negative/cascade_dbdrop.q.out (revision 0) +++ hbase-handler/src/test/results/negative/cascade_dbdrop.q.out (revision 0) @@ -0,0 +1,21 @@ +PREHOOK: query: CREATE DATABASE hbaseDB +PREHOOK: type: CREATEDATABASE +POSTHOOK: query: CREATE DATABASE hbaseDB +POSTHOOK: type: CREATEDATABASE +PREHOOK: query: CREATE TABLE hbaseDB.hbase_table_0(key int, value string) +STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' +WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,cf:string") +TBLPROPERTIES ("hbase.table.name" = "hbase_table_0") +PREHOOK: type: CREATETABLE +POSTHOOK: query: CREATE TABLE hbaseDB.hbase_table_0(key int, value string) +STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' +WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,cf:string") +TBLPROPERTIES ("hbase.table.name" = "hbase_table_0") +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: hbaseDB@hbase_table_0 +PREHOOK: query: DROP DATABASE IF EXISTS hbaseDB CASCADE +PREHOOK: type: DROPDATABASE +POSTHOOK: query: DROP DATABASE IF EXISTS hbaseDB CASCADE +POSTHOOK: type: DROPDATABASE +Command failed with exit code = 1 +#### A masked pattern was here #### Index: hbase-handler/src/test/results/hbase_queries.q.out =================================================================== --- hbase-handler/src/test/results/hbase_queries.q.out (revision 1239976) +++ hbase-handler/src/test/results/hbase_queries.q.out (working copy) @@ -1,980 +0,0 @@ -PREHOOK: query: DROP TABLE hbase_table_1 -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE hbase_table_1 -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE TABLE hbase_table_1(key int, value string) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = "cf:string") -TBLPROPERTIES ("hbase.table.name" = "hbase_table_0") -PREHOOK: type: CREATETABLE -POSTHOOK: query: CREATE TABLE hbase_table_1(key int, value string) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = "cf:string") -TBLPROPERTIES ("hbase.table.name" = "hbase_table_0") -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@hbase_table_1 -PREHOOK: query: DESCRIBE EXTENDED hbase_table_1 -PREHOOK: type: DESCTABLE -POSTHOOK: query: DESCRIBE EXTENDED hbase_table_1 -POSTHOOK: type: DESCTABLE -key int from deserializer -value string from deserializer - -#### A masked pattern was here #### -PREHOOK: query: select * from hbase_table_1 -PREHOOK: type: QUERY -PREHOOK: Input: default@hbase_table_1 -#### A masked pattern was here #### -POSTHOOK: query: select * from hbase_table_1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@hbase_table_1 -#### A masked pattern was here #### -PREHOOK: query: EXPLAIN FROM src INSERT OVERWRITE TABLE hbase_table_1 SELECT * WHERE (key%2)=0 -PREHOOK: type: QUERY -POSTHOOK: query: EXPLAIN FROM src INSERT OVERWRITE TABLE hbase_table_1 SELECT * WHERE (key%2)=0 -POSTHOOK: type: QUERY -ABSTRACT SYNTAX TREE: - (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME hbase_table_1))) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (= (% (TOK_TABLE_OR_COL key) 2) 0)))) - -STAGE DEPENDENCIES: - Stage-0 is a root stage - -STAGE PLANS: - Stage: Stage-0 - Map Reduce - Alias -> Map Operator Tree: - src - TableScan - alias: src - Filter Operator - predicate: - expr: ((key % 2) = 0) - type: boolean - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: UDFToInteger(_col0) - type: int - expr: _col1 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.hive.hbase.HiveHBaseTableInputFormat - output format: org.apache.hadoop.hive.hbase.HiveHBaseTableOutputFormat - serde: org.apache.hadoop.hive.hbase.HBaseSerDe - name: default.hbase_table_1 - - -PREHOOK: query: FROM src INSERT OVERWRITE TABLE hbase_table_1 SELECT * WHERE (key%2)=0 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@hbase_table_1 -POSTHOOK: query: FROM src INSERT OVERWRITE TABLE hbase_table_1 SELECT * WHERE (key%2)=0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@hbase_table_1 -PREHOOK: query: DROP TABLE hbase_table_2 -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE hbase_table_2 -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE EXTERNAL TABLE hbase_table_2(key int, value string) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = "cf:string") -TBLPROPERTIES ("hbase.table.name" = "hbase_table_0") -PREHOOK: type: CREATETABLE -POSTHOOK: query: CREATE EXTERNAL TABLE hbase_table_2(key int, value string) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = "cf:string") -TBLPROPERTIES ("hbase.table.name" = "hbase_table_0") -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@hbase_table_2 -PREHOOK: query: EXPLAIN -SELECT Y.* -FROM -(SELECT hbase_table_1.* FROM hbase_table_1) x -JOIN -(SELECT src.* FROM src) Y -ON (x.key = Y.key) -ORDER BY key, value LIMIT 20 -PREHOOK: type: QUERY -POSTHOOK: query: EXPLAIN -SELECT Y.* -FROM -(SELECT hbase_table_1.* FROM hbase_table_1) x -JOIN -(SELECT src.* FROM src) Y -ON (x.key = Y.key) -ORDER BY key, value LIMIT 20 -POSTHOOK: type: QUERY -ABSTRACT SYNTAX TREE: - (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME hbase_table_1))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME hbase_table_1)))))) x) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME src)))))) Y) (= (. (TOK_TABLE_OR_COL x) key) (. (TOK_TABLE_OR_COL Y) key)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME Y)))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL value))) (TOK_LIMIT 20))) - -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 is a root stage - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Alias -> Map Operator Tree: - x:hbase_table_1 - TableScan - alias: hbase_table_1 - Select Operator - expressions: - expr: key - type: int - outputColumnNames: _col0 - Reduce Output Operator - key expressions: - expr: UDFToDouble(_col0) - type: double - sort order: + - Map-reduce partition columns: - expr: UDFToDouble(_col0) - type: double - tag: 0 - y:src - TableScan - alias: src - Select Operator - expressions: - expr: key - type: string - expr: value - type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: UDFToDouble(_col0) - type: double - sort order: + - Map-reduce partition columns: - expr: UDFToDouble(_col0) - type: double - tag: 1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - condition expressions: - 0 - 1 {VALUE._col0} {VALUE._col1} - handleSkewJoin: false - outputColumnNames: _col2, _col3 - Select Operator - expressions: - expr: _col2 - type: string - expr: _col3 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - Stage: Stage-2 - Map Reduce - Alias -> Map Operator Tree: -#### A masked pattern was here #### - Reduce Output Operator - key expressions: - expr: _col0 - type: string - expr: _col1 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: string - expr: _col1 - type: string - Reduce Operator Tree: - Extract - Limit - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - Stage: Stage-0 - Fetch Operator - limit: 20 - - -PREHOOK: query: SELECT Y.* -FROM -(SELECT hbase_table_1.* FROM hbase_table_1) x -JOIN -(SELECT src.* FROM src) Y -ON (x.key = Y.key) -ORDER BY key, value LIMIT 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@hbase_table_1 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: SELECT Y.* -FROM -(SELECT hbase_table_1.* FROM hbase_table_1) x -JOIN -(SELECT src.* FROM src) Y -ON (x.key = Y.key) -ORDER BY key, value LIMIT 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@hbase_table_1 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 val_0 -0 val_0 -0 val_0 -10 val_10 -100 val_100 -100 val_100 -104 val_104 -104 val_104 -114 val_114 -116 val_116 -118 val_118 -118 val_118 -12 val_12 -12 val_12 -120 val_120 -120 val_120 -126 val_126 -128 val_128 -128 val_128 -128 val_128 -PREHOOK: query: EXPLAIN -SELECT Y.* -FROM -(SELECT hbase_table_1.* FROM hbase_table_1 WHERE hbase_table_1.key > 100) x -JOIN -(SELECT hbase_table_2.* FROM hbase_table_2 WHERE hbase_table_2.key < 120) Y -ON (x.key = Y.key) -ORDER BY key, value -PREHOOK: type: QUERY -POSTHOOK: query: EXPLAIN -SELECT Y.* -FROM -(SELECT hbase_table_1.* FROM hbase_table_1 WHERE hbase_table_1.key > 100) x -JOIN -(SELECT hbase_table_2.* FROM hbase_table_2 WHERE hbase_table_2.key < 120) Y -ON (x.key = Y.key) -ORDER BY key, value -POSTHOOK: type: QUERY -ABSTRACT SYNTAX TREE: - (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME hbase_table_1))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME hbase_table_1)))) (TOK_WHERE (> (. (TOK_TABLE_OR_COL hbase_table_1) key) 100)))) x) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME hbase_table_2))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME hbase_table_2)))) (TOK_WHERE (< (. (TOK_TABLE_OR_COL hbase_table_2) key) 120)))) Y) (= (. (TOK_TABLE_OR_COL x) key) (. (TOK_TABLE_OR_COL Y) key)))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME Y)))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL key)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL value))))) - -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 is a root stage - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Alias -> Map Operator Tree: - x:hbase_table_1 - TableScan - alias: hbase_table_1 - Filter Operator - predicate: - expr: (key > 100) - type: boolean - Select Operator - expressions: - expr: key - type: int - outputColumnNames: _col0 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - sort order: + - Map-reduce partition columns: - expr: _col0 - type: int - tag: 0 - y:hbase_table_2 - TableScan - alias: hbase_table_2 - Filter Operator - predicate: - expr: (key < 120) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: int - sort order: + - Map-reduce partition columns: - expr: _col0 - type: int - tag: 1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - condition expressions: - 0 - 1 {VALUE._col0} {VALUE._col1} - handleSkewJoin: false - outputColumnNames: _col2, _col3 - Select Operator - expressions: - expr: _col2 - type: int - expr: _col3 - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - Stage: Stage-2 - Map Reduce - Alias -> Map Operator Tree: -#### A masked pattern was here #### - Reduce Output Operator - key expressions: - expr: _col0 - type: int - expr: _col1 - type: string - sort order: ++ - tag: -1 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string - Reduce Operator Tree: - Extract - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - Stage: Stage-0 - Fetch Operator - limit: -1 - - -PREHOOK: query: SELECT Y.* -FROM -(SELECT hbase_table_1.* FROM hbase_table_1 WHERE hbase_table_1.key > 100) x -JOIN -(SELECT hbase_table_2.* FROM hbase_table_2 WHERE hbase_table_2.key < 120) Y -ON (x.key = Y.key) -ORDER BY key,value -PREHOOK: type: QUERY -PREHOOK: Input: default@hbase_table_1 -PREHOOK: Input: default@hbase_table_2 -#### A masked pattern was here #### -POSTHOOK: query: SELECT Y.* -FROM -(SELECT hbase_table_1.* FROM hbase_table_1 WHERE hbase_table_1.key > 100) x -JOIN -(SELECT hbase_table_2.* FROM hbase_table_2 WHERE hbase_table_2.key < 120) Y -ON (x.key = Y.key) -ORDER BY key,value -POSTHOOK: type: QUERY -POSTHOOK: Input: default@hbase_table_1 -POSTHOOK: Input: default@hbase_table_2 -#### A masked pattern was here #### -104 val_104 -114 val_114 -116 val_116 -118 val_118 -PREHOOK: query: DROP TABLE empty_hbase_table -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE empty_hbase_table -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE TABLE empty_hbase_table(key int, value string) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = "cf:string") -PREHOOK: type: CREATETABLE -POSTHOOK: query: CREATE TABLE empty_hbase_table(key int, value string) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = "cf:string") -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@empty_hbase_table -PREHOOK: query: DROP TABLE empty_normal_table -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE empty_normal_table -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE TABLE empty_normal_table(key int, value string) -PREHOOK: type: CREATETABLE -POSTHOOK: query: CREATE TABLE empty_normal_table(key int, value string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@empty_normal_table -PREHOOK: query: select * from (select count(1) as c from empty_normal_table union all select count(1) as c from empty_hbase_table) x order by c -PREHOOK: type: QUERY -PREHOOK: Input: default@empty_hbase_table -PREHOOK: Input: default@empty_normal_table -#### A masked pattern was here #### -POSTHOOK: query: select * from (select count(1) as c from empty_normal_table union all select count(1) as c from empty_hbase_table) x order by c -POSTHOOK: type: QUERY -POSTHOOK: Input: default@empty_hbase_table -POSTHOOK: Input: default@empty_normal_table -#### A masked pattern was here #### -0 -0 -PREHOOK: query: select * from (select count(1) c from empty_normal_table union all select count(1) as c from hbase_table_1) x order by c -PREHOOK: type: QUERY -PREHOOK: Input: default@empty_normal_table -PREHOOK: Input: default@hbase_table_1 -#### A masked pattern was here #### -POSTHOOK: query: select * from (select count(1) c from empty_normal_table union all select count(1) as c from hbase_table_1) x order by c -POSTHOOK: type: QUERY -POSTHOOK: Input: default@empty_normal_table -POSTHOOK: Input: default@hbase_table_1 -#### A masked pattern was here #### -0 -155 -PREHOOK: query: select * from (select count(1) c from src union all select count(1) as c from empty_hbase_table) x order by c -PREHOOK: type: QUERY -PREHOOK: Input: default@empty_hbase_table -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select * from (select count(1) c from src union all select count(1) as c from empty_hbase_table) x order by c -POSTHOOK: type: QUERY -POSTHOOK: Input: default@empty_hbase_table -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 -500 -PREHOOK: query: select * from (select count(1) c from src union all select count(1) as c from hbase_table_1) x order by c -PREHOOK: type: QUERY -PREHOOK: Input: default@hbase_table_1 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select * from (select count(1) c from src union all select count(1) as c from hbase_table_1) x order by c -POSTHOOK: type: QUERY -POSTHOOK: Input: default@hbase_table_1 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -155 -500 -PREHOOK: query: CREATE TABLE hbase_table_3(key int, value string, count int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "cf:val,cf2:count" -) -PREHOOK: type: CREATETABLE -POSTHOOK: query: CREATE TABLE hbase_table_3(key int, value string, count int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "cf:val,cf2:count" -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@hbase_table_3 -PREHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE hbase_table_3 -SELECT x.key, x.value, Y.count -FROM -(SELECT hbase_table_1.* FROM hbase_table_1) x -JOIN -(SELECT src.key, count(src.key) as count FROM src GROUP BY src.key) Y -ON (x.key = Y.key) -PREHOOK: type: QUERY -POSTHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE hbase_table_3 -SELECT x.key, x.value, Y.count -FROM -(SELECT hbase_table_1.* FROM hbase_table_1) x -JOIN -(SELECT src.key, count(src.key) as count FROM src GROUP BY src.key) Y -ON (x.key = Y.key) -POSTHOOK: type: QUERY -ABSTRACT SYNTAX TREE: - (TOK_QUERY (TOK_FROM (TOK_JOIN (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME hbase_table_1))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_ALLCOLREF (TOK_TABNAME hbase_table_1)))))) x) (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL src) key)) (TOK_SELEXPR (TOK_FUNCTION count (. (TOK_TABLE_OR_COL src) key)) count)) (TOK_GROUPBY (. (TOK_TABLE_OR_COL src) key)))) Y) (= (. (TOK_TABLE_OR_COL x) key) (. (TOK_TABLE_OR_COL Y) key)))) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME hbase_table_3))) (TOK_SELECT (TOK_SELEXPR (. (TOK_TABLE_OR_COL x) key)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL x) value)) (TOK_SELEXPR (. (TOK_TABLE_OR_COL Y) count))))) - -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Alias -> Map Operator Tree: - y:src - TableScan - alias: src - Select Operator - expressions: - expr: key - type: string - outputColumnNames: key - Group By Operator - aggregations: - expr: count(key) - bucketGroup: false - keys: - expr: key - type: string - mode: hash - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: _col0 - type: string - sort order: + - Map-reduce partition columns: - expr: _col0 - type: string - tag: -1 - value expressions: - expr: _col1 - type: bigint - Reduce Operator Tree: - Group By Operator - aggregations: - expr: count(VALUE._col0) - bucketGroup: false - keys: - expr: KEY._col0 - type: string - mode: mergepartial - outputColumnNames: _col0, _col1 - Select Operator - expressions: - expr: _col0 - type: string - expr: _col1 - type: bigint - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - - Stage: Stage-0 - Map Reduce - Alias -> Map Operator Tree: - $INTNAME - Reduce Output Operator - key expressions: - expr: UDFToDouble(_col0) - type: double - sort order: + - Map-reduce partition columns: - expr: UDFToDouble(_col0) - type: double - tag: 1 - value expressions: - expr: _col1 - type: bigint - x:hbase_table_1 - TableScan - alias: hbase_table_1 - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - Reduce Output Operator - key expressions: - expr: UDFToDouble(_col0) - type: double - sort order: + - Map-reduce partition columns: - expr: UDFToDouble(_col0) - type: double - tag: 0 - value expressions: - expr: _col0 - type: int - expr: _col1 - type: string - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - condition expressions: - 0 {VALUE._col0} {VALUE._col1} - 1 {VALUE._col1} - handleSkewJoin: false - outputColumnNames: _col0, _col1, _col3 - Select Operator - expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: _col3 - type: bigint - outputColumnNames: _col0, _col1, _col2 - Select Operator - expressions: - expr: _col0 - type: int - expr: _col1 - type: string - expr: UDFToInteger(_col2) - type: int - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - GlobalTableId: 1 - table: - input format: org.apache.hadoop.hive.hbase.HiveHBaseTableInputFormat - output format: org.apache.hadoop.hive.hbase.HiveHBaseTableOutputFormat - serde: org.apache.hadoop.hive.hbase.HBaseSerDe - name: default.hbase_table_3 - - -PREHOOK: query: INSERT OVERWRITE TABLE hbase_table_3 -SELECT x.key, x.value, Y.count -FROM -(SELECT hbase_table_1.* FROM hbase_table_1) x -JOIN -(SELECT src.key, count(src.key) as count FROM src GROUP BY src.key) Y -ON (x.key = Y.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@hbase_table_1 -PREHOOK: Input: default@src -PREHOOK: Output: default@hbase_table_3 -POSTHOOK: query: INSERT OVERWRITE TABLE hbase_table_3 -SELECT x.key, x.value, Y.count -FROM -(SELECT hbase_table_1.* FROM hbase_table_1) x -JOIN -(SELECT src.key, count(src.key) as count FROM src GROUP BY src.key) Y -ON (x.key = Y.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@hbase_table_1 -POSTHOOK: Input: default@src -POSTHOOK: Output: default@hbase_table_3 -PREHOOK: query: select count(1) from hbase_table_3 -PREHOOK: type: QUERY -PREHOOK: Input: default@hbase_table_3 -#### A masked pattern was here #### -POSTHOOK: query: select count(1) from hbase_table_3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@hbase_table_3 -#### A masked pattern was here #### -155 -PREHOOK: query: select * from hbase_table_3 order by key, value limit 5 -PREHOOK: type: QUERY -PREHOOK: Input: default@hbase_table_3 -#### A masked pattern was here #### -POSTHOOK: query: select * from hbase_table_3 order by key, value limit 5 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@hbase_table_3 -#### A masked pattern was here #### -0 val_0 3 -2 val_2 1 -4 val_4 1 -8 val_8 1 -10 val_10 1 -PREHOOK: query: select key, count from hbase_table_3 order by key, count desc limit 5 -PREHOOK: type: QUERY -PREHOOK: Input: default@hbase_table_3 -#### A masked pattern was here #### -POSTHOOK: query: select key, count from hbase_table_3 order by key, count desc limit 5 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@hbase_table_3 -#### A masked pattern was here #### -0 3 -2 1 -4 1 -8 1 -10 1 -PREHOOK: query: DROP TABLE hbase_table_4 -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE hbase_table_4 -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE TABLE hbase_table_4(key int, value1 string, value2 int, value3 int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "a:b,a:c,d:e" -) -PREHOOK: type: CREATETABLE -POSTHOOK: query: CREATE TABLE hbase_table_4(key int, value1 string, value2 int, value3 int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "a:b,a:c,d:e" -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@hbase_table_4 -PREHOOK: query: INSERT OVERWRITE TABLE hbase_table_4 SELECT key, value, key+1, key+2 -FROM src WHERE key=98 OR key=100 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@hbase_table_4 -POSTHOOK: query: INSERT OVERWRITE TABLE hbase_table_4 SELECT key, value, key+1, key+2 -FROM src WHERE key=98 OR key=100 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@hbase_table_4 -PREHOOK: query: SELECT * FROM hbase_table_4 ORDER BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@hbase_table_4 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM hbase_table_4 ORDER BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@hbase_table_4 -#### A masked pattern was here #### -98 val_98 99 100 -100 val_100 101 102 -PREHOOK: query: DROP TABLE hbase_table_5 -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE hbase_table_5 -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE EXTERNAL TABLE hbase_table_5(key int, value map) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = "a:") -TBLPROPERTIES ("hbase.table.name" = "hbase_table_4") -PREHOOK: type: CREATETABLE -POSTHOOK: query: CREATE EXTERNAL TABLE hbase_table_5(key int, value map) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = "a:") -TBLPROPERTIES ("hbase.table.name" = "hbase_table_4") -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@hbase_table_5 -PREHOOK: query: SELECT * FROM hbase_table_5 ORDER BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@hbase_table_5 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM hbase_table_5 ORDER BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@hbase_table_5 -#### A masked pattern was here #### -98 {"b":"val_98","c":"99"} -100 {"b":"val_100","c":"101"} -PREHOOK: query: DROP TABLE hbase_table_6 -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE hbase_table_6 -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE TABLE hbase_table_6(key int, value map) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = ":key,cf:" -) -PREHOOK: type: CREATETABLE -POSTHOOK: query: CREATE TABLE hbase_table_6(key int, value map) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = ":key,cf:" -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@hbase_table_6 -PREHOOK: query: INSERT OVERWRITE TABLE hbase_table_6 SELECT key, map(value, key) FROM src -WHERE key=98 OR key=100 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@hbase_table_6 -POSTHOOK: query: INSERT OVERWRITE TABLE hbase_table_6 SELECT key, map(value, key) FROM src -WHERE key=98 OR key=100 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@hbase_table_6 -PREHOOK: query: SELECT * FROM hbase_table_6 ORDER BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@hbase_table_6 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM hbase_table_6 ORDER BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@hbase_table_6 -#### A masked pattern was here #### -98 {"val_98":"98"} -100 {"val_100":"100"} -PREHOOK: query: DROP TABLE hbase_table_7 -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE hbase_table_7 -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE TABLE hbase_table_7(value map, key int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "cf:,:key" -) -PREHOOK: type: CREATETABLE -POSTHOOK: query: CREATE TABLE hbase_table_7(value map, key int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "cf:,:key" -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@hbase_table_7 -PREHOOK: query: INSERT OVERWRITE TABLE hbase_table_7 -SELECT map(value, key, upper(value), key+1), key FROM src -WHERE key=98 OR key=100 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@hbase_table_7 -POSTHOOK: query: INSERT OVERWRITE TABLE hbase_table_7 -SELECT map(value, key, upper(value), key+1), key FROM src -WHERE key=98 OR key=100 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@hbase_table_7 -PREHOOK: query: SELECT * FROM hbase_table_7 ORDER BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@hbase_table_7 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM hbase_table_7 ORDER BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@hbase_table_7 -#### A masked pattern was here #### -{"VAL_98":"99.0","val_98":"98"} 98 -{"VAL_100":"101.0","val_100":"100"} 100 -PREHOOK: query: DROP TABLE hbase_table_8 -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE hbase_table_8 -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE TABLE hbase_table_8(key int, value1 string, value2 int, value3 int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "a:b,a:c,d:e" -) -PREHOOK: type: CREATETABLE -POSTHOOK: query: CREATE TABLE hbase_table_8(key int, value1 string, value2 int, value3 int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "a:b,a:c,d:e" -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@hbase_table_8 -PREHOOK: query: INSERT OVERWRITE TABLE hbase_table_8 SELECT key, value, key+1, key+2 -FROM src WHERE key=98 OR key=100 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@hbase_table_8 -POSTHOOK: query: INSERT OVERWRITE TABLE hbase_table_8 SELECT key, value, key+1, key+2 -FROM src WHERE key=98 OR key=100 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@hbase_table_8 -PREHOOK: query: SELECT * FROM hbase_table_8 ORDER BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@hbase_table_8 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM hbase_table_8 ORDER BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@hbase_table_8 -#### A masked pattern was here #### -98 val_98 99 100 -100 val_100 101 102 -PREHOOK: query: DROP TABLE hbase_table_1 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@hbase_table_1 -PREHOOK: Output: default@hbase_table_1 -POSTHOOK: query: DROP TABLE hbase_table_1 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@hbase_table_1 -POSTHOOK: Output: default@hbase_table_1 -PREHOOK: query: DROP TABLE hbase_table_2 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@hbase_table_2 -PREHOOK: Output: default@hbase_table_2 -POSTHOOK: query: DROP TABLE hbase_table_2 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@hbase_table_2 -POSTHOOK: Output: default@hbase_table_2 -PREHOOK: query: DROP TABLE hbase_table_3 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@hbase_table_3 -PREHOOK: Output: default@hbase_table_3 -POSTHOOK: query: DROP TABLE hbase_table_3 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@hbase_table_3 -POSTHOOK: Output: default@hbase_table_3 -PREHOOK: query: DROP TABLE hbase_table_4 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@hbase_table_4 -PREHOOK: Output: default@hbase_table_4 -POSTHOOK: query: DROP TABLE hbase_table_4 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@hbase_table_4 -POSTHOOK: Output: default@hbase_table_4 -PREHOOK: query: DROP TABLE hbase_table_5 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@hbase_table_5 -PREHOOK: Output: default@hbase_table_5 -POSTHOOK: query: DROP TABLE hbase_table_5 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@hbase_table_5 -POSTHOOK: Output: default@hbase_table_5 -PREHOOK: query: DROP TABLE hbase_table_6 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@hbase_table_6 -PREHOOK: Output: default@hbase_table_6 -POSTHOOK: query: DROP TABLE hbase_table_6 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@hbase_table_6 -POSTHOOK: Output: default@hbase_table_6 -PREHOOK: query: DROP TABLE hbase_table_7 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@hbase_table_7 -PREHOOK: Output: default@hbase_table_7 -POSTHOOK: query: DROP TABLE hbase_table_7 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@hbase_table_7 -POSTHOOK: Output: default@hbase_table_7 -PREHOOK: query: DROP TABLE hbase_table_8 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@hbase_table_8 -PREHOOK: Output: default@hbase_table_8 -POSTHOOK: query: DROP TABLE hbase_table_8 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@hbase_table_8 -POSTHOOK: Output: default@hbase_table_8 -PREHOOK: query: DROP TABLE empty_hbase_table -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@empty_hbase_table -PREHOOK: Output: default@empty_hbase_table -POSTHOOK: query: DROP TABLE empty_hbase_table -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@empty_hbase_table -POSTHOOK: Output: default@empty_hbase_table -PREHOOK: query: DROP TABLE empty_normal_table -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@empty_normal_table -PREHOOK: Output: default@empty_normal_table -POSTHOOK: query: DROP TABLE empty_normal_table -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@empty_normal_table -POSTHOOK: Output: default@empty_normal_table Index: hbase-handler/src/test/results/hbase_stats.q.out =================================================================== --- hbase-handler/src/test/results/hbase_stats.q.out (revision 1239976) +++ hbase-handler/src/test/results/hbase_stats.q.out (working copy) @@ -1,390 +0,0 @@ -PREHOOK: query: create table stats_src like src -PREHOOK: type: CREATETABLE -POSTHOOK: query: create table stats_src like src -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@stats_src -PREHOOK: query: insert overwrite table stats_src select * from src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@stats_src -POSTHOOK: query: insert overwrite table stats_src select * from src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@stats_src -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: analyze table stats_src compute statistics -PREHOOK: type: QUERY -PREHOOK: Input: default@stats_src -PREHOOK: Output: default@stats_src -POSTHOOK: query: analyze table stats_src compute statistics -POSTHOOK: type: QUERY -POSTHOOK: Input: default@stats_src -POSTHOOK: Output: default@stats_src -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: desc formatted stats_src -PREHOOK: type: DESCTABLE -POSTHOOK: query: desc formatted stats_src -POSTHOOK: type: DESCTABLE -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default - -# Detailed Table Information -Database: default -#### A masked pattern was here #### -Protect Mode: None -Retention: 0 -#### A masked pattern was here #### -Table Type: MANAGED_TABLE -Table Parameters: - numFiles 1 - numPartitions 0 - numRows 500 - rawDataSize 5312 - totalSize 5812 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -InputFormat: org.apache.hadoop.mapred.TextInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Storage Desc Params: - serialization.format 1 -PREHOOK: query: create table stats_part like srcpart -PREHOOK: type: CREATETABLE -POSTHOOK: query: create table stats_part like srcpart -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@stats_part -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: insert overwrite table stats_part partition (ds='2010-04-08', hr = '11') select key, value from src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@stats_part@ds=2010-04-08/hr=11 -POSTHOOK: query: insert overwrite table stats_part partition (ds='2010-04-08', hr = '11') select key, value from src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@stats_part@ds=2010-04-08/hr=11 -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: insert overwrite table stats_part partition (ds='2010-04-08', hr = '12') select key, value from src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@stats_part@ds=2010-04-08/hr=12 -POSTHOOK: query: insert overwrite table stats_part partition (ds='2010-04-08', hr = '12') select key, value from src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@stats_part@ds=2010-04-08/hr=12 -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: analyze table stats_part partition(ds='2010-04-08', hr='11') compute statistics -PREHOOK: type: QUERY -PREHOOK: Input: default@stats_part@ds=2010-04-08/hr=11 -PREHOOK: Output: default@stats_part -PREHOOK: Output: default@stats_part@ds=2010-04-08/hr=11 -POSTHOOK: query: analyze table stats_part partition(ds='2010-04-08', hr='11') compute statistics -POSTHOOK: type: QUERY -POSTHOOK: Input: default@stats_part@ds=2010-04-08/hr=11 -POSTHOOK: Output: default@stats_part -POSTHOOK: Output: default@stats_part@ds=2010-04-08/hr=11 -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: analyze table stats_part partition(ds='2010-04-08', hr='12') compute statistics -PREHOOK: type: QUERY -PREHOOK: Input: default@stats_part@ds=2010-04-08/hr=12 -PREHOOK: Output: default@stats_part -PREHOOK: Output: default@stats_part@ds=2010-04-08/hr=12 -POSTHOOK: query: analyze table stats_part partition(ds='2010-04-08', hr='12') compute statistics -POSTHOOK: type: QUERY -POSTHOOK: Input: default@stats_part@ds=2010-04-08/hr=12 -POSTHOOK: Output: default@stats_part -POSTHOOK: Output: default@stats_part@ds=2010-04-08/hr=12 -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: insert overwrite table stats_part partition (ds='2010-04-08', hr = '13') select key, value from src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@stats_part@ds=2010-04-08/hr=13 -POSTHOOK: query: insert overwrite table stats_part partition (ds='2010-04-08', hr = '13') select key, value from src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@stats_part@ds=2010-04-08/hr=13 -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: desc formatted stats_part -PREHOOK: type: DESCTABLE -POSTHOOK: query: desc formatted stats_part -POSTHOOK: type: DESCTABLE -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default - -# Partition Information -# col_name data_type comment - -ds string None -hr string None - -# Detailed Table Information -Database: default -#### A masked pattern was here #### -Protect Mode: None -Retention: 0 -#### A masked pattern was here #### -Table Type: MANAGED_TABLE -Table Parameters: - numFiles 3 - numPartitions 3 - numRows 1500 - rawDataSize 15936 - totalSize 17436 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -InputFormat: org.apache.hadoop.mapred.TextInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Storage Desc Params: - serialization.format 1 -PREHOOK: query: desc formatted stats_part partition (ds='2010-04-08', hr = '11') -PREHOOK: type: DESCTABLE -POSTHOOK: query: desc formatted stats_part partition (ds='2010-04-08', hr = '11') -POSTHOOK: type: DESCTABLE -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default - -# Partition Information -# col_name data_type comment - -ds string None -hr string None - -# Detailed Partition Information -Partition Value: [2010-04-08, 11] -Database: default -Table: stats_part -#### A masked pattern was here #### -Protect Mode: None -#### A masked pattern was here #### -Partition Parameters: - numFiles 1 - numRows 500 - rawDataSize 5312 - totalSize 5812 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -InputFormat: org.apache.hadoop.mapred.TextInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Storage Desc Params: - serialization.format 1 -PREHOOK: query: desc formatted stats_part partition (ds='2010-04-08', hr = '12') -PREHOOK: type: DESCTABLE -POSTHOOK: query: desc formatted stats_part partition (ds='2010-04-08', hr = '12') -POSTHOOK: type: DESCTABLE -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default - -# Partition Information -# col_name data_type comment - -ds string None -hr string None - -# Detailed Partition Information -Partition Value: [2010-04-08, 12] -Database: default -Table: stats_part -#### A masked pattern was here #### -Protect Mode: None -#### A masked pattern was here #### -Partition Parameters: - numFiles 1 - numRows 500 - rawDataSize 5312 - totalSize 5812 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -InputFormat: org.apache.hadoop.mapred.TextInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Storage Desc Params: - serialization.format 1 -PREHOOK: query: analyze table stats_part partition(ds, hr) compute statistics -PREHOOK: type: QUERY -PREHOOK: Input: default@stats_part@ds=2010-04-08/hr=11 -PREHOOK: Input: default@stats_part@ds=2010-04-08/hr=12 -PREHOOK: Input: default@stats_part@ds=2010-04-08/hr=13 -PREHOOK: Output: default@stats_part -PREHOOK: Output: default@stats_part@ds=2010-04-08/hr=11 -PREHOOK: Output: default@stats_part@ds=2010-04-08/hr=12 -PREHOOK: Output: default@stats_part@ds=2010-04-08/hr=13 -POSTHOOK: query: analyze table stats_part partition(ds, hr) compute statistics -POSTHOOK: type: QUERY -POSTHOOK: Input: default@stats_part@ds=2010-04-08/hr=11 -POSTHOOK: Input: default@stats_part@ds=2010-04-08/hr=12 -POSTHOOK: Input: default@stats_part@ds=2010-04-08/hr=13 -POSTHOOK: Output: default@stats_part -POSTHOOK: Output: default@stats_part@ds=2010-04-08/hr=11 -POSTHOOK: Output: default@stats_part@ds=2010-04-08/hr=12 -POSTHOOK: Output: default@stats_part@ds=2010-04-08/hr=13 -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: desc formatted stats_part -PREHOOK: type: DESCTABLE -POSTHOOK: query: desc formatted stats_part -POSTHOOK: type: DESCTABLE -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default - -# Partition Information -# col_name data_type comment - -ds string None -hr string None - -# Detailed Table Information -Database: default -#### A masked pattern was here #### -Protect Mode: None -Retention: 0 -#### A masked pattern was here #### -Table Type: MANAGED_TABLE -Table Parameters: - numFiles 3 - numPartitions 3 - numRows 1500 - rawDataSize 15936 - totalSize 17436 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -InputFormat: org.apache.hadoop.mapred.TextInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Storage Desc Params: - serialization.format 1 -PREHOOK: query: drop table stats_src -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@stats_src -PREHOOK: Output: default@stats_src -POSTHOOK: query: drop table stats_src -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@stats_src -POSTHOOK: Output: default@stats_src -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: drop table stats_part -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@stats_part -PREHOOK: Output: default@stats_part -POSTHOOK: query: drop table stats_part -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@stats_part -POSTHOOK: Output: default@stats_part -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] Index: hbase-handler/src/test/results/hbase_stats2.q.out =================================================================== --- hbase-handler/src/test/results/hbase_stats2.q.out (revision 1239976) +++ hbase-handler/src/test/results/hbase_stats2.q.out (working copy) @@ -1,390 +0,0 @@ -PREHOOK: query: create table stats_src like src -PREHOOK: type: CREATETABLE -POSTHOOK: query: create table stats_src like src -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@stats_src -PREHOOK: query: insert overwrite table stats_src select * from src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@stats_src -POSTHOOK: query: insert overwrite table stats_src select * from src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@stats_src -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: analyze table stats_src compute statistics -PREHOOK: type: QUERY -PREHOOK: Input: default@stats_src -PREHOOK: Output: default@stats_src -POSTHOOK: query: analyze table stats_src compute statistics -POSTHOOK: type: QUERY -POSTHOOK: Input: default@stats_src -POSTHOOK: Output: default@stats_src -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: desc formatted stats_src -PREHOOK: type: DESCTABLE -POSTHOOK: query: desc formatted stats_src -POSTHOOK: type: DESCTABLE -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default - -# Detailed Table Information -Database: default -#### A masked pattern was here #### -Protect Mode: None -Retention: 0 -#### A masked pattern was here #### -Table Type: MANAGED_TABLE -Table Parameters: - numFiles 1 - numPartitions 0 - numRows 500 - rawDataSize 5312 - totalSize 5812 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -InputFormat: org.apache.hadoop.mapred.TextInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Storage Desc Params: - serialization.format 1 -PREHOOK: query: create table stats_part like srcpart -PREHOOK: type: CREATETABLE -POSTHOOK: query: create table stats_part like srcpart -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@stats_part -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: insert overwrite table stats_part partition (ds='2010-04-08', hr = '11') select key, value from src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@stats_part@ds=2010-04-08/hr=11 -POSTHOOK: query: insert overwrite table stats_part partition (ds='2010-04-08', hr = '11') select key, value from src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@stats_part@ds=2010-04-08/hr=11 -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: insert overwrite table stats_part partition (ds='2010-04-08', hr = '12') select key, value from src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@stats_part@ds=2010-04-08/hr=12 -POSTHOOK: query: insert overwrite table stats_part partition (ds='2010-04-08', hr = '12') select key, value from src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@stats_part@ds=2010-04-08/hr=12 -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: analyze table stats_part partition(ds='2010-04-08', hr='11') compute statistics -PREHOOK: type: QUERY -PREHOOK: Input: default@stats_part@ds=2010-04-08/hr=11 -PREHOOK: Output: default@stats_part -PREHOOK: Output: default@stats_part@ds=2010-04-08/hr=11 -POSTHOOK: query: analyze table stats_part partition(ds='2010-04-08', hr='11') compute statistics -POSTHOOK: type: QUERY -POSTHOOK: Input: default@stats_part@ds=2010-04-08/hr=11 -POSTHOOK: Output: default@stats_part -POSTHOOK: Output: default@stats_part@ds=2010-04-08/hr=11 -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: analyze table stats_part partition(ds='2010-04-08', hr='12') compute statistics -PREHOOK: type: QUERY -PREHOOK: Input: default@stats_part@ds=2010-04-08/hr=12 -PREHOOK: Output: default@stats_part -PREHOOK: Output: default@stats_part@ds=2010-04-08/hr=12 -POSTHOOK: query: analyze table stats_part partition(ds='2010-04-08', hr='12') compute statistics -POSTHOOK: type: QUERY -POSTHOOK: Input: default@stats_part@ds=2010-04-08/hr=12 -POSTHOOK: Output: default@stats_part -POSTHOOK: Output: default@stats_part@ds=2010-04-08/hr=12 -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: insert overwrite table stats_part partition (ds='2010-04-08', hr = '13') select key, value from src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@stats_part@ds=2010-04-08/hr=13 -POSTHOOK: query: insert overwrite table stats_part partition (ds='2010-04-08', hr = '13') select key, value from src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@stats_part@ds=2010-04-08/hr=13 -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: desc formatted stats_part -PREHOOK: type: DESCTABLE -POSTHOOK: query: desc formatted stats_part -POSTHOOK: type: DESCTABLE -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default - -# Partition Information -# col_name data_type comment - -ds string None -hr string None - -# Detailed Table Information -Database: default -#### A masked pattern was here #### -Protect Mode: None -Retention: 0 -#### A masked pattern was here #### -Table Type: MANAGED_TABLE -Table Parameters: - numFiles 3 - numPartitions 3 - numRows 1500 - rawDataSize 15936 - totalSize 17436 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -InputFormat: org.apache.hadoop.mapred.TextInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Storage Desc Params: - serialization.format 1 -PREHOOK: query: desc formatted stats_part partition (ds='2010-04-08', hr = '11') -PREHOOK: type: DESCTABLE -POSTHOOK: query: desc formatted stats_part partition (ds='2010-04-08', hr = '11') -POSTHOOK: type: DESCTABLE -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default - -# Partition Information -# col_name data_type comment - -ds string None -hr string None - -# Detailed Partition Information -Partition Value: [2010-04-08, 11] -Database: default -Table: stats_part -#### A masked pattern was here #### -Protect Mode: None -#### A masked pattern was here #### -Partition Parameters: - numFiles 1 - numRows 500 - rawDataSize 5312 - totalSize 5812 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -InputFormat: org.apache.hadoop.mapred.TextInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Storage Desc Params: - serialization.format 1 -PREHOOK: query: desc formatted stats_part partition (ds='2010-04-08', hr = '12') -PREHOOK: type: DESCTABLE -POSTHOOK: query: desc formatted stats_part partition (ds='2010-04-08', hr = '12') -POSTHOOK: type: DESCTABLE -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default - -# Partition Information -# col_name data_type comment - -ds string None -hr string None - -# Detailed Partition Information -Partition Value: [2010-04-08, 12] -Database: default -Table: stats_part -#### A masked pattern was here #### -Protect Mode: None -#### A masked pattern was here #### -Partition Parameters: - numFiles 1 - numRows 500 - rawDataSize 5312 - totalSize 5812 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -InputFormat: org.apache.hadoop.mapred.TextInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Storage Desc Params: - serialization.format 1 -PREHOOK: query: analyze table stats_part partition(ds, hr) compute statistics -PREHOOK: type: QUERY -PREHOOK: Input: default@stats_part@ds=2010-04-08/hr=11 -PREHOOK: Input: default@stats_part@ds=2010-04-08/hr=12 -PREHOOK: Input: default@stats_part@ds=2010-04-08/hr=13 -PREHOOK: Output: default@stats_part -PREHOOK: Output: default@stats_part@ds=2010-04-08/hr=11 -PREHOOK: Output: default@stats_part@ds=2010-04-08/hr=12 -PREHOOK: Output: default@stats_part@ds=2010-04-08/hr=13 -POSTHOOK: query: analyze table stats_part partition(ds, hr) compute statistics -POSTHOOK: type: QUERY -POSTHOOK: Input: default@stats_part@ds=2010-04-08/hr=11 -POSTHOOK: Input: default@stats_part@ds=2010-04-08/hr=12 -POSTHOOK: Input: default@stats_part@ds=2010-04-08/hr=13 -POSTHOOK: Output: default@stats_part -POSTHOOK: Output: default@stats_part@ds=2010-04-08/hr=11 -POSTHOOK: Output: default@stats_part@ds=2010-04-08/hr=12 -POSTHOOK: Output: default@stats_part@ds=2010-04-08/hr=13 -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: desc formatted stats_part -PREHOOK: type: DESCTABLE -POSTHOOK: query: desc formatted stats_part -POSTHOOK: type: DESCTABLE -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -# col_name data_type comment - -key string default -value string default - -# Partition Information -# col_name data_type comment - -ds string None -hr string None - -# Detailed Table Information -Database: default -#### A masked pattern was here #### -Protect Mode: None -Retention: 0 -#### A masked pattern was here #### -Table Type: MANAGED_TABLE -Table Parameters: - numFiles 3 - numPartitions 3 - numRows 1500 - rawDataSize 15936 - totalSize 17436 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -InputFormat: org.apache.hadoop.mapred.TextInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Storage Desc Params: - serialization.format 1 -PREHOOK: query: drop table stats_src -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@stats_src -PREHOOK: Output: default@stats_src -POSTHOOK: query: drop table stats_src -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@stats_src -POSTHOOK: Output: default@stats_src -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: drop table stats_part -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@stats_part -PREHOOK: Output: default@stats_part -POSTHOOK: query: drop table stats_part -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@stats_part -POSTHOOK: Output: default@stats_part -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=11).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=12).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_part PARTITION(ds=2010-04-08,hr=13).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: stats_src.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] Index: hbase-handler/src/test/results/hbase_joins.q.out =================================================================== --- hbase-handler/src/test/results/hbase_joins.q.out (revision 1239976) +++ hbase-handler/src/test/results/hbase_joins.q.out (working copy) @@ -1,267 +0,0 @@ -PREHOOK: query: DROP TABLE users -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE users -POSTHOOK: type: DROPTABLE -PREHOOK: query: DROP TABLE states -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE states -POSTHOOK: type: DROPTABLE -PREHOOK: query: DROP TABLE countries -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE countries -POSTHOOK: type: DROPTABLE -PREHOOK: query: DROP TABLE users_level -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE users_level -POSTHOOK: type: DROPTABLE -PREHOOK: query: -- From HIVE-1257 - -CREATE TABLE users(key string, state string, country string, country_id int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "info:state,info:country,info:country_id" -) -PREHOOK: type: CREATETABLE -POSTHOOK: query: -- From HIVE-1257 - -CREATE TABLE users(key string, state string, country string, country_id int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "info:state,info:country,info:country_id" -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@users -PREHOOK: query: CREATE TABLE states(key string, name string) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "state:name" -) -PREHOOK: type: CREATETABLE -POSTHOOK: query: CREATE TABLE states(key string, name string) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "state:name" -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@states -PREHOOK: query: CREATE TABLE countries(key string, name string, country string, country_id int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "info:name,info:country,info:country_id" -) -PREHOOK: type: CREATETABLE -POSTHOOK: query: CREATE TABLE countries(key string, name string, country string, country_id int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "info:name,info:country,info:country_id" -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@countries -PREHOOK: query: INSERT OVERWRITE TABLE users SELECT 'user1', 'IA', 'USA', 0 -FROM src WHERE key=100 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@users -POSTHOOK: query: INSERT OVERWRITE TABLE users SELECT 'user1', 'IA', 'USA', 0 -FROM src WHERE key=100 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@users -PREHOOK: query: INSERT OVERWRITE TABLE states SELECT 'IA', 'Iowa' -FROM src WHERE key=100 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@states -POSTHOOK: query: INSERT OVERWRITE TABLE states SELECT 'IA', 'Iowa' -FROM src WHERE key=100 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@states -PREHOOK: query: INSERT OVERWRITE TABLE countries SELECT 'USA', 'United States', 'USA', 1 -FROM src WHERE key=100 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@countries -POSTHOOK: query: INSERT OVERWRITE TABLE countries SELECT 'USA', 'United States', 'USA', 1 -FROM src WHERE key=100 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@countries -PREHOOK: query: SELECT u.key, u.country, c.name, c.key FROM users u JOIN countries c -ON (u.country = c.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@countries -PREHOOK: Input: default@users -#### A masked pattern was here #### -POSTHOOK: query: SELECT u.key, u.country, c.name, c.key FROM users u JOIN countries c -ON (u.country = c.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@countries -POSTHOOK: Input: default@users -#### A masked pattern was here #### -user1 USA United States USA -PREHOOK: query: SELECT u.key, u.country, c.name, c.key FROM users u JOIN countries c -ON (u.country = c.country) -PREHOOK: type: QUERY -PREHOOK: Input: default@countries -PREHOOK: Input: default@users -#### A masked pattern was here #### -POSTHOOK: query: SELECT u.key, u.country, c.name, c.key FROM users u JOIN countries c -ON (u.country = c.country) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@countries -POSTHOOK: Input: default@users -#### A masked pattern was here #### -user1 USA United States USA -PREHOOK: query: SELECT u.key, u.country, c.name, c.key FROM users u JOIN countries c -ON (u.country_id = c.country_id) -PREHOOK: type: QUERY -PREHOOK: Input: default@countries -PREHOOK: Input: default@users -#### A masked pattern was here #### -POSTHOOK: query: SELECT u.key, u.country, c.name, c.key FROM users u JOIN countries c -ON (u.country_id = c.country_id) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@countries -POSTHOOK: Input: default@users -#### A masked pattern was here #### -PREHOOK: query: SELECT u.key, u.state, s.name FROM users u JOIN states s -ON (u.state = s.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@states -PREHOOK: Input: default@users -#### A masked pattern was here #### -POSTHOOK: query: SELECT u.key, u.state, s.name FROM users u JOIN states s -ON (u.state = s.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@states -POSTHOOK: Input: default@users -#### A masked pattern was here #### -user1 IA Iowa -PREHOOK: query: SELECT u.key, u.country, c.name, c.key FROM users u JOIN countries c -ON (u.country = c.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@countries -PREHOOK: Input: default@users -#### A masked pattern was here #### -POSTHOOK: query: SELECT u.key, u.country, c.name, c.key FROM users u JOIN countries c -ON (u.country = c.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@countries -POSTHOOK: Input: default@users -#### A masked pattern was here #### -user1 USA United States USA -PREHOOK: query: SELECT u.key, u.country, c.name, c.key FROM users u JOIN countries c -ON (u.country = c.country) -PREHOOK: type: QUERY -PREHOOK: Input: default@countries -PREHOOK: Input: default@users -#### A masked pattern was here #### -POSTHOOK: query: SELECT u.key, u.country, c.name, c.key FROM users u JOIN countries c -ON (u.country = c.country) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@countries -POSTHOOK: Input: default@users -#### A masked pattern was here #### -user1 USA United States USA -PREHOOK: query: SELECT u.key, u.country, c.name, c.key FROM users u JOIN countries c -ON (u.country_id = c.country_id) -PREHOOK: type: QUERY -PREHOOK: Input: default@countries -PREHOOK: Input: default@users -#### A masked pattern was here #### -POSTHOOK: query: SELECT u.key, u.country, c.name, c.key FROM users u JOIN countries c -ON (u.country_id = c.country_id) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@countries -POSTHOOK: Input: default@users -#### A masked pattern was here #### -PREHOOK: query: SELECT u.key, u.state, s.name FROM users u JOIN states s -ON (u.state = s.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@states -PREHOOK: Input: default@users -#### A masked pattern was here #### -POSTHOOK: query: SELECT u.key, u.state, s.name FROM users u JOIN states s -ON (u.state = s.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@states -POSTHOOK: Input: default@users -#### A masked pattern was here #### -user1 IA Iowa -PREHOOK: query: DROP TABLE users -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@users -PREHOOK: Output: default@users -POSTHOOK: query: DROP TABLE users -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@users -POSTHOOK: Output: default@users -PREHOOK: query: DROP TABLE states -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@states -PREHOOK: Output: default@states -POSTHOOK: query: DROP TABLE states -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@states -POSTHOOK: Output: default@states -PREHOOK: query: DROP TABLE countries -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@countries -PREHOOK: Output: default@countries -POSTHOOK: query: DROP TABLE countries -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@countries -POSTHOOK: Output: default@countries -PREHOOK: query: CREATE TABLE users(key int, userid int, username string, created int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,f:userid,f:nickname,f:created") -PREHOOK: type: CREATETABLE -POSTHOOK: query: CREATE TABLE users(key int, userid int, username string, created int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,f:userid,f:nickname,f:created") -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@users -PREHOOK: query: CREATE TABLE users_level(key int, userid int, level int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,f:userid,f:level") -PREHOOK: type: CREATETABLE -POSTHOOK: query: CREATE TABLE users_level(key int, userid int, level int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,f:userid,f:level") -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@users_level -PREHOOK: query: -- HIVE-1903: the problem fixed here showed up even without any data, --- so no need to load any to test it -SELECT year(from_unixtime(users.created)) AS year, level, count(users.userid) AS num - FROM users JOIN users_level ON (users.userid = users_level.userid) - GROUP BY year(from_unixtime(users.created)), level -PREHOOK: type: QUERY -PREHOOK: Input: default@users -PREHOOK: Input: default@users_level -#### A masked pattern was here #### -POSTHOOK: query: -- HIVE-1903: the problem fixed here showed up even without any data, --- so no need to load any to test it -SELECT year(from_unixtime(users.created)) AS year, level, count(users.userid) AS num - FROM users JOIN users_level ON (users.userid = users_level.userid) - GROUP BY year(from_unixtime(users.created)), level -POSTHOOK: type: QUERY -POSTHOOK: Input: default@users -POSTHOOK: Input: default@users_level -#### A masked pattern was here #### -PREHOOK: query: DROP TABLE users -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@users -PREHOOK: Output: default@users -POSTHOOK: query: DROP TABLE users -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@users -POSTHOOK: Output: default@users -PREHOOK: query: DROP TABLE users_level -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@users_level -PREHOOK: Output: default@users_level -POSTHOOK: query: DROP TABLE users_level -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@users_level -POSTHOOK: Output: default@users_level Index: hbase-handler/src/test/results/hbase_pushdown.q.out =================================================================== --- hbase-handler/src/test/results/hbase_pushdown.q.out (revision 1239976) +++ hbase-handler/src/test/results/hbase_pushdown.q.out (working copy) @@ -1,403 +0,0 @@ -PREHOOK: query: CREATE TABLE hbase_pushdown(key int, value string) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,cf:string") -PREHOOK: type: CREATETABLE -POSTHOOK: query: CREATE TABLE hbase_pushdown(key int, value string) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,cf:string") -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: default@hbase_pushdown -PREHOOK: query: INSERT OVERWRITE TABLE hbase_pushdown -SELECT * -FROM src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@hbase_pushdown -POSTHOOK: query: INSERT OVERWRITE TABLE hbase_pushdown -SELECT * -FROM src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@hbase_pushdown -PREHOOK: query: -- with full pushdown -explain select * from hbase_pushdown where key=90 -PREHOOK: type: QUERY -POSTHOOK: query: -- with full pushdown -explain select * from hbase_pushdown where key=90 -POSTHOOK: type: QUERY -ABSTRACT SYNTAX TREE: - (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME hbase_pushdown))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (= (TOK_TABLE_OR_COL key) 90)))) - -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 is a root stage - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Alias -> Map Operator Tree: - hbase_pushdown - TableScan - alias: hbase_pushdown - filterExpr: - expr: (key = 90) - type: boolean - Filter Operator - predicate: - expr: (key = 90) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - Stage: Stage-0 - Fetch Operator - limit: -1 - - -PREHOOK: query: select * from hbase_pushdown where key=90 -PREHOOK: type: QUERY -PREHOOK: Input: default@hbase_pushdown -#### A masked pattern was here #### -POSTHOOK: query: select * from hbase_pushdown where key=90 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@hbase_pushdown -#### A masked pattern was here #### -90 val_90 -PREHOOK: query: -- with partial pushdown - -explain select * from hbase_pushdown where key=90 and value like '%90%' -PREHOOK: type: QUERY -POSTHOOK: query: -- with partial pushdown - -explain select * from hbase_pushdown where key=90 and value like '%90%' -POSTHOOK: type: QUERY -ABSTRACT SYNTAX TREE: - (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME hbase_pushdown))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (and (= (TOK_TABLE_OR_COL key) 90) (like (TOK_TABLE_OR_COL value) '%90%'))))) - -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 is a root stage - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Alias -> Map Operator Tree: - hbase_pushdown - TableScan - alias: hbase_pushdown - filterExpr: - expr: (key = 90) - type: boolean - Filter Operator - predicate: - expr: (value like '%90%') - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - Stage: Stage-0 - Fetch Operator - limit: -1 - - -PREHOOK: query: select * from hbase_pushdown where key=90 and value like '%90%' -PREHOOK: type: QUERY -PREHOOK: Input: default@hbase_pushdown -#### A masked pattern was here #### -POSTHOOK: query: select * from hbase_pushdown where key=90 and value like '%90%' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@hbase_pushdown -#### A masked pattern was here #### -90 val_90 -PREHOOK: query: -- with two residuals - -explain select * from hbase_pushdown -where key=90 and value like '%90%' and key=cast(value as int) -PREHOOK: type: QUERY -POSTHOOK: query: -- with two residuals - -explain select * from hbase_pushdown -where key=90 and value like '%90%' and key=cast(value as int) -POSTHOOK: type: QUERY -ABSTRACT SYNTAX TREE: - (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME hbase_pushdown))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (and (and (= (TOK_TABLE_OR_COL key) 90) (like (TOK_TABLE_OR_COL value) '%90%')) (= (TOK_TABLE_OR_COL key) (TOK_FUNCTION TOK_INT (TOK_TABLE_OR_COL value))))))) - -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 is a root stage - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Alias -> Map Operator Tree: - hbase_pushdown - TableScan - alias: hbase_pushdown - filterExpr: - expr: (key = 90) - type: boolean - Filter Operator - predicate: - expr: ((value like '%90%') and (key = UDFToInteger(value))) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - Stage: Stage-0 - Fetch Operator - limit: -1 - - -PREHOOK: query: -- with contradictory pushdowns - -explain select * from hbase_pushdown -where key=80 and key=90 and value like '%90%' -PREHOOK: type: QUERY -POSTHOOK: query: -- with contradictory pushdowns - -explain select * from hbase_pushdown -where key=80 and key=90 and value like '%90%' -POSTHOOK: type: QUERY -ABSTRACT SYNTAX TREE: - (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME hbase_pushdown))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (and (and (= (TOK_TABLE_OR_COL key) 80) (= (TOK_TABLE_OR_COL key) 90)) (like (TOK_TABLE_OR_COL value) '%90%'))))) - -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 is a root stage - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Alias -> Map Operator Tree: - hbase_pushdown - TableScan - alias: hbase_pushdown - Filter Operator - predicate: - expr: (((key = 80) and (key = 90)) and (value like '%90%')) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - Stage: Stage-0 - Fetch Operator - limit: -1 - - -PREHOOK: query: select * from hbase_pushdown -where key=80 and key=90 and value like '%90%' -PREHOOK: type: QUERY -PREHOOK: Input: default@hbase_pushdown -#### A masked pattern was here #### -POSTHOOK: query: select * from hbase_pushdown -where key=80 and key=90 and value like '%90%' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@hbase_pushdown -#### A masked pattern was here #### -PREHOOK: query: -- with nothing to push down - -explain select * from hbase_pushdown -PREHOOK: type: QUERY -POSTHOOK: query: -- with nothing to push down - -explain select * from hbase_pushdown -POSTHOOK: type: QUERY -ABSTRACT SYNTAX TREE: - (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME hbase_pushdown))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)))) - -STAGE DEPENDENCIES: - Stage-0 is a root stage - -STAGE PLANS: - Stage: Stage-0 - Fetch Operator - limit: -1 - - -PREHOOK: query: -- with a predicate which is not actually part of the filter, so --- it should be ignored by pushdown - -explain select * from hbase_pushdown -where (case when key=90 then 2 else 4 end) > 3 -PREHOOK: type: QUERY -POSTHOOK: query: -- with a predicate which is not actually part of the filter, so --- it should be ignored by pushdown - -explain select * from hbase_pushdown -where (case when key=90 then 2 else 4 end) > 3 -POSTHOOK: type: QUERY -ABSTRACT SYNTAX TREE: - (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME hbase_pushdown))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (> (TOK_FUNCTION when (= (TOK_TABLE_OR_COL key) 90) 2 4) 3)))) - -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 is a root stage - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Alias -> Map Operator Tree: - hbase_pushdown - TableScan - alias: hbase_pushdown - Filter Operator - predicate: - expr: (CASE WHEN ((key = 90)) THEN (2) ELSE (4) END > 3) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - Stage: Stage-0 - Fetch Operator - limit: -1 - - -PREHOOK: query: -- with a predicate which is under an OR, so it should --- be ignored by pushdown - -explain select * from hbase_pushdown -where key=80 or value like '%90%' -PREHOOK: type: QUERY -POSTHOOK: query: -- with a predicate which is under an OR, so it should --- be ignored by pushdown - -explain select * from hbase_pushdown -where key=80 or value like '%90%' -POSTHOOK: type: QUERY -ABSTRACT SYNTAX TREE: - (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME hbase_pushdown))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (or (= (TOK_TABLE_OR_COL key) 80) (like (TOK_TABLE_OR_COL value) '%90%'))))) - -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 is a root stage - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Alias -> Map Operator Tree: - hbase_pushdown - TableScan - alias: hbase_pushdown - Filter Operator - predicate: - expr: ((key = 80) or (value like '%90%')) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - Stage: Stage-0 - Fetch Operator - limit: -1 - - -PREHOOK: query: -- with pushdown disabled - -explain select * from hbase_pushdown where key=90 -PREHOOK: type: QUERY -POSTHOOK: query: -- with pushdown disabled - -explain select * from hbase_pushdown where key=90 -POSTHOOK: type: QUERY -ABSTRACT SYNTAX TREE: - (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME hbase_pushdown))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR TOK_ALLCOLREF)) (TOK_WHERE (= (TOK_TABLE_OR_COL key) 90)))) - -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 is a root stage - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Alias -> Map Operator Tree: - hbase_pushdown - TableScan - alias: hbase_pushdown - Filter Operator - predicate: - expr: (key = 90) - type: boolean - Select Operator - expressions: - expr: key - type: int - expr: value - type: string - outputColumnNames: _col0, _col1 - File Output Operator - compressed: false - GlobalTableId: 0 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - - Stage: Stage-0 - Fetch Operator - limit: -1 - - Index: hbase-handler/src/test/queries/hbase_stats.q =================================================================== --- hbase-handler/src/test/queries/hbase_stats.q (revision 1239976) +++ hbase-handler/src/test/queries/hbase_stats.q (working copy) @@ -1,30 +0,0 @@ -set datanucleus.cache.collections=false; -set hive.stats.autogather=true; -set hive.ststs.atomic=false; - -set hive.stats.dbclass=hbase; - -create table stats_src like src; -insert overwrite table stats_src select * from src; -analyze table stats_src compute statistics; -desc formatted stats_src; - -create table stats_part like srcpart; - -insert overwrite table stats_part partition (ds='2010-04-08', hr = '11') select key, value from src; -insert overwrite table stats_part partition (ds='2010-04-08', hr = '12') select key, value from src; - -analyze table stats_part partition(ds='2010-04-08', hr='11') compute statistics; -analyze table stats_part partition(ds='2010-04-08', hr='12') compute statistics; - -insert overwrite table stats_part partition (ds='2010-04-08', hr = '13') select key, value from src; - -desc formatted stats_part; -desc formatted stats_part partition (ds='2010-04-08', hr = '11'); -desc formatted stats_part partition (ds='2010-04-08', hr = '12'); - -analyze table stats_part partition(ds, hr) compute statistics; -desc formatted stats_part; - -drop table stats_src; -drop table stats_part; Index: hbase-handler/src/test/queries/hbase_stats2.q =================================================================== --- hbase-handler/src/test/queries/hbase_stats2.q (revision 1239976) +++ hbase-handler/src/test/queries/hbase_stats2.q (working copy) @@ -1,31 +0,0 @@ -set datanucleus.cache.collections=false; -set hive.stats.autogather=true; -set hive.stats.atomic=false; -set hive.stats.collect.uncompressedsize=false; - -set hive.stats.dbclass=hbase; - -create table stats_src like src; -insert overwrite table stats_src select * from src; -analyze table stats_src compute statistics; -desc formatted stats_src; - -create table stats_part like srcpart; - -insert overwrite table stats_part partition (ds='2010-04-08', hr = '11') select key, value from src; -insert overwrite table stats_part partition (ds='2010-04-08', hr = '12') select key, value from src; - -analyze table stats_part partition(ds='2010-04-08', hr='11') compute statistics; -analyze table stats_part partition(ds='2010-04-08', hr='12') compute statistics; - -insert overwrite table stats_part partition (ds='2010-04-08', hr = '13') select key, value from src; - -desc formatted stats_part; -desc formatted stats_part partition (ds='2010-04-08', hr = '11'); -desc formatted stats_part partition (ds='2010-04-08', hr = '12'); - -analyze table stats_part partition(ds, hr) compute statistics; -desc formatted stats_part; - -drop table stats_src; -drop table stats_part; Index: hbase-handler/src/test/queries/negative/cascade_dbdrop.q =================================================================== --- hbase-handler/src/test/queries/negative/cascade_dbdrop.q (revision 0) +++ hbase-handler/src/test/queries/negative/cascade_dbdrop.q (revision 0) @@ -0,0 +1,17 @@ + +CREATE DATABASE hbaseDB; + +CREATE TABLE hbaseDB.hbase_table_0(key int, value string) +STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' +WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,cf:string") +TBLPROPERTIES ("hbase.table.name" = "hbase_table_0"); + +DROP DATABASE IF EXISTS hbaseDB CASCADE; + +!ls ../build/ql/tmp/hbase/hbase_table_0; + + + + + + Index: hbase-handler/src/test/queries/positive/table_create.q =================================================================== --- hbase-handler/src/test/queries/positive/table_create.q (revision 0) +++ hbase-handler/src/test/queries/positive/table_create.q (revision 0) @@ -0,0 +1,8 @@ +DROP TABLE hbase_table_1; + +CREATE TABLE hbase_table_1(key int, value string) +STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' +WITH SERDEPROPERTIES ("hbase.columns.mapping" = "cf:string") +TBLPROPERTIES ("hbase.table.name" = "hbase_table_0"); + +!cd ../build/ql/tmp/hbase/hbase_table_0; Index: hbase-handler/src/test/queries/hbase_joins.q =================================================================== --- hbase-handler/src/test/queries/hbase_joins.q (revision 1239976) +++ hbase-handler/src/test/queries/hbase_joins.q (working copy) @@ -1,82 +0,0 @@ -DROP TABLE users; -DROP TABLE states; -DROP TABLE countries; -DROP TABLE users_level; - --- From HIVE-1257 - -CREATE TABLE users(key string, state string, country string, country_id int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "info:state,info:country,info:country_id" -); - -CREATE TABLE states(key string, name string) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "state:name" -); - -CREATE TABLE countries(key string, name string, country string, country_id int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "info:name,info:country,info:country_id" -); - -INSERT OVERWRITE TABLE users SELECT 'user1', 'IA', 'USA', 0 -FROM src WHERE key=100; - -INSERT OVERWRITE TABLE states SELECT 'IA', 'Iowa' -FROM src WHERE key=100; - -INSERT OVERWRITE TABLE countries SELECT 'USA', 'United States', 'USA', 1 -FROM src WHERE key=100; - -set hive.input.format = org.apache.hadoop.hive.ql.io.HiveInputFormat; - -SELECT u.key, u.country, c.name, c.key FROM users u JOIN countries c -ON (u.country = c.key); - -SELECT u.key, u.country, c.name, c.key FROM users u JOIN countries c -ON (u.country = c.country); - -SELECT u.key, u.country, c.name, c.key FROM users u JOIN countries c -ON (u.country_id = c.country_id); - -SELECT u.key, u.state, s.name FROM users u JOIN states s -ON (u.state = s.key); - -set hive.input.format = org.apache.hadoop.hive.ql.io.CombineHiveInputFormat; - -SELECT u.key, u.country, c.name, c.key FROM users u JOIN countries c -ON (u.country = c.key); - -SELECT u.key, u.country, c.name, c.key FROM users u JOIN countries c -ON (u.country = c.country); - -SELECT u.key, u.country, c.name, c.key FROM users u JOIN countries c -ON (u.country_id = c.country_id); - -SELECT u.key, u.state, s.name FROM users u JOIN states s -ON (u.state = s.key); - -DROP TABLE users; -DROP TABLE states; -DROP TABLE countries; - -CREATE TABLE users(key int, userid int, username string, created int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,f:userid,f:nickname,f:created"); - -CREATE TABLE users_level(key int, userid int, level int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,f:userid,f:level"); - --- HIVE-1903: the problem fixed here showed up even without any data, --- so no need to load any to test it -SELECT year(from_unixtime(users.created)) AS year, level, count(users.userid) AS num - FROM users JOIN users_level ON (users.userid = users_level.userid) - GROUP BY year(from_unixtime(users.created)), level; - -DROP TABLE users; -DROP TABLE users_level; Index: hbase-handler/src/test/queries/hbase_pushdown.q =================================================================== --- hbase-handler/src/test/queries/hbase_pushdown.q (revision 1239976) +++ hbase-handler/src/test/queries/hbase_pushdown.q (working copy) @@ -1,53 +0,0 @@ -CREATE TABLE hbase_pushdown(key int, value string) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,cf:string"); - -INSERT OVERWRITE TABLE hbase_pushdown -SELECT * -FROM src; - --- with full pushdown -explain select * from hbase_pushdown where key=90; - -select * from hbase_pushdown where key=90; - --- with partial pushdown - -explain select * from hbase_pushdown where key=90 and value like '%90%'; - -select * from hbase_pushdown where key=90 and value like '%90%'; - --- with two residuals - -explain select * from hbase_pushdown -where key=90 and value like '%90%' and key=cast(value as int); - --- with contradictory pushdowns - -explain select * from hbase_pushdown -where key=80 and key=90 and value like '%90%'; - -select * from hbase_pushdown -where key=80 and key=90 and value like '%90%'; - --- with nothing to push down - -explain select * from hbase_pushdown; - --- with a predicate which is not actually part of the filter, so --- it should be ignored by pushdown - -explain select * from hbase_pushdown -where (case when key=90 then 2 else 4 end) > 3; - --- with a predicate which is under an OR, so it should --- be ignored by pushdown - -explain select * from hbase_pushdown -where key=80 or value like '%90%'; - -set hive.optimize.ppd.storage=false; - --- with pushdown disabled - -explain select * from hbase_pushdown where key=90; Index: hbase-handler/src/test/queries/hbase_bulk.m =================================================================== --- hbase-handler/src/test/queries/hbase_bulk.m (revision 1239976) +++ hbase-handler/src/test/queries/hbase_bulk.m (working copy) @@ -1,59 +0,0 @@ -drop table hbsort; -drop table hbpartition; - -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; - --- this is a dummy table used for controlling how the HFiles are --- created -create table hbsort(key string, val string, val2 string) -stored as -INPUTFORMAT 'org.apache.hadoop.mapred.TextInputFormat' -OUTPUTFORMAT 'org.apache.hadoop.hive.hbase.HiveHFileOutputFormat' -TBLPROPERTIES ('hfile.family.path' = '/tmp/hbsort/cf'); - --- this is a dummy table used for controlling how the input file --- for TotalOrderPartitioner is created -create table hbpartition(part_break string) -row format serde -'org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe' -stored as -inputformat -'org.apache.hadoop.mapred.TextInputFormat' -outputformat -'org.apache.hadoop.hive.ql.io.HiveNullValueSequenceFileOutputFormat'; - --- this should produce one file, but we do not --- know what it will be called, so we will copy it to a well known --- filename /tmp/hbpartition.lst -insert overwrite table hbpartition -select distinct value -from src -where value='val_100' or value='val_200'; - -dfs -count /build/ql/test/data/warehouse/hbpartition; -dfs -cp /build/ql/test/data/warehouse/hbpartition/* /tmp/hbpartition.lst; - -set mapred.reduce.tasks=3; -set hive.mapred.partitioner=org.apache.hadoop.mapred.lib.TotalOrderPartitioner; -set total.order.partitioner.natural.order=false; -set total.order.partitioner.path=/tmp/hbpartition.lst; - --- this should produce three files in /tmp/hbsort/cf --- include some trailing blanks and nulls to make sure we handle them correctly -insert overwrite table hbsort -select distinct value, - case when key=103 then cast(null as string) else key end, - case when key=103 then '' - else cast(key+1 as string) end -from src -cluster by value; - -dfs -count /tmp/hbsort/cf; - --- To get the files out to your local filesystem for loading into --- HBase, run mkdir -p /tmp/blah/cf, then uncomment and --- semicolon-terminate the line below before running this test: --- dfs -copyToLocal /tmp/hbsort/cf/* /tmp/blah/cf - -drop table hbsort; -drop table hbpartition; Index: hbase-handler/src/test/queries/hbase_queries.q =================================================================== --- hbase-handler/src/test/queries/hbase_queries.q (revision 1239976) +++ hbase-handler/src/test/queries/hbase_queries.q (working copy) @@ -1,160 +0,0 @@ -DROP TABLE hbase_table_1; -CREATE TABLE hbase_table_1(key int, value string) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = "cf:string") -TBLPROPERTIES ("hbase.table.name" = "hbase_table_0"); - -DESCRIBE EXTENDED hbase_table_1; - -select * from hbase_table_1; - -EXPLAIN FROM src INSERT OVERWRITE TABLE hbase_table_1 SELECT * WHERE (key%2)=0; -FROM src INSERT OVERWRITE TABLE hbase_table_1 SELECT * WHERE (key%2)=0; - -DROP TABLE hbase_table_2; -CREATE EXTERNAL TABLE hbase_table_2(key int, value string) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = "cf:string") -TBLPROPERTIES ("hbase.table.name" = "hbase_table_0"); - -EXPLAIN -SELECT Y.* -FROM -(SELECT hbase_table_1.* FROM hbase_table_1) x -JOIN -(SELECT src.* FROM src) Y -ON (x.key = Y.key) -ORDER BY key, value LIMIT 20; - -SELECT Y.* -FROM -(SELECT hbase_table_1.* FROM hbase_table_1) x -JOIN -(SELECT src.* FROM src) Y -ON (x.key = Y.key) -ORDER BY key, value LIMIT 20; - -EXPLAIN -SELECT Y.* -FROM -(SELECT hbase_table_1.* FROM hbase_table_1 WHERE hbase_table_1.key > 100) x -JOIN -(SELECT hbase_table_2.* FROM hbase_table_2 WHERE hbase_table_2.key < 120) Y -ON (x.key = Y.key) -ORDER BY key, value; - -SELECT Y.* -FROM -(SELECT hbase_table_1.* FROM hbase_table_1 WHERE hbase_table_1.key > 100) x -JOIN -(SELECT hbase_table_2.* FROM hbase_table_2 WHERE hbase_table_2.key < 120) Y -ON (x.key = Y.key) -ORDER BY key,value; - -DROP TABLE empty_hbase_table; -CREATE TABLE empty_hbase_table(key int, value string) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = "cf:string"); - -DROP TABLE empty_normal_table; -CREATE TABLE empty_normal_table(key int, value string); - -select * from (select count(1) as c from empty_normal_table union all select count(1) as c from empty_hbase_table) x order by c; -select * from (select count(1) c from empty_normal_table union all select count(1) as c from hbase_table_1) x order by c; -select * from (select count(1) c from src union all select count(1) as c from empty_hbase_table) x order by c; -select * from (select count(1) c from src union all select count(1) as c from hbase_table_1) x order by c; - -CREATE TABLE hbase_table_3(key int, value string, count int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "cf:val,cf2:count" -); - -EXPLAIN -INSERT OVERWRITE TABLE hbase_table_3 -SELECT x.key, x.value, Y.count -FROM -(SELECT hbase_table_1.* FROM hbase_table_1) x -JOIN -(SELECT src.key, count(src.key) as count FROM src GROUP BY src.key) Y -ON (x.key = Y.key); - -INSERT OVERWRITE TABLE hbase_table_3 -SELECT x.key, x.value, Y.count -FROM -(SELECT hbase_table_1.* FROM hbase_table_1) x -JOIN -(SELECT src.key, count(src.key) as count FROM src GROUP BY src.key) Y -ON (x.key = Y.key); - -select count(1) from hbase_table_3; -select * from hbase_table_3 order by key, value limit 5; -select key, count from hbase_table_3 order by key, count desc limit 5; - -DROP TABLE hbase_table_4; -CREATE TABLE hbase_table_4(key int, value1 string, value2 int, value3 int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "a:b,a:c,d:e" -); - -INSERT OVERWRITE TABLE hbase_table_4 SELECT key, value, key+1, key+2 -FROM src WHERE key=98 OR key=100; - -SELECT * FROM hbase_table_4 ORDER BY key; - -DROP TABLE hbase_table_5; -CREATE EXTERNAL TABLE hbase_table_5(key int, value map) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ("hbase.columns.mapping" = "a:") -TBLPROPERTIES ("hbase.table.name" = "hbase_table_4"); - -SELECT * FROM hbase_table_5 ORDER BY key; - -DROP TABLE hbase_table_6; -CREATE TABLE hbase_table_6(key int, value map) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = ":key,cf:" -); -INSERT OVERWRITE TABLE hbase_table_6 SELECT key, map(value, key) FROM src -WHERE key=98 OR key=100; - -SELECT * FROM hbase_table_6 ORDER BY key; - -DROP TABLE hbase_table_7; -CREATE TABLE hbase_table_7(value map, key int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "cf:,:key" -); -INSERT OVERWRITE TABLE hbase_table_7 -SELECT map(value, key, upper(value), key+1), key FROM src -WHERE key=98 OR key=100; - -SELECT * FROM hbase_table_7 ORDER BY key; - -set hive.hbase.wal.enabled=false; - -DROP TABLE hbase_table_8; -CREATE TABLE hbase_table_8(key int, value1 string, value2 int, value3 int) -STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' -WITH SERDEPROPERTIES ( -"hbase.columns.mapping" = "a:b,a:c,d:e" -); - -INSERT OVERWRITE TABLE hbase_table_8 SELECT key, value, key+1, key+2 -FROM src WHERE key=98 OR key=100; - -SELECT * FROM hbase_table_8 ORDER BY key; - -DROP TABLE hbase_table_1; -DROP TABLE hbase_table_2; -DROP TABLE hbase_table_3; -DROP TABLE hbase_table_4; -DROP TABLE hbase_table_5; -DROP TABLE hbase_table_6; -DROP TABLE hbase_table_7; -DROP TABLE hbase_table_8; -DROP TABLE empty_hbase_table; -DROP TABLE empty_normal_table; Index: hbase-handler/src/test/templates/TestHBaseNegativeCliDriver.vm =================================================================== --- hbase-handler/src/test/templates/TestHBaseNegativeCliDriver.vm (revision 0) +++ hbase-handler/src/test/templates/TestHBaseNegativeCliDriver.vm (revision 0) @@ -0,0 +1,119 @@ +package org.apache.hadoop.hive.cli; + +import junit.framework.Test; +import junit.framework.TestCase; +import junit.framework.TestSuite; + +import java.io.*; +import java.util.*; + +import org.apache.hadoop.hive.hbase.HBaseQTestUtil; +import org.apache.hadoop.hive.hbase.HBaseTestSetup; +import org.apache.hadoop.hive.ql.exec.Task; +import org.apache.hadoop.hive.ql.history.HiveHistoryViewer; +import org.apache.hadoop.hive.ql.history.HiveHistory.QueryInfo; +import org.apache.hadoop.hive.ql.history.HiveHistory.Keys; +import org.apache.hadoop.hive.ql.history.HiveHistory.TaskInfo; +import org.apache.hadoop.hive.ql.session.SessionState; + + + +import org.antlr.runtime.*; +import org.antlr.runtime.tree.*; + +public class $className extends TestCase { + + private HBaseQTestUtil qt; + private HBaseTestSetup setup; + + public $className(String name, HBaseTestSetup setup) { + super(name); + qt = null; + this.setup = setup; + } + + @Override + protected void setUp() { + try { + boolean miniMR = false; + if ("$clusterMode".equals("miniMR")) { + miniMR = true; + } + + qt = new HBaseQTestUtil( + "$resultsDir.getCanonicalPath()", + "$logDir.getCanonicalPath()", miniMR, setup); + +#foreach ($qf in $qfiles) + qt.addFile("$qf.getCanonicalPath()"); +#end + } catch (Exception e) { + System.out.println("Exception: " + e.getMessage()); + e.printStackTrace(); + System.out.flush(); + fail("Unexpected exception in setup"); + } + } + + @Override + protected void tearDown() { + try { + qt.shutdown(); + } + catch (Exception e) { + System.out.println("Exception: " + e.getMessage()); + e.printStackTrace(); + System.out.flush(); + fail("Unexpected exception in tearDown"); + } + } + + public static Test suite() { + TestSuite suite = new TestSuite(); + HBaseTestSetup setup = new HBaseTestSetup(suite); +#foreach ($qf in $qfiles) + #set ($fname = $qf.getName()) + #set ($eidx = $fname.indexOf('.')) + #set ($tname = $fname.substring(0, $eidx)) + suite.addTest(new $className("testCliDriver_$tname", setup)); +#end + return setup; + } + +#foreach ($qf in $qfiles) + #set ($fname = $qf.getName()) + #set ($eidx = $fname.indexOf('.')) + #set ($tname = $fname.substring(0, $eidx)) + public void testCliDriver_$tname() throws Exception { + long startTime = System.currentTimeMillis(); + try { + System.out.println("Begin query: " + "$fname"); + qt.cliInit("$fname"); + qt.clearTestSideEffects(); + int ecode = qt.executeClient("$fname"); + if (ecode == 0) { + fail("Client Execution failed with error code = " + ecode); + } + + ecode = qt.checkCliDriverResults("$fname"); + if (ecode != 0) { + fail("Client execution results failed with error code = " + ecode); + } + qt.clearPostTestEffects(); + + } catch (Throwable e) { + System.out.println("Exception: " + e.getMessage()); + e.printStackTrace(); + System.out.println("Failed query: " + "$fname"); + System.out.flush(); + fail("Unexpected exception"); + } + + long elapsedTime = System.currentTimeMillis() - startTime; + System.out.println("Done query: " + "$fname" + " elapsedTime=" + elapsedTime/1000 + "s"); + assertTrue("Test passed", true); + } + +#end +} + Index: hbase-handler/build.xml =================================================================== --- hbase-handler/build.xml (revision 1239976) +++ hbase-handler/build.xml (working copy) @@ -74,22 +74,31 @@ + Index: metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java =================================================================== --- metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java (revision 1239976) +++ metastore/src/java/org/apache/hadoop/hive/metastore/HiveMetaStoreClient.java (working copy) @@ -452,6 +452,13 @@ } return; } + + if (cascade) { + List tableList = getAllTables(name); + for (String table : tableList) { + dropTable(name, table, deleteData, false); + } + } client.drop_database(name, deleteData, cascade); }