diff --git ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java index 65a795c..928fd61 100644 --- ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java +++ ql/src/java/org/apache/hadoop/hive/ql/io/AcidUtils.java @@ -40,6 +40,8 @@ * are used by the compactor and cleaner and thus must be format agnostic. */ public class AcidUtils { + // This key will be put in the conf file when planning an acid operation + public static final String CONF_ACID_KEY = "hive.doing.acid"; public static final String BASE_PREFIX = "base_"; public static final String DELTA_PREFIX = "delta_"; public static final PathFilter deltaFileFilter = new PathFilter() { diff --git ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcInputFormat.java ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcInputFormat.java index 0310fdf..9007771 100644 --- ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcInputFormat.java +++ ql/src/java/org/apache/hadoop/hive/ql/io/orc/OrcInputFormat.java @@ -132,7 +132,7 @@ @Override public boolean shouldSkipCombine(Path path, Configuration conf) throws IOException { - return AcidUtils.isAcid(path, conf); + return (conf.get(AcidUtils.CONF_ACID_KEY) != null) || AcidUtils.isAcid(path, conf); } private static class OrcRecordReader diff --git ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java index 41c75ef..1fe187a 100644 --- ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java +++ ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java @@ -6256,6 +6256,7 @@ private void checkAcidConstraints(QB qb, TableDesc tableDesc) throws SemanticExc LOG.info("Modifying config values for ACID write"); conf.setBoolVar(ConfVars.HIVEOPTREDUCEDEDUPLICATION, false); conf.setBoolVar(ConfVars.HIVE_HADOOP_SUPPORTS_SUBDIRECTORIES, true); + conf.set(AcidUtils.CONF_ACID_KEY, "true"); } /** diff --git ql/src/test/queries/clientnegative/acid_overwrite.q ql/src/test/queries/clientnegative/acid_overwrite.q index 9f6c1f8..5515fc4 100644 --- ql/src/test/queries/clientnegative/acid_overwrite.q +++ ql/src/test/queries/clientnegative/acid_overwrite.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_uanp(a int, b varchar(128)) clustered by (a) into 2 buckets stored as orc; diff --git ql/src/test/queries/clientnegative/authorization_delete_nodeletepriv.q ql/src/test/queries/clientnegative/authorization_delete_nodeletepriv.q index 090495a..31a3297 100644 --- ql/src/test/queries/clientnegative/authorization_delete_nodeletepriv.q +++ ql/src/test/queries/clientnegative/authorization_delete_nodeletepriv.q @@ -5,7 +5,6 @@ set hive.security.authorization.enabled=true; set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; diff --git ql/src/test/queries/clientnegative/authorization_update_noupdatepriv.q ql/src/test/queries/clientnegative/authorization_update_noupdatepriv.q index 922beba..5310dbe 100644 --- ql/src/test/queries/clientnegative/authorization_update_noupdatepriv.q +++ ql/src/test/queries/clientnegative/authorization_update_noupdatepriv.q @@ -5,7 +5,6 @@ set hive.security.authorization.enabled=true; set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; diff --git ql/src/test/queries/clientnegative/update_no_such_table.q ql/src/test/queries/clientnegative/update_no_such_table.q index 522c46d..07239cf 100644 --- ql/src/test/queries/clientnegative/update_no_such_table.q +++ ql/src/test/queries/clientnegative/update_no_such_table.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; diff --git ql/src/test/queries/clientnegative/update_partition_col.q ql/src/test/queries/clientnegative/update_partition_col.q index 918d312..929cc59 100644 --- ql/src/test/queries/clientnegative/update_partition_col.q +++ ql/src/test/queries/clientnegative/update_partition_col.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table foo(a int, b varchar(128)) partitioned by (ds string) clustered by (a) into 2 buckets stored as orc; diff --git ql/src/test/queries/clientpositive/acid_vectorization.q ql/src/test/queries/clientpositive/acid_vectorization.q index 804144a..a1bf971 100644 --- ql/src/test/queries/clientpositive/acid_vectorization.q +++ ql/src/test/queries/clientpositive/acid_vectorization.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; set hive.exec.dynamic.partition.mode=nonstrict; set hive.vectorized.execution.enabled=true; diff --git ql/src/test/queries/clientpositive/authorization_delete.q ql/src/test/queries/clientpositive/authorization_delete.q index ebd0315..973fb8e 100644 --- ql/src/test/queries/clientpositive/authorization_delete.q +++ ql/src/test/queries/clientpositive/authorization_delete.q @@ -4,7 +4,6 @@ set hive.security.authenticator.manager=org.apache.hadoop.hive.ql.security.Sessi set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; set user.name=user1; diff --git ql/src/test/queries/clientpositive/authorization_delete_own_table.q ql/src/test/queries/clientpositive/authorization_delete_own_table.q index 19dbbeb..8b4c89e 100644 --- ql/src/test/queries/clientpositive/authorization_delete_own_table.q +++ ql/src/test/queries/clientpositive/authorization_delete_own_table.q @@ -5,7 +5,6 @@ set hive.security.authorization.enabled=true; set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; diff --git ql/src/test/queries/clientpositive/authorization_update.q ql/src/test/queries/clientpositive/authorization_update.q index 18ceadb..fd9f670 100644 --- ql/src/test/queries/clientpositive/authorization_update.q +++ ql/src/test/queries/clientpositive/authorization_update.q @@ -4,7 +4,6 @@ set hive.security.authenticator.manager=org.apache.hadoop.hive.ql.security.Sessi set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; set user.name=user1; diff --git ql/src/test/queries/clientpositive/authorization_update_own_table.q ql/src/test/queries/clientpositive/authorization_update_own_table.q index 46beb49..489117b 100644 --- ql/src/test/queries/clientpositive/authorization_update_own_table.q +++ ql/src/test/queries/clientpositive/authorization_update_own_table.q @@ -5,7 +5,6 @@ set hive.security.authorization.enabled=true; set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; diff --git ql/src/test/queries/clientpositive/delete_all_non_partitioned.q ql/src/test/queries/clientpositive/delete_all_non_partitioned.q index 80a5991..f15391b 100644 --- ql/src/test/queries/clientpositive/delete_all_non_partitioned.q +++ ql/src/test/queries/clientpositive/delete_all_non_partitioned.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_danp(a int, b varchar(128)) clustered by (a) into 2 buckets stored as orc; diff --git ql/src/test/queries/clientpositive/delete_all_partitioned.q ql/src/test/queries/clientpositive/delete_all_partitioned.q index b848319..9ae6a54 100644 --- ql/src/test/queries/clientpositive/delete_all_partitioned.q +++ ql/src/test/queries/clientpositive/delete_all_partitioned.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_dap(a int, b varchar(128)) partitioned by (ds string) clustered by (a) into 2 buckets stored as orc; diff --git ql/src/test/queries/clientpositive/delete_orig_table.q ql/src/test/queries/clientpositive/delete_orig_table.q index e1759f6..d859eae 100644 --- ql/src/test/queries/clientpositive/delete_orig_table.q +++ ql/src/test/queries/clientpositive/delete_orig_table.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; dfs ${system:test.dfs.mkdir} ${system:test.tmp.dir}/delete_orig_table; diff --git ql/src/test/queries/clientpositive/delete_tmp_table.q ql/src/test/queries/clientpositive/delete_tmp_table.q index 5563b3c..6f5a73d 100644 --- ql/src/test/queries/clientpositive/delete_tmp_table.q +++ ql/src/test/queries/clientpositive/delete_tmp_table.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create temporary table acid_dtt(a int, b varchar(128)) clustered by (a) into 2 buckets stored as orc; diff --git ql/src/test/queries/clientpositive/delete_where_no_match.q ql/src/test/queries/clientpositive/delete_where_no_match.q index 8ebff45..16a3839 100644 --- ql/src/test/queries/clientpositive/delete_where_no_match.q +++ ql/src/test/queries/clientpositive/delete_where_no_match.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_dwnm(a int, b varchar(128)) clustered by (a) into 2 buckets stored as orc; diff --git ql/src/test/queries/clientpositive/delete_where_non_partitioned.q ql/src/test/queries/clientpositive/delete_where_non_partitioned.q index b37ec80..b703768 100644 --- ql/src/test/queries/clientpositive/delete_where_non_partitioned.q +++ ql/src/test/queries/clientpositive/delete_where_non_partitioned.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_dwnp(a int, b varchar(128)) clustered by (a) into 2 buckets stored as orc; diff --git ql/src/test/queries/clientpositive/delete_where_partitioned.q ql/src/test/queries/clientpositive/delete_where_partitioned.q index cce89f4..5959c3c 100644 --- ql/src/test/queries/clientpositive/delete_where_partitioned.q +++ ql/src/test/queries/clientpositive/delete_where_partitioned.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_dwp(a int, b varchar(128)) partitioned by (ds string) clustered by (a) into 2 buckets stored as orc; diff --git ql/src/test/queries/clientpositive/delete_whole_partition.q ql/src/test/queries/clientpositive/delete_whole_partition.q index 2cb3e74..1444090 100644 --- ql/src/test/queries/clientpositive/delete_whole_partition.q +++ ql/src/test/queries/clientpositive/delete_whole_partition.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_dwhp(a int, b varchar(128)) partitioned by (ds string) clustered by (a) into 2 buckets stored as orc; diff --git ql/src/test/queries/clientpositive/insert_orig_table.q ql/src/test/queries/clientpositive/insert_orig_table.q index 2c6df88..e1114d5 100644 --- ql/src/test/queries/clientpositive/insert_orig_table.q +++ ql/src/test/queries/clientpositive/insert_orig_table.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_iot( diff --git ql/src/test/queries/clientpositive/insert_update_delete.q ql/src/test/queries/clientpositive/insert_update_delete.q index 34350df..a3ef181 100644 --- ql/src/test/queries/clientpositive/insert_update_delete.q +++ ql/src/test/queries/clientpositive/insert_update_delete.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_iud(a int, b varchar(128)) clustered by (a) into 2 buckets stored as orc; diff --git ql/src/test/queries/clientpositive/insert_values_dynamic_partitioned.q ql/src/test/queries/clientpositive/insert_values_dynamic_partitioned.q index bde2e71..bea67a4 100644 --- ql/src/test/queries/clientpositive/insert_values_dynamic_partitioned.q +++ ql/src/test/queries/clientpositive/insert_values_dynamic_partitioned.q @@ -1,7 +1,6 @@ set hive.exec.dynamic.partition.mode=nonstrict; set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table ivdp(i int, diff --git ql/src/test/queries/clientpositive/insert_values_non_partitioned.q ql/src/test/queries/clientpositive/insert_values_non_partitioned.q index 9d57f23..f69100c 100644 --- ql/src/test/queries/clientpositive/insert_values_non_partitioned.q +++ ql/src/test/queries/clientpositive/insert_values_non_partitioned.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_ivnp(ti tinyint, diff --git ql/src/test/queries/clientpositive/insert_values_orig_table.q ql/src/test/queries/clientpositive/insert_values_orig_table.q index d3d68d2..703ed81 100644 --- ql/src/test/queries/clientpositive/insert_values_orig_table.q +++ ql/src/test/queries/clientpositive/insert_values_orig_table.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_ivot( diff --git ql/src/test/queries/clientpositive/insert_values_partitioned.q ql/src/test/queries/clientpositive/insert_values_partitioned.q index 23d6d4c..70bd03b 100644 --- ql/src/test/queries/clientpositive/insert_values_partitioned.q +++ ql/src/test/queries/clientpositive/insert_values_partitioned.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_ivp(ti tinyint, diff --git ql/src/test/queries/clientpositive/insert_values_tmp_table.q ql/src/test/queries/clientpositive/insert_values_tmp_table.q index fd8ec29..0b440d8 100644 --- ql/src/test/queries/clientpositive/insert_values_tmp_table.q +++ ql/src/test/queries/clientpositive/insert_values_tmp_table.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create temporary table acid_ivtt(i int, de decimal(5,2), vc varchar(128)) clustered by (vc) into 2 buckets stored as orc; diff --git ql/src/test/queries/clientpositive/update_after_multiple_inserts.q ql/src/test/queries/clientpositive/update_after_multiple_inserts.q index 04d2df5..b440cf3 100644 --- ql/src/test/queries/clientpositive/update_after_multiple_inserts.q +++ ql/src/test/queries/clientpositive/update_after_multiple_inserts.q @@ -1,7 +1,6 @@ set hive.exec.dynamic.partition.mode=nonstrict; set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_uami(i int, diff --git ql/src/test/queries/clientpositive/update_all_non_partitioned.q ql/src/test/queries/clientpositive/update_all_non_partitioned.q index 67d6ba9..7c427df 100644 --- ql/src/test/queries/clientpositive/update_all_non_partitioned.q +++ ql/src/test/queries/clientpositive/update_all_non_partitioned.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_uanp(a int, b varchar(128)) clustered by (a) into 2 buckets stored as orc; diff --git ql/src/test/queries/clientpositive/update_all_partitioned.q ql/src/test/queries/clientpositive/update_all_partitioned.q index 0b6c767..80d40c1 100644 --- ql/src/test/queries/clientpositive/update_all_partitioned.q +++ ql/src/test/queries/clientpositive/update_all_partitioned.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_uap(a int, b varchar(128)) partitioned by (ds string) clustered by (a) into 2 buckets stored as orc; diff --git ql/src/test/queries/clientpositive/update_all_types.q ql/src/test/queries/clientpositive/update_all_types.q index 39fe73d..c60201d 100644 --- ql/src/test/queries/clientpositive/update_all_types.q +++ ql/src/test/queries/clientpositive/update_all_types.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_uat(ti tinyint, diff --git ql/src/test/queries/clientpositive/update_orig_table.q ql/src/test/queries/clientpositive/update_orig_table.q index f09ad32..61622d9 100644 --- ql/src/test/queries/clientpositive/update_orig_table.q +++ ql/src/test/queries/clientpositive/update_orig_table.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; dfs ${system:test.dfs.mkdir} ${system:test.tmp.dir}/update_orig_table; diff --git ql/src/test/queries/clientpositive/update_tmp_table.q ql/src/test/queries/clientpositive/update_tmp_table.q index c863cd6..5fbad2c 100644 --- ql/src/test/queries/clientpositive/update_tmp_table.q +++ ql/src/test/queries/clientpositive/update_tmp_table.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_utt(a int, b varchar(128)) clustered by (b) into 2 buckets stored as orc; diff --git ql/src/test/queries/clientpositive/update_two_cols.q ql/src/test/queries/clientpositive/update_two_cols.q index 3233d2f..716df7d 100644 --- ql/src/test/queries/clientpositive/update_two_cols.q +++ ql/src/test/queries/clientpositive/update_two_cols.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_utc(a int, b varchar(128), c float) clustered by (a) into 2 buckets stored as orc; diff --git ql/src/test/queries/clientpositive/update_where_no_match.q ql/src/test/queries/clientpositive/update_where_no_match.q index 00583c3..d3252de 100644 --- ql/src/test/queries/clientpositive/update_where_no_match.q +++ ql/src/test/queries/clientpositive/update_where_no_match.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_wnm(a int, b varchar(128)) clustered by (a) into 2 buckets stored as orc; diff --git ql/src/test/queries/clientpositive/update_where_non_partitioned.q ql/src/test/queries/clientpositive/update_where_non_partitioned.q index 378cf94..e52d9e8 100644 --- ql/src/test/queries/clientpositive/update_where_non_partitioned.q +++ ql/src/test/queries/clientpositive/update_where_non_partitioned.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_uwnp(a int, b varchar(128)) clustered by (a) into 2 buckets stored as orc; diff --git ql/src/test/queries/clientpositive/update_where_partitioned.q ql/src/test/queries/clientpositive/update_where_partitioned.q index c5b6d04..d84da5b 100644 --- ql/src/test/queries/clientpositive/update_where_partitioned.q +++ ql/src/test/queries/clientpositive/update_where_partitioned.q @@ -1,6 +1,5 @@ set hive.support.concurrency=true; set hive.txn.manager=org.apache.hadoop.hive.ql.lockmgr.DbTxnManager; -set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set hive.enforce.bucketing=true; create table acid_uwp(a int, b varchar(128)) partitioned by (ds string) clustered by (a) into 2 buckets stored as orc;