commit 426b119e3f8f2ab369e8ca5f242c2d71c833c852 Author: Sahil Takiar Date: Wed Jan 31 13:45:00 2018 -0800 HIVE-18596: Synchronize value of hive.spark.client.connect.timeout across unit tests diff --git a/data/conf/spark/standalone/hive-site.xml b/data/conf/spark/standalone/hive-site.xml index 1e5bd65758..7095979f10 100644 --- a/data/conf/spark/standalone/hive-site.xml +++ b/data/conf/spark/standalone/hive-site.xml @@ -255,6 +255,11 @@ Internal marker for test. Used for masking env-dependent values + + hive.spark.client.connect.timeout + 30000ms + + hive.stats.column.autogather diff --git a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java index 0fee3c6d8b..2a4da20df9 100644 --- a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java +++ b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestJdbcWithLocalClusterSpark.java @@ -70,6 +70,7 @@ private static HiveConf createHiveConf() { conf.set("hive.execution.engine", "spark"); conf.set("spark.serializer", "org.apache.spark.serializer.KryoSerializer"); conf.set("spark.master", "local-cluster[2,2,1024]"); + conf.set("hive.spark.client.connect.timeout", "30000ms"); // FIXME: Hadoop3 made the incompatible change for dfs.client.datanode-restart.timeout // while spark2 is still using Hadoop2. // Spark requires Hive to support Hadoop3 first then Spark can start diff --git a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java index 93fe7762c2..9f72e51187 100644 --- a/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java +++ b/itests/hive-unit/src/test/java/org/apache/hive/jdbc/TestMultiSessionsHS2WithLocalClusterSpark.java @@ -79,6 +79,7 @@ private static HiveConf createHiveConf() { conf.set("spark.serializer", "org.apache.spark.serializer.KryoSerializer"); conf.set("spark.master", "local-cluster[2,2,1024]"); conf.set("spark.deploy.defaultCores", "2"); + conf.set("hive.spark.client.connect.timeout", "30000ms"); // FIXME: Hadoop3 made the incompatible change for dfs.client.datanode-restart.timeout // while spark2 is still using Hadoop2. // Spark requires Hive to support Hadoop3 first then Spark can start diff --git a/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java b/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java index 23df7920d3..579ca0738a 100644 --- a/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java +++ b/spark-client/src/test/java/org/apache/hive/spark/client/TestSparkClient.java @@ -68,6 +68,10 @@ private static final long TIMEOUT = 20; private static final HiveConf HIVECONF = new HiveConf(); + static { + HIVECONF.set("hive.spark.client.connect.timeout", "30000ms"); + } + private Map createConf() { Map conf = new HashMap();