diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/HiveSparkClientFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/HiveSparkClientFactory.java index 6e9ba7c6df..194585e0f0 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/HiveSparkClientFactory.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/HiveSparkClientFactory.java @@ -95,21 +95,21 @@ public static HiveSparkClient createHiveSparkClient(HiveConf hiveconf) throws Ex inputStream = HiveSparkClientFactory.class.getClassLoader() .getResourceAsStream(SPARK_DEFAULT_CONF_FILE); if (inputStream != null) { - LOG.info("loading spark properties from:" + SPARK_DEFAULT_CONF_FILE); + LOG.info("loading spark properties from: " + SPARK_DEFAULT_CONF_FILE); Properties properties = new Properties(); properties.load(new InputStreamReader(inputStream, CharsetNames.UTF_8)); for (String propertyName : properties.stringPropertyNames()) { if (propertyName.startsWith("spark")) { String value = properties.getProperty(propertyName); sparkConf.put(propertyName, properties.getProperty(propertyName)); - LOG.info(String.format( + LOG.debug(String.format( "load spark property from %s (%s -> %s).", SPARK_DEFAULT_CONF_FILE, propertyName, LogUtils.maskIfPassword(propertyName,value))); } } } } catch (IOException e) { - LOG.info("Failed to open spark configuration file:" + LOG.info("Failed to open spark configuration file: " + SPARK_DEFAULT_CONF_FILE, e); } finally { if (inputStream != null) { @@ -156,7 +156,7 @@ public static HiveSparkClient createHiveSparkClient(HiveConf hiveconf) throws Ex if (propertyName.startsWith("spark")) { String value = hiveConf.get(propertyName); sparkConf.put(propertyName, value); - LOG.info(String.format( + LOG.debug(String.format( "load spark property from hive configuration (%s -> %s).", propertyName, LogUtils.maskIfPassword(propertyName,value))); } else if (propertyName.startsWith("yarn") && @@ -166,7 +166,7 @@ public static HiveSparkClient createHiveSparkClient(HiveConf hiveconf) throws Ex // started with spark prefix, Spark would remove spark.hadoop prefix lately and add // it to its hadoop configuration. sparkConf.put("spark.hadoop." + propertyName, value); - LOG.info(String.format( + LOG.debug(String.format( "load yarn property from hive configuration in %s mode (%s -> %s).", sparkMaster, propertyName, LogUtils.maskIfPassword(propertyName,value))); } else if (propertyName.equals(CommonConfigurationKeysPublic.FS_DEFAULT_NAME_KEY)) { @@ -180,19 +180,19 @@ public static HiveSparkClient createHiveSparkClient(HiveConf hiveconf) throws Ex // Spark problem. String value = hiveConf.get(propertyName); sparkConf.put("spark.hadoop." + propertyName, value); - LOG.info(String.format( + LOG.debug(String.format( "load HBase configuration (%s -> %s).", propertyName, LogUtils.maskIfPassword(propertyName,value))); } else if (propertyName.startsWith("oozie")) { String value = hiveConf.get(propertyName); sparkConf.put("spark." + propertyName, value); - LOG.info(String.format( + LOG.debug(String.format( "Pass Oozie configuration (%s -> %s).", propertyName, LogUtils.maskIfPassword(propertyName,value))); } if (RpcConfiguration.HIVE_SPARK_RSC_CONFIGS.contains(propertyName)) { String value = RpcConfiguration.getValue(hiveConf, propertyName); sparkConf.put(propertyName, value); - LOG.info(String.format( + LOG.debug(String.format( "load RPC property from hive configuration (%s -> %s).", propertyName, LogUtils.maskIfPassword(propertyName,value))); } diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/RemoteHiveSparkClient.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/RemoteHiveSparkClient.java index dbcf26a2ad..102e41b9c4 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/RemoteHiveSparkClient.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/RemoteHiveSparkClient.java @@ -359,12 +359,12 @@ public Serializable call(JobContext jc) throws Exception { } private void logConfigurations(JobConf localJobConf) { - if (LOG.isInfoEnabled()) { - LOG.info("Logging job configuration: "); + if (LOG.isDebugEnabled()) { + LOG.debug("Logging job configuration: "); StringBuilder outWriter = new StringBuilder(); // redact sensitive information before logging HiveConfUtil.dumpConfig(localJobConf, outWriter); - LOG.info(outWriter.toString()); + LOG.debug(outWriter.toString()); } } }