From 4dea7cf1e979b2b00036f7ea80af2c91b4e1846c Mon Sep 17 00:00:00 2001 From: Ashutosh Chauhan Date: Thu, 22 Oct 2015 19:22:09 -0700 Subject: [PATCH] HIVE-12237 : Replace log4j direct references with slf4j facade --- .../hadoop/hive/accumulo/LazyAccumuloRow.java | 5 +++-- .../org/apache/hadoop/hive/accumulo/Utils.java | 5 +++-- .../hadoop/hive/accumulo/columns/ColumnMapper.java | 5 +++-- .../accumulo/columns/ColumnMappingFactory.java | 5 +++-- .../columns/HiveAccumuloColumnMapping.java | 5 +++-- .../hadoop/hive/accumulo/mr/HiveAccumuloSplit.java | 5 +++-- .../predicate/AccumuloPredicateHandler.java | 5 +++-- .../predicate/PrimitiveComparisonFilter.java | 5 +++-- .../hive/accumulo/predicate/PushdownTuple.java | 5 +++-- .../accumulo/predicate/compare/StringCompare.java | 3 --- .../hive/accumulo/serde/AccumuloRowSerializer.java | 5 +++-- .../accumulo/serde/AccumuloSerDeParameters.java | 5 +++-- .../serde/CompositeAccumuloRowIdFactory.java | 5 +++-- .../predicate/TestAccumuloPredicateHandler.java | 3 --- .../serde/DelimitedAccumuloRowIdFactory.java | 5 +++-- .../serde/FirstCharAccumuloCompositeRowId.java | 5 +++-- .../hive/accumulo/serde/TestAccumuloSerDe.java | 3 --- .../hive/metastore/TestHiveMetaStoreTxns.java | 3 ++- .../java/org/apache/hadoop/hive/ql/QTestUtil.java | 4 ---- .../java/org/apache/hive/jdbc/HiveDataSource.java | 24 ++++++++++++++-------- jdbc/src/java/org/apache/hive/jdbc/HiveDriver.java | 6 ++++++ .../hadoop/hive/llap/daemon/HistoryLogger.java | 5 +++-- .../registry/impl/LlapFixedRegistryImpl.java | 7 ++++--- .../daemon/registry/impl/LlapRegistryService.java | 5 +++-- .../daemon/registry/impl/LlapYarnRegistryImpl.java | 5 +++-- .../llap/daemon/services/impl/LlapWebServices.java | 2 -- .../hadoop/hive/ql/lockmgr/TestDbTxnManager.java | 13 +++++------- .../hive/ql/lockmgr/TestDummyTxnManager.java | 10 ++++----- .../hive/serde2/avro/TestTypeInfoToSchema.java | 7 ++++--- .../service/cli/operation/LogDivertAppender.java | 7 ++++--- .../service/cli/operation/OperationManager.java | 3 ++- 31 files changed, 98 insertions(+), 82 deletions(-) diff --git a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/LazyAccumuloRow.java b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/LazyAccumuloRow.java index 4597f5c..d5af7a8 100644 --- a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/LazyAccumuloRow.java +++ b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/LazyAccumuloRow.java @@ -35,7 +35,8 @@ import org.apache.hadoop.hive.serde2.lazy.objectinspector.LazySimpleStructObjectInspector; import org.apache.hadoop.hive.serde2.objectinspector.StructField; import org.apache.hadoop.io.Text; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * @@ -43,7 +44,7 @@ * */ public class LazyAccumuloRow extends LazyStruct { - private static final Logger log = Logger.getLogger(LazyAccumuloRow.class); + private static final Logger log = LoggerFactory.getLogger(LazyAccumuloRow.class); private AccumuloHiveRow row; private List columnMappings; diff --git a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/Utils.java b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/Utils.java index dc4782a..407ecbd 100644 --- a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/Utils.java +++ b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/Utils.java @@ -44,7 +44,8 @@ import org.apache.hadoop.fs.Path; import org.apache.hadoop.hive.common.JavaUtils; import org.apache.hadoop.util.StringUtils; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import com.google.common.base.Preconditions; @@ -53,7 +54,7 @@ * helpful */ public class Utils { - private static final Logger log = Logger.getLogger(Utils.class); + private static final Logger log = LoggerFactory.getLogger(Utils.class); // Thanks, HBase public static void addDependencyJars(Configuration conf, Class... classes) throws IOException { diff --git a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/ColumnMapper.java b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/ColumnMapper.java index 0498bab..b06b44a 100644 --- a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/ColumnMapper.java +++ b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/ColumnMapper.java @@ -25,7 +25,8 @@ import org.apache.hadoop.hive.accumulo.serde.TooManyAccumuloColumnsException; import org.apache.hadoop.hive.serde.serdeConstants; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import com.google.common.base.Preconditions; @@ -33,7 +34,7 @@ * */ public class ColumnMapper { - private static final Logger log = Logger.getLogger(ColumnMapper.class); + private static final Logger log = LoggerFactory.getLogger(ColumnMapper.class); private List columnMappings; private int rowIdOffset; diff --git a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/ColumnMappingFactory.java b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/ColumnMappingFactory.java index a241882..63d496e 100644 --- a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/ColumnMappingFactory.java +++ b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/ColumnMappingFactory.java @@ -20,7 +20,8 @@ import org.apache.hadoop.hive.accumulo.AccumuloHiveConstants; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import com.google.common.base.Preconditions; import com.google.common.collect.Maps; @@ -29,7 +30,7 @@ * */ public class ColumnMappingFactory { - private static final Logger log = Logger.getLogger(ColumnMappingFactory.class); + private static final Logger log = LoggerFactory.getLogger(ColumnMappingFactory.class); /** * Generate the proper instance of a ColumnMapping diff --git a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/HiveAccumuloColumnMapping.java b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/HiveAccumuloColumnMapping.java index d09ade1..85d883e 100644 --- a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/HiveAccumuloColumnMapping.java +++ b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/columns/HiveAccumuloColumnMapping.java @@ -17,7 +17,8 @@ package org.apache.hadoop.hive.accumulo.columns; import org.apache.hadoop.hive.accumulo.AccumuloHiveConstants; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import com.google.common.base.Charsets; @@ -26,7 +27,7 @@ */ public class HiveAccumuloColumnMapping extends ColumnMapping { @SuppressWarnings("unused") - private static final Logger log = Logger.getLogger(HiveAccumuloColumnMapping.class); + private static final Logger log = LoggerFactory.getLogger(HiveAccumuloColumnMapping.class); protected String columnFamily, columnQualifier; protected byte[] columnFamilyBytes, columnQualifierBytes; diff --git a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/mr/HiveAccumuloSplit.java b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/mr/HiveAccumuloSplit.java index 530f232..2a6be86 100644 --- a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/mr/HiveAccumuloSplit.java +++ b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/mr/HiveAccumuloSplit.java @@ -28,7 +28,8 @@ import org.apache.hadoop.mapred.FileSplit; import org.apache.hadoop.mapred.InputSplit; import org.apache.hadoop.util.StringUtils; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * Wraps RangeInputSplit into a FileSplit so Hadoop won't complain when it tries to make its own @@ -40,7 +41,7 @@ * error */ public class HiveAccumuloSplit extends FileSplit implements InputSplit { - private static final Logger log = Logger.getLogger(HiveAccumuloSplit.class); + private static final Logger log = LoggerFactory.getLogger(HiveAccumuloSplit.class); private RangeInputSplit split; diff --git a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/AccumuloPredicateHandler.java b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/AccumuloPredicateHandler.java index 534e77f..2c0e3c2 100644 --- a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/AccumuloPredicateHandler.java +++ b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/AccumuloPredicateHandler.java @@ -69,7 +69,8 @@ import org.apache.hadoop.hive.ql.udf.generic.GenericUDFOPNotEqual; import org.apache.hadoop.hive.serde.serdeConstants; import org.apache.hadoop.hive.serde2.SerDeException; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import com.google.common.collect.Lists; import com.google.common.collect.Maps; @@ -91,7 +92,7 @@ // Want to start sufficiently "high" enough in the iterator stack private static int iteratorCount = 50; - private static final Logger log = Logger.getLogger(AccumuloPredicateHandler.class); + private static final Logger log = LoggerFactory.getLogger(AccumuloPredicateHandler.class); static { compareOps.put(GenericUDFOPEqual.class.getName(), Equal.class); compareOps.put(GenericUDFOPNotEqual.class.getName(), NotEqual.class); diff --git a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/PrimitiveComparisonFilter.java b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/PrimitiveComparisonFilter.java index 4b5fae6..17d5529 100644 --- a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/PrimitiveComparisonFilter.java +++ b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/PrimitiveComparisonFilter.java @@ -39,7 +39,8 @@ import org.apache.hadoop.hive.common.JavaUtils; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory; import org.apache.hadoop.io.Text; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import com.google.common.collect.Lists; @@ -53,7 +54,7 @@ */ public class PrimitiveComparisonFilter extends WholeRowIterator { @SuppressWarnings("unused") - private static final Logger log = Logger.getLogger(PrimitiveComparisonFilter.class); + private static final Logger log = LoggerFactory.getLogger(PrimitiveComparisonFilter.class); public static final String FILTER_PREFIX = "accumulo.filter.compare.iterator."; public static final String P_COMPARE_CLASS = "accumulo.filter.iterator.p.compare.class"; diff --git a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/PushdownTuple.java b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/PushdownTuple.java index 32d143a..f326d52 100644 --- a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/PushdownTuple.java +++ b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/PushdownTuple.java @@ -33,7 +33,8 @@ import org.apache.hadoop.io.LongWritable; import org.apache.hadoop.io.Writable; import org.apache.hadoop.util.StringUtils; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * For use in IteratorSetting construction. @@ -41,7 +42,7 @@ * encapsulates a constant byte [], PrimitiveCompare instance, and CompareOp instance. */ public class PushdownTuple { - private static final Logger log = Logger.getLogger(PushdownTuple.class); + private static final Logger log = LoggerFactory.getLogger(PushdownTuple.class); private byte[] constVal; private PrimitiveComparison pCompare; diff --git a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/compare/StringCompare.java b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/compare/StringCompare.java index 0e038ad..3d6d55c 100644 --- a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/compare/StringCompare.java +++ b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/predicate/compare/StringCompare.java @@ -19,8 +19,6 @@ import java.util.regex.Pattern; -import org.apache.log4j.Logger; - /** * Set of comparison operations over a string constant. Used for Hive predicates involving string * comparison. @@ -29,7 +27,6 @@ */ public class StringCompare implements PrimitiveComparison { @SuppressWarnings("unused") - private static final Logger log = Logger.getLogger(StringCompare.class); private String constant; diff --git a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/serde/AccumuloRowSerializer.java b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/serde/AccumuloRowSerializer.java index 14facff..7ad6a45 100644 --- a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/serde/AccumuloRowSerializer.java +++ b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/serde/AccumuloRowSerializer.java @@ -42,7 +42,8 @@ import org.apache.hadoop.hive.serde2.objectinspector.primitive.PrimitiveObjectInspectorFactory; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoUtils; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import com.google.common.base.Charsets; import com.google.common.base.Preconditions; @@ -52,7 +53,7 @@ * {@link ColumnMapping}s */ public class AccumuloRowSerializer { - private static final Logger log = Logger.getLogger(AccumuloRowSerializer.class); + private static final Logger log = LoggerFactory.getLogger(AccumuloRowSerializer.class); private final int rowIdOffset; private final ByteStream.Output output; diff --git a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/serde/AccumuloSerDeParameters.java b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/serde/AccumuloSerDeParameters.java index 4dac675..09c5f24 100644 --- a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/serde/AccumuloSerDeParameters.java +++ b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/serde/AccumuloSerDeParameters.java @@ -34,7 +34,8 @@ import org.apache.hadoop.hive.serde2.lazy.LazySerDeParameters; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo; import org.apache.hadoop.util.ReflectionUtils; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import com.google.common.base.Preconditions; @@ -42,7 +43,7 @@ * */ public class AccumuloSerDeParameters extends AccumuloConnectionParameters { - private static final Logger log = Logger.getLogger(AccumuloSerDeParameters.class); + private static final Logger log = LoggerFactory.getLogger(AccumuloSerDeParameters.class); public static final String COLUMN_MAPPINGS = "accumulo.columns.mapping"; public static final String ITERATOR_PUSHDOWN_KEY = "accumulo.iterator.pushdown"; diff --git a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/serde/CompositeAccumuloRowIdFactory.java b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/serde/CompositeAccumuloRowIdFactory.java index 574a8aa..02d9736 100644 --- a/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/serde/CompositeAccumuloRowIdFactory.java +++ b/accumulo-handler/src/java/org/apache/hadoop/hive/accumulo/serde/CompositeAccumuloRowIdFactory.java @@ -27,7 +27,8 @@ import org.apache.hadoop.hive.serde2.SerDeException; import org.apache.hadoop.hive.serde2.lazy.objectinspector.LazySimpleStructObjectInspector; import org.apache.hadoop.hive.serde2.objectinspector.ObjectInspector; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * {@link AccumuloRowIdFactory} designed for injection of the {@link AccumuloCompositeRowId} to be @@ -39,7 +40,7 @@ public class CompositeAccumuloRowIdFactory extends DefaultAccumuloRowIdFactory { - public static final Logger log = Logger.getLogger(CompositeAccumuloRowIdFactory.class); + public static final Logger log = LoggerFactory.getLogger(CompositeAccumuloRowIdFactory.class); private final Class keyClass; private final Constructor constructor; diff --git a/accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/predicate/TestAccumuloPredicateHandler.java b/accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/predicate/TestAccumuloPredicateHandler.java index 97e14a2..15ccda7 100644 --- a/accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/predicate/TestAccumuloPredicateHandler.java +++ b/accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/predicate/TestAccumuloPredicateHandler.java @@ -77,7 +77,6 @@ import org.apache.hadoop.io.Text; import org.apache.hadoop.mapred.JobConf; import org.apache.hadoop.util.StringUtils; -import org.apache.log4j.Logger; import org.junit.Assert; import org.junit.Before; import org.junit.Test; @@ -87,8 +86,6 @@ import com.google.common.collect.Lists; public class TestAccumuloPredicateHandler { - @SuppressWarnings("unused") - private static final Logger log = Logger.getLogger(TestAccumuloPredicateHandler.class); private AccumuloPredicateHandler handler = AccumuloPredicateHandler.getInstance(); private JobConf conf; diff --git a/accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/serde/DelimitedAccumuloRowIdFactory.java b/accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/serde/DelimitedAccumuloRowIdFactory.java index 4bb5419..f885eba 100644 --- a/accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/serde/DelimitedAccumuloRowIdFactory.java +++ b/accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/serde/DelimitedAccumuloRowIdFactory.java @@ -29,14 +29,15 @@ import org.apache.hadoop.hive.serde2.objectinspector.StructField; import org.apache.hadoop.hive.serde2.objectinspector.StructObjectInspector; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * Example AccumuloRowIdFactory which accepts a delimiter that is used to separate the components of * some struct to place in the rowId. */ public class DelimitedAccumuloRowIdFactory extends DefaultAccumuloRowIdFactory { - private static final Logger log = Logger.getLogger(DelimitedAccumuloRowIdFactory.class); + private static final Logger log = LoggerFactory.getLogger(DelimitedAccumuloRowIdFactory.class); public static final String ACCUMULO_COMPOSITE_DELIMITER = "accumulo.composite.delimiter"; private byte separator; diff --git a/accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/serde/FirstCharAccumuloCompositeRowId.java b/accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/serde/FirstCharAccumuloCompositeRowId.java index f835a96..ed28e18 100644 --- a/accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/serde/FirstCharAccumuloCompositeRowId.java +++ b/accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/serde/FirstCharAccumuloCompositeRowId.java @@ -22,13 +22,14 @@ import org.apache.commons.lang.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.serde2.lazy.objectinspector.LazySimpleStructObjectInspector; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; /** * Gets the first character of each string in a struct */ public class FirstCharAccumuloCompositeRowId extends AccumuloCompositeRowId { - private static final Logger log = Logger.getLogger(FirstCharAccumuloCompositeRowId.class); + private static final Logger log = LoggerFactory.getLogger(FirstCharAccumuloCompositeRowId.class); private Properties tbl; private Configuration conf; diff --git a/accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/serde/TestAccumuloSerDe.java b/accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/serde/TestAccumuloSerDe.java index fb4f82b..58cac88 100644 --- a/accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/serde/TestAccumuloSerDe.java +++ b/accumulo-handler/src/test/org/apache/hadoop/hive/accumulo/serde/TestAccumuloSerDe.java @@ -55,7 +55,6 @@ import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory; import org.apache.hadoop.io.Text; -import org.apache.log4j.Logger; import org.junit.Assert; import org.junit.Before; import org.junit.Test; @@ -63,8 +62,6 @@ import com.google.common.base.Joiner; public class TestAccumuloSerDe { - @SuppressWarnings("unused") - private static final Logger log = Logger.getLogger(TestAccumuloSerDe.class); protected AccumuloSerDe serde; diff --git a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/TestHiveMetaStoreTxns.java b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/TestHiveMetaStoreTxns.java index 5ad5f35..e9ce789 100644 --- a/itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/TestHiveMetaStoreTxns.java +++ b/itests/hive-unit/src/test/java/org/apache/hadoop/hive/metastore/TestHiveMetaStoreTxns.java @@ -18,6 +18,7 @@ package org.apache.hadoop.hive.metastore; import junit.framework.Assert; + import org.apache.hadoop.hive.common.ValidTxnList; import org.apache.hadoop.hive.common.ValidReadTxnList; import org.apache.hadoop.hive.conf.HiveConf; @@ -47,7 +48,7 @@ */ public class TestHiveMetaStoreTxns { - private HiveConf conf = new HiveConf(); + private final HiveConf conf = new HiveConf(); private IMetaStoreClient client; public TestHiveMetaStoreTxns() throws Exception { diff --git a/itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java b/itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java index 6ddd8e4..1ca6618 100644 --- a/itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java +++ b/itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java @@ -102,8 +102,6 @@ import org.apache.hadoop.hive.shims.ShimLoader; import org.apache.hadoop.util.Shell; import org.apache.hive.common.util.StreamPrinter; -import org.apache.log4j.Level; -import org.apache.log4j.Logger; import org.apache.tools.ant.BuildException; import org.apache.zookeeper.WatchedEvent; import org.apache.zookeeper.Watcher; @@ -387,8 +385,6 @@ public QTestUtil(String outDir, String logDir, MiniClusterType clusterType, this.logDir = logDir; this.useHBaseMetastore = useHBaseMetastore; - Logger hadoopLog = Logger.getLogger("org.apache.hadoop"); - hadoopLog.setLevel(Level.INFO); if (confDir != null && !confDir.isEmpty()) { HiveConf.setHiveSiteLocation(new URL("file://"+ new File(confDir).toURI().getPath() + "/hive-site.xml")); System.out.println("Setting hive-site: "+HiveConf.getHiveSiteLocation()); diff --git a/jdbc/src/java/org/apache/hive/jdbc/HiveDataSource.java b/jdbc/src/java/org/apache/hive/jdbc/HiveDataSource.java index 459f08d..58feb97 100644 --- a/jdbc/src/java/org/apache/hive/jdbc/HiveDataSource.java +++ b/jdbc/src/java/org/apache/hive/jdbc/HiveDataSource.java @@ -41,20 +41,22 @@ public HiveDataSource() { /* * (non-Javadoc) - * + * * @see javax.sql.DataSource#getConnection() */ + @Override public Connection getConnection() throws SQLException { return getConnection("", ""); } /* * (non-Javadoc) - * + * * @see javax.sql.DataSource#getConnection(java.lang.String, java.lang.String) */ + @Override public Connection getConnection(String username, String password) throws SQLException { try { @@ -66,10 +68,11 @@ public Connection getConnection(String username, String password) /* * (non-Javadoc) - * + * * @see javax.sql.CommonDataSource#getLogWriter() */ + @Override public PrintWriter getLogWriter() throws SQLException { // TODO Auto-generated method stub throw new SQLException("Method not supported"); @@ -77,10 +80,11 @@ public PrintWriter getLogWriter() throws SQLException { /* * (non-Javadoc) - * + * * @see javax.sql.CommonDataSource#getLoginTimeout() */ + @Override public int getLoginTimeout() throws SQLException { // TODO Auto-generated method stub throw new SQLException("Method not supported"); @@ -93,10 +97,11 @@ public Logger getParentLogger() throws SQLFeatureNotSupportedException { /* * (non-Javadoc) - * + * * @see javax.sql.CommonDataSource#setLogWriter(java.io.PrintWriter) */ + @Override public void setLogWriter(PrintWriter arg0) throws SQLException { // TODO Auto-generated method stub throw new SQLException("Method not supported"); @@ -104,10 +109,11 @@ public void setLogWriter(PrintWriter arg0) throws SQLException { /* * (non-Javadoc) - * + * * @see javax.sql.CommonDataSource#setLoginTimeout(int) */ + @Override public void setLoginTimeout(int arg0) throws SQLException { // TODO Auto-generated method stub throw new SQLException("Method not supported"); @@ -115,10 +121,11 @@ public void setLoginTimeout(int arg0) throws SQLException { /* * (non-Javadoc) - * + * * @see java.sql.Wrapper#isWrapperFor(java.lang.Class) */ + @Override public boolean isWrapperFor(Class arg0) throws SQLException { // TODO Auto-generated method stub throw new SQLException("Method not supported"); @@ -126,10 +133,11 @@ public boolean isWrapperFor(Class arg0) throws SQLException { /* * (non-Javadoc) - * + * * @see java.sql.Wrapper#unwrap(java.lang.Class) */ + @Override public T unwrap(Class arg0) throws SQLException { // TODO Auto-generated method stub throw new SQLException("Method not supported"); diff --git a/jdbc/src/java/org/apache/hive/jdbc/HiveDriver.java b/jdbc/src/java/org/apache/hive/jdbc/HiveDriver.java index 396c314..dbc9612 100644 --- a/jdbc/src/java/org/apache/hive/jdbc/HiveDriver.java +++ b/jdbc/src/java/org/apache/hive/jdbc/HiveDriver.java @@ -92,6 +92,7 @@ public HiveDriver() { * TODO: - write a better regex. - decide on uri format */ + @Override public boolean acceptsURL(String url) throws SQLException { return Pattern.matches(Utils.URL_PREFIX + ".*", url); } @@ -101,6 +102,7 @@ public boolean acceptsURL(String url) throws SQLException { * "If the Driver implementation understands the URL, it will return a Connection object; * otherwise it returns null" */ + @Override public Connection connect(String url, Properties info) throws SQLException { return acceptsURL(url) ? new HiveConnection(url, info) : null; } @@ -156,6 +158,7 @@ static int getMinorDriverVersion() { /** * Returns the major version of this driver. */ + @Override public int getMajorVersion() { return HiveDriver.getMajorDriverVersion(); } @@ -163,6 +166,7 @@ public int getMajorVersion() { /** * Returns the minor version of this driver. */ + @Override public int getMinorVersion() { return HiveDriver.getMinorDriverVersion(); } @@ -172,6 +176,7 @@ public Logger getParentLogger() throws SQLFeatureNotSupportedException { throw new SQLFeatureNotSupportedException("Method not supported"); } + @Override public DriverPropertyInfo[] getPropertyInfo(String url, Properties info) throws SQLException { if (info == null) { info = new Properties(); @@ -208,6 +213,7 @@ public Logger getParentLogger() throws SQLFeatureNotSupportedException { /** * Returns whether the driver is JDBC compliant. */ + @Override public boolean jdbcCompliant() { return JDBC_COMPLIANT; } diff --git a/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/HistoryLogger.java b/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/HistoryLogger.java index b4fc618..3c9ad24 100644 --- a/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/HistoryLogger.java +++ b/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/HistoryLogger.java @@ -14,7 +14,8 @@ package org.apache.hadoop.hive.llap.daemon; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; public class HistoryLogger { @@ -36,7 +37,7 @@ private static final String EVENT_TYPE_FRAGMENT_START = "FRAGMENT_START"; private static final String EVENT_TYPE_FRAGMENT_END = "FRAGMENT_END"; - private static final Logger HISTORY_LOGGER = Logger.getLogger(HistoryLogger.class); + private static final Logger HISTORY_LOGGER = LoggerFactory.getLogger(HistoryLogger.class); public static void logFragmentStart(String applicationIdStr, String containerIdStr, String hostname, diff --git a/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/registry/impl/LlapFixedRegistryImpl.java b/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/registry/impl/LlapFixedRegistryImpl.java index 57aa1e7..621a6a6 100644 --- a/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/registry/impl/LlapFixedRegistryImpl.java +++ b/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/registry/impl/LlapFixedRegistryImpl.java @@ -31,11 +31,12 @@ import org.apache.hadoop.net.NetUtils; import org.apache.hadoop.util.StringUtils; import org.apache.hadoop.yarn.api.records.Resource; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; public class LlapFixedRegistryImpl implements ServiceRegistry { - private static final Logger LOG = Logger.getLogger(LlapFixedRegistryImpl.class); + private static final Logger LOG = LoggerFactory.getLogger(LlapFixedRegistryImpl.class); @InterfaceAudience.Private // This is primarily for testing to avoid the host lookup @@ -219,4 +220,4 @@ public ServiceInstanceSet getInstances(String component) throws IOException { public String toString() { return String.format("FixedRegistry hosts=%s", StringUtils.join(",", this.hosts)); } -} \ No newline at end of file +} diff --git a/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/registry/impl/LlapRegistryService.java b/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/registry/impl/LlapRegistryService.java index d3647d0..6550940 100644 --- a/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/registry/impl/LlapRegistryService.java +++ b/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/registry/impl/LlapRegistryService.java @@ -20,11 +20,12 @@ import org.apache.hadoop.hive.llap.daemon.registry.ServiceInstanceSet; import org.apache.hadoop.hive.llap.daemon.registry.ServiceRegistry; import org.apache.hadoop.service.AbstractService; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; public class LlapRegistryService extends AbstractService { - private static final Logger LOG = Logger.getLogger(LlapRegistryService.class); + private static final Logger LOG = LoggerFactory.getLogger(LlapRegistryService.class); private ServiceRegistry registry = null; private final boolean isDaemon; diff --git a/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/registry/impl/LlapYarnRegistryImpl.java b/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/registry/impl/LlapYarnRegistryImpl.java index cb1b1d0..599da13 100644 --- a/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/registry/impl/LlapYarnRegistryImpl.java +++ b/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/registry/impl/LlapYarnRegistryImpl.java @@ -48,7 +48,8 @@ import org.apache.hadoop.registry.client.types.ServiceRecord; import org.apache.hadoop.yarn.api.records.Resource; import org.apache.hadoop.yarn.conf.YarnConfiguration; -import org.apache.log4j.Logger; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import org.apache.tez.dag.api.TezUncheckedException; import org.apache.zookeeper.CreateMode; @@ -56,7 +57,7 @@ public class LlapYarnRegistryImpl implements ServiceRegistry { - private static final Logger LOG = Logger.getLogger(LlapYarnRegistryImpl.class); + private static final Logger LOG = LoggerFactory.getLogger(LlapYarnRegistryImpl.class); private final RegistryOperationsService client; private final Configuration conf; diff --git a/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/services/impl/LlapWebServices.java b/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/services/impl/LlapWebServices.java index 2275719..9520413 100644 --- a/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/services/impl/LlapWebServices.java +++ b/llap-server/src/java/org/apache/hadoop/hive/llap/daemon/services/impl/LlapWebServices.java @@ -6,11 +6,9 @@ import org.apache.hadoop.yarn.conf.YarnConfiguration; import org.apache.hadoop.yarn.webapp.WebApp; import org.apache.hadoop.yarn.webapp.WebApps; -import org.apache.log4j.Logger; public class LlapWebServices extends AbstractService { - private static final Logger LOG = Logger.getLogger(LlapWebServices.class); private int port; private boolean ssl; diff --git a/ql/src/test/org/apache/hadoop/hive/ql/lockmgr/TestDbTxnManager.java b/ql/src/test/org/apache/hadoop/hive/ql/lockmgr/TestDbTxnManager.java index 68c6542..c1e5c81 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/lockmgr/TestDbTxnManager.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/lockmgr/TestDbTxnManager.java @@ -33,7 +33,7 @@ import org.apache.hadoop.hive.ql.session.SessionState; import org.apache.hadoop.hive.ql.txn.AcidHouseKeeperService; import org.apache.log4j.Level; -import org.apache.log4j.LogManager; +import org.slf4j.LoggerFactory; import static org.hamcrest.CoreMatchers.is; import org.junit.After; import org.junit.Assert; @@ -49,12 +49,11 @@ */ public class TestDbTxnManager { - private HiveConf conf = new HiveConf(); + private final HiveConf conf = new HiveConf(); private HiveTxnManager txnMgr; private AcidHouseKeeperService houseKeeperService = null; - private Context ctx; + private final Context ctx; private int nextInput; - private int nextOutput; HashSet readEntities; HashSet writeEntities; @@ -62,7 +61,6 @@ public TestDbTxnManager() throws Exception { TxnDbUtil.setConfValues(conf); SessionState.start(conf); ctx = new Context(conf); - LogManager.getRootLogger().setLevel(Level.DEBUG); tearDown(); } @@ -363,7 +361,6 @@ public void setUp() throws Exception { txnMgr = TxnManagerFactory.getTxnManagerFactory().getTxnManager(conf); Assert.assertTrue(txnMgr instanceof DbTxnManager); nextInput = 1; - nextOutput = 1; readEntities = new HashSet(); writeEntities = new HashSet(); conf.setTimeVar(HiveConf.ConfVars.HIVE_TIMEDOUT_TXN_REAPER_START, 0, TimeUnit.SECONDS); @@ -379,8 +376,8 @@ public void tearDown() throws Exception { } private static class MockQueryPlan extends QueryPlan { - private HashSet inputs; - private HashSet outputs; + private final HashSet inputs; + private final HashSet outputs; MockQueryPlan(TestDbTxnManager test) { HashSet r = test.readEntities; diff --git a/ql/src/test/org/apache/hadoop/hive/ql/lockmgr/TestDummyTxnManager.java b/ql/src/test/org/apache/hadoop/hive/ql/lockmgr/TestDummyTxnManager.java index 19f82ad..0fc87ae 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/lockmgr/TestDummyTxnManager.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/lockmgr/TestDummyTxnManager.java @@ -33,8 +33,6 @@ import org.apache.hadoop.hive.ql.lockmgr.zookeeper.ZooKeeperHiveLock; import org.apache.hadoop.hive.ql.metadata.Table; import org.apache.hadoop.hive.ql.session.SessionState; -import org.apache.log4j.Level; -import org.apache.log4j.LogManager; import org.junit.Test; import org.junit.runner.RunWith; import org.mockito.ArgumentCaptor; @@ -50,7 +48,7 @@ @RunWith(MockitoJUnitRunner.class) public class TestDummyTxnManager { - private HiveConf conf = new HiveConf(); + private final HiveConf conf = new HiveConf(); private HiveTxnManager txnMgr; private Context ctx; private int nextInput = 1; @@ -67,7 +65,6 @@ public void setUp() throws Exception { conf.setVar(HiveConf.ConfVars.HIVE_TXN_MANAGER, DummyTxnManager.class.getName()); SessionState.start(conf); ctx = new Context(conf); - LogManager.getRootLogger().setLevel(Level.DEBUG); txnMgr = TxnManagerFactory.getTxnManagerFactory().getTxnManager(conf); Assert.assertTrue(txnMgr instanceof DummyTxnManager); @@ -116,8 +113,8 @@ public void testSingleReadTable() throws Exception { Assert.assertEquals(expectedLocks.get(1).getHiveLockMode(), resultLocks.get(1).getHiveLockMode()); Assert.assertEquals(expectedLocks.get(0).getHiveLockObject().getName(), resultLocks.get(0).getHiveLockObject().getName()); - verify(mockLockManager).lock((List)lockObjsCaptor.capture(), eq(false)); - List lockObjs = (List)lockObjsCaptor.getValue(); + verify(mockLockManager).lock(lockObjsCaptor.capture(), eq(false)); + List lockObjs = lockObjsCaptor.getValue(); Assert.assertEquals(2, lockObjs.size()); Assert.assertEquals("default", lockObjs.get(0).getName()); Assert.assertEquals(HiveLockMode.SHARED, lockObjs.get(0).mode); @@ -157,6 +154,7 @@ public void testDedupLockObjects() { Assert.assertEquals("Locks should be deduped", 2, lockObjs.size()); Comparator cmp = new Comparator() { + @Override public int compare(HiveLockObj lock1, HiveLockObj lock2) { return lock1.getName().compareTo(lock2.getName()); } diff --git a/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestTypeInfoToSchema.java b/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestTypeInfoToSchema.java index abbf038..ac0a8ee 100644 --- a/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestTypeInfoToSchema.java +++ b/serde/src/test/org/apache/hadoop/hive/serde2/avro/TestTypeInfoToSchema.java @@ -33,9 +33,10 @@ import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory; import org.apache.hadoop.hive.serde2.typeinfo.UnionTypeInfo; import org.apache.hadoop.hive.serde2.typeinfo.VarcharTypeInfo; -import org.apache.log4j.Logger; import org.junit.Before; import org.junit.Test; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; import java.io.IOException; import java.util.ArrayList; @@ -44,7 +45,7 @@ public class TestTypeInfoToSchema { - private static Logger LOGGER = Logger.getLogger(TestTypeInfoToSchema.class); + private static Logger LOGGER = LoggerFactory.getLogger(TestTypeInfoToSchema.class); private static final List COLUMN_NAMES = Arrays.asList("testCol"); private static final TypeInfo STRING = TypeInfoFactory.getPrimitiveTypeInfo( serdeConstants.STRING_TYPE_NAME); @@ -434,4 +435,4 @@ public void createAvroNestedStructSchema() throws IOException { Assert.assertEquals("Test for nested struct's avro schema failed", expectedSchema, getAvroSchemaString(superStructTypeInfo)); } -} \ No newline at end of file +} diff --git a/service/src/java/org/apache/hive/service/cli/operation/LogDivertAppender.java b/service/src/java/org/apache/hive/service/cli/operation/LogDivertAppender.java index c1bc547..9cb6439 100644 --- a/service/src/java/org/apache/hive/service/cli/operation/LogDivertAppender.java +++ b/service/src/java/org/apache/hive/service/cli/operation/LogDivertAppender.java @@ -39,6 +39,7 @@ import org.apache.logging.log4j.core.config.Configuration; import org.apache.logging.log4j.core.filter.AbstractFilter; import org.apache.logging.log4j.core.layout.PatternLayout; +import org.slf4j.LoggerFactory; import com.google.common.base.Joiner; @@ -47,7 +48,7 @@ */ public class LogDivertAppender extends AbstractOutputStreamAppender { - private static final Logger LOG = LogManager.getLogger(LogDivertAppender.class.getName()); + private static final org.slf4j.Logger LOG = LoggerFactory.getLogger(LogDivertAppender.class.getName()); private static LoggerContext context = (LoggerContext) LogManager.getContext(false); private static Configuration configuration = context.getConfiguration(); public static final Layout verboseLayout = PatternLayout.createLayout( @@ -56,7 +57,7 @@ "%-5p : %m%n", configuration, null, null, true, false, null, null); private final OperationManager operationManager; - private StringOutputStreamManager manager; + private final StringOutputStreamManager manager; private boolean isVerbose; private final Layout layout; @@ -105,7 +106,7 @@ protected LogDivertAppender(String name, Filter filter, private static class NameFilter extends AbstractFilter { private Pattern namePattern; private OperationLog.LoggingLevel loggingMode; - private OperationManager operationManager; + private final OperationManager operationManager; /* Patterns that are excluded in verbose logging level. * Filter out messages coming from log processing classes, or we'll run an infinite loop. diff --git a/service/src/java/org/apache/hive/service/cli/operation/OperationManager.java b/service/src/java/org/apache/hive/service/cli/operation/OperationManager.java index e29b4b6..a089b67 100644 --- a/service/src/java/org/apache/hive/service/cli/operation/OperationManager.java +++ b/service/src/java/org/apache/hive/service/cli/operation/OperationManager.java @@ -49,6 +49,7 @@ import org.apache.logging.log4j.core.LoggerContext; import org.apache.logging.log4j.core.config.Configuration; import org.apache.logging.log4j.core.config.LoggerConfig; +import org.slf4j.LoggerFactory; /** * OperationManager. @@ -91,7 +92,7 @@ private void initOperationLogCapture(String loggingMode) { Appender ap = LogDivertAppender.createInstance(this, OperationLog.getLoggingLevel(loggingMode)); LoggerContext context = (LoggerContext) LogManager.getContext(false); Configuration configuration = context.getConfiguration(); - LoggerConfig loggerConfig = configuration.getLoggerConfig(LogManager.getLogger().getName()); + LoggerConfig loggerConfig = configuration.getLoggerConfig(LoggerFactory.getLogger(getClass()).getName()); loggerConfig.addAppender(ap, null, null); context.updateLoggers(); ap.start(); -- 1.7.12.4 (Apple Git-37)