diff --git a/conf/log4j.properties b/conf/log4j.properties index 9d59c6e..bb5ce19 100644 --- a/conf/log4j.properties +++ b/conf/log4j.properties @@ -74,7 +74,7 @@ log4j.appender.console.layout.ConversionPattern=%d{ISO8601} %-5p [%t] %c{2}: %m% log4j.logger.org.apache.zookeeper=INFO #log4j.logger.org.apache.hadoop.fs.FSNamesystem=DEBUG -log4j.logger.org.apache.hadoop.hbase=DEBUG +log4j.logger.org.apache.hadoop.hbase=INFO # Make these two classes INFO-level. Make them DEBUG to see more zk debug. log4j.logger.org.apache.hadoop.hbase.zookeeper.ZKUtil=INFO log4j.logger.org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher=INFO diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionUtils.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionUtils.java index 1e04b7f..69ed527 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionUtils.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ConnectionUtils.java @@ -90,6 +90,6 @@ public class ConnectionUtils { int serversideMultiplier = c.getInt("hbase.client.serverside.retries.multiplier", 10); int retries = hcRetries * serversideMultiplier; c.setInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, retries); - log.debug(sn + " HConnection server-to-server retries=" + retries); + log.info(sn + " server-side HConnection retries=" + retries); } -} +} \ No newline at end of file diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java index 565ca3b..047d3f9 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java @@ -163,7 +163,7 @@ public class RecoverableZooKeeper { "previous attempt succeeded."); return; } - LOG.warn("Node " + path + " already deleted, retry=" + isRetry); + LOG.info("Node " + path + " already deleted, retry=" + isRetry); throw e; case CONNECTIONLOSS: @@ -505,8 +505,7 @@ public class RecoverableZooKeeper { Bytes.toStringBinary(data)); throw e; } - LOG.info("Node " + path + " already exists and this is not a " + - "retry"); + LOG.info("Node " + path + " already exists"); throw e; case CONNECTIONLOSS: diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperNodeTracker.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperNodeTracker.java index b13acc7..d81b3be 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperNodeTracker.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperNodeTracker.java @@ -35,7 +35,6 @@ import org.apache.zookeeper.KeeperException; */ @InterfaceAudience.Private public abstract class ZooKeeperNodeTracker extends ZooKeeperListener { - static final Log LOG = LogFactory.getLog(ZooKeeperNodeTracker.class); /** Path of node being tracked */ protected final String node; @@ -144,7 +143,7 @@ public abstract class ZooKeeperNodeTracker extends ZooKeeperListener { // It did not exists, and now it does. if (nodeExistsChecked){ - LOG.info("Node "+node+" now exists, resetting a watcher"); + LOG.debug("Node " + node + " now exists, resetting a watcher"); try { // This does not create a watch if the node does not exists this.data = ZKUtil.getDataAndWatch(watcher, node); diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ChecksumType.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ChecksumType.java index e160dac..7de74bd 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ChecksumType.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/ChecksumType.java @@ -65,7 +65,7 @@ public enum ChecksumType { // check if hadoop library is available try { ctor = ChecksumFactory.newConstructor(PURECRC32); - LOG.info("Checksum using " + PURECRC32); + LOG.debug("Checksum using " + PURECRC32); } catch (Exception e) { LOG.trace(PURECRC32 + " not available."); } @@ -74,7 +74,7 @@ public enum ChecksumType { // This is available on all JVMs. if (ctor == null) { ctor = ChecksumFactory.newConstructor(JDKCRC); - LOG.info("Checksum can use " + JDKCRC); + LOG.debug("Checksum can use " + JDKCRC); } } catch (Exception e) { LOG.trace(JDKCRC + " not available."); @@ -108,7 +108,7 @@ public enum ChecksumType { LOG = LogFactory.getLog(ChecksumType.class); try { ctor = ChecksumFactory.newConstructor(PURECRC32C); - LOG.info("Checksum can use " + PURECRC32C); + LOG.debug("Checksum can use " + PURECRC32C); } catch (Exception e) { LOG.trace(PURECRC32C + " not available."); } diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java index 2276543..292afe8 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java @@ -29,7 +29,6 @@ import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.io.hfile.BlockType.BlockCategory; import org.apache.hadoop.hbase.io.hfile.bucket.BucketCache; -import org.apache.hadoop.hbase.regionserver.StoreFile; import org.apache.hadoop.hbase.util.DirectMemoryUtils; import org.apache.hadoop.util.StringUtils; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java index f7208b6..e65b1d6 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java @@ -1458,7 +1458,7 @@ public class AssignmentManager extends ZooKeeperListener { if (regionCount == 0) { return true; } - LOG.debug("Assigning " + regionCount + " region(s) to " + destination.toString()); + LOG.info("Assigning " + regionCount + " region(s) to " + destination.toString()); Set encodedNames = new HashSet(regionCount); for (HRegionInfo region : regions) { encodedNames.add(region.getEncodedName()); @@ -1509,8 +1509,8 @@ public class AssignmentManager extends ZooKeeperListener { for (int oldCounter = 0; !server.isStopped();) { int count = counter.get(); if (oldCounter != count) { - LOG.info(destination.toString() + " unassigned znodes=" + count + - " of total=" + total); + LOG.debug(destination.toString() + " unassigned znodes=" + count + + " of total=" + total + "; oldCounter=" + oldCounter); oldCounter = count; } if (count >= total) break; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/RegionServerProcedureManagerHost.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/RegionServerProcedureManagerHost.java index f617cbd..6aa567a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/RegionServerProcedureManagerHost.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/RegionServerProcedureManagerHost.java @@ -40,17 +40,17 @@ public class RegionServerProcedureManagerHost extends public void initialize(RegionServerServices rss) throws KeeperException { for (RegionServerProcedureManager proc : procedures) { - LOG.info("Procedure " + proc.getProcedureSignature() + " is initializing"); + LOG.debug("Procedure " + proc.getProcedureSignature() + " is initializing"); proc.initialize(rss); - LOG.info("Procedure " + proc.getProcedureSignature() + " is initialized"); + LOG.debug("Procedure " + proc.getProcedureSignature() + " is initialized"); } } public void start() { for (RegionServerProcedureManager proc : procedures) { - LOG.info("Procedure " + proc.getProcedureSignature() + " is starting"); + LOG.debug("Procedure " + proc.getProcedureSignature() + " is starting"); proc.start(); - LOG.info("Procedure " + proc.getProcedureSignature() + " is started"); + LOG.debug("Procedure " + proc.getProcedureSignature() + " is started"); } } diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java index d94f029..07a43df 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java @@ -1785,7 +1785,7 @@ public class HRegionServer implements ClientProtos.ClientService.BlockingInterfa public void postOpenDeployTasks(final HRegion r, final CatalogTracker ct) throws KeeperException, IOException { checkOpen(); - LOG.info("Post open deploy tasks for region=" + r.getRegionNameAsString()); + LOG.debug("Post open deploy tasks for region=" + r.getRegionNameAsString()); // Do checks to see if we need to compact (references or too many files) for (Store s : r.getStores().values()) { if (s.hasReferences() || s.needsCompaction()) { @@ -1810,7 +1810,7 @@ public class HRegionServer implements ClientProtos.ClientService.BlockingInterfa MetaEditor.updateRegionLocation(ct, r.getRegionInfo(), this.serverNameFromMasterPOV, openSeqNum); } - LOG.info("Finished post open deploy task for " + r.getRegionNameAsString()); + LOG.debug("Finished post open deploy task for " + r.getRegionNameAsString()); } diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ShutdownHook.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ShutdownHook.java index 57a01d5..f01c55f 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ShutdownHook.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ShutdownHook.java @@ -84,7 +84,7 @@ public class ShutdownHook { Runnable fsShutdownHook = suppressHdfsShutdownHook(fs); Thread t = new ShutdownHookThread(conf, stop, threadToJoin, fsShutdownHook); ShutdownHookManager.affixShutdownHook(t, 0); - LOG.info("Installed shutdown hook thread: " + t.getName()); + LOG.debug("Installed shutdown hook thread: " + t.getName()); } /* diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogSplitter.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogSplitter.java index 15986d1..bbe6861 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogSplitter.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogSplitter.java @@ -354,7 +354,7 @@ public class HLogSplitter { e = RemoteExceptionHandler.checkIOException(e); throw e; } finally { - LOG.info("Finishing writing output logs and closing down."); + LOG.debug("Finishing writing output logs and closing down."); if (outputSinkStarted) { progress_failed = outputSink.finishWritingAndClose() == null; } @@ -450,7 +450,7 @@ public class HLogSplitter { if (!FSUtils.renameAndSetModifyTime(fs, p, newPath)) { LOG.warn("Unable to move " + p + " to " + newPath); } else { - LOG.debug("Archived processed log " + p + " to " + newPath); + LOG.info("Archived processed log " + p + " to " + newPath); } } } diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceManager.java index b9ac55b..9826489 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceManager.java @@ -47,7 +47,6 @@ import org.apache.hadoop.hbase.replication.ReplicationListener; import org.apache.hadoop.hbase.replication.ReplicationPeers; import org.apache.hadoop.hbase.replication.ReplicationQueues; import org.apache.hadoop.hbase.replication.ReplicationTracker; -import org.apache.zookeeper.KeeperException; import com.google.common.util.concurrent.ThreadFactoryBuilder;