WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-common/src/main/java/org/apache/hadoop/hbase/util/DynamicClassLoader.java:167 try { URL url = file.toURI().toURL(); addURL(url); } catch (MalformedURLException mue) { LOG.warn("Failed to load new jar " + fileName, mue); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-common/src/main/java/org/apache/hadoop/hbase/util/DynamicClassLoader.java:218 try { File dst = new File(localDir, fileName); remoteDirFs.copyToLocalFile(path, new Path(dst.getPath())); jarModifiedTime.put(fileName, Long.valueOf(dst.lastModified())); URL url = dst.toURI().toURL(); addURL(url); } catch (IOException ioe) { LOG.warn("Failed to load new jar " + fileName, ioe); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java:127 try { checkZk(); } catch (Exception x) { } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncProcess.java:957 try { pool.submit(replicaRunnable); } catch (RejectedExecutionException ree) { LOG.warn("#" + id + ", replica task was rejected by the pool - no replica calls", ree); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AsyncProcess.java:1148 try { this.callback.update(regionName, sentAction.getAction().getRow(), (CResult)result); } catch (Throwable t) { LOG.error("User callback threw an exception for " + Bytes.toStringBinary(regionName) + ", ignoring", t); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeerZKImpl.java:248 try { readPeerStateZnode(); } catch (DeserializationException e) { LOG.warn("Failed deserializing the content of " + path, e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaTableLocator.java:369 try { ZKUtil.deleteNode(zookeeper, zookeeper.metaServerZNode); } catch (KeeperException.NoNodeException nne) { } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKLeaderManager.java:143 try { leaderExists.wait(); } catch (InterruptedException ie) { LOG.debug("Interrupted waiting on leader", ie); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/RegexStringComparator.java:158 try { comparator.setCharset(Charset.forName(charset)); } catch (IllegalCharsetNameException e) { LOG.error("invalid charset", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java:1706 try { String clusterKey = ZooKeeperProtos.ReplicationPeer.newBuilder().mergeFrom(data, pblen, data.length - pblen).getClusterkey(); sb.append("\n").append(znodeToProcess).append(": ").append(clusterKey); appendPeerState(zkw, znodeToProcess, sb); } catch (InvalidProtocolBufferException ipbe) { LOG.warn("Got Exception while parsing peer: " + znodeToProcess, ipbe); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java:579 try { start = elapsed = EnvironmentEdgeManager.currentTime(); processingList.clear(); failedCount = 0; queue.drainTo(processingList); currentProcessingPutCount.set(processingList.size()); if (processingList.size() > 0) { ArrayList list = new ArrayList(processingList.size()); for (PutStatus putStatus : processingList) { list.add(putStatus.getPut()); } List failed = null; Object[] results = new Object[list.size()]; try { htable.batch(list, results); } catch (IOException e) { LOG.debug("Caught some exceptions " + e + " when flushing puts to region server " + addr.getHostnamePort()); } finally { for (int i = results.length - 1; i >= 0; i--) { if (results[i] instanceof Result) { list.remove(i); } } failed = list; } if (failed != null) { if (failed.size() == processingList.size()) { for (PutStatus putStatus : processingList) { if (!resubmitFailedPut(putStatus, this.addr)) { failedCount++; } } } else { Set failedPutSet = new HashSet(failed); for (PutStatus putStatus : processingList) { if (failedPutSet.contains(putStatus.getPut()) && !resubmitFailedPut(putStatus, this.addr)) { failedCount++; } } } } this.totalFailedPutCount.addAndGet(failedCount); elapsed = EnvironmentEdgeManager.currentTime() - start; averageLatency.add(elapsed); if (elapsed > maxLatency.get()) { maxLatency.set(elapsed); } if (LOG.isDebugEnabled()) { LOG.debug("Processed " + currentProcessingPutCount + " put requests for " + addr.getHostnamePort() + " and " + failedCount + " failed" + ", latency for this send: " + elapsed); } currentProcessingPutCount.set(0); } if (elapsed == start) { elapsed = EnvironmentEdgeManager.currentTime() - start; } if (elapsed < frequency) { try { Thread.sleep(frequency - elapsed); } catch (InterruptedException e) { LOG.warn("Interrupted while sleeping"); Thread.currentThread().interrupt(); } } } catch (Exception e) { LOG.debug("Caught some exceptions " + e + " when flushing puts to region server " + addr.getHostnamePort()); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java:494 try { createTableAsync(desc, splitKeys); } catch (SocketTimeoutException ste) { LOG.warn("Creating " + desc.getTableName() + " took too long", ste); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java:3162 try { LOG.info("Deleting restore-failsafe snapshot: " + failSafeSnapshotSnapshotName); deleteSnapshot(failSafeSnapshotSnapshotName); } catch (IOException e) { LOG.error("Unable to remove the failsafe snapshot: " + failSafeSnapshotSnapshotName, e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java:1031 try { getRegionPlan(regionState.getRegion(), sn, true); new ClosedRegionHandler(server, this, regionState.getRegion()).process(); } catch (HBaseIOException e) { LOG.warn("Failed to get region plan", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java:2311 try { processFavoredNodes(regions); } catch (IOException ie) { LOG.warn("Ignoring exception in processFavoredNodes " + ie); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java:3399 try { getRegionPlan(hri, sn, true); } catch (HBaseIOException e) { LOG.warn("Failed to get region plan", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/TableLockManager.java:412 try { for (String tableName : getTableNames()) { String tableLockZNode = ZKUtil.joinZNode(zkWatcher.tableLockZNode, tableName); ZKInterProcessReadWriteLock lock = new ZKInterProcessReadWriteLock(zkWatcher, tableLockZNode, null); lock.writeLock(null).reapAllLocks(); } } catch (IOException ex) { throw ex; } catch (Exception ex) { LOG.warn("Caught exception while reaping table write locks", ex); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java:303 try { handleCoprocessorThrowable(env, e); } catch (IOException ioe) { LOG.warn("handleCoprocessorThrowable() threw an IOException while attempting to handle Throwable " + e + ". Ignoring.", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java:338 try { execOperation(coprocessors.isEmpty() ? null : new RegionOperation(){ () { super(); } @Override() public void call(RegionObserver oserver, ObserverContext ctx) throws IOException { oserver.postOpen(ctx); } }); } catch (IOException e) { LOG.warn(e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java:355 try { execOperation(coprocessors.isEmpty() ? null : new RegionOperation(){ () { super(); } @Override() public void call(RegionObserver oserver, ObserverContext ctx) throws IOException { oserver.postLogReplay(ctx); } }); } catch (IOException e) { LOG.warn(e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RegionCoprocessorHost.java:450 try { execOperation(coprocessors.isEmpty() ? null : new RegionOperation(){ () { super(); } @Override() public void call(RegionObserver oserver, ObserverContext ctx) throws IOException { oserver.postCompactSelection(ctx, store, selected, request); } }); } catch (IOException e) { LOG.warn(e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java:524 try { in.readByte(); } catch (EOFException e) { } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/FSHLog.java:634 try { nextWriter.sync(); postSync(System.nanoTime() - startTimeNanos, 0); } catch (IOException e) { LOG.warn("pre-sync failed but an optimization so keep going", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/wal/FSHLog.java:1317 try { while (true) { takeSyncFuture = this.syncFutures.take(); currentSequence = this.sequence; long syncFutureSequence = takeSyncFuture.getRingBufferSequence(); if (syncFutureSequence > currentSequence) { throw new IllegalStateException("currentSequence=" + syncFutureSequence + ", syncFutureSequence=" + syncFutureSequence); } long currentHighestSyncedSequence = highestSyncedSequence.get(); if (currentSequence < currentHighestSyncedSequence) { syncCount += releaseSyncFuture(takeSyncFuture, currentHighestSyncedSequence, null); continue; } break; } TraceScope scope = Trace.continueSpan(takeSyncFuture.getSpan()); long start = System.nanoTime(); Throwable t = null; try { Trace.addTimelineAnnotation("syncing writer"); writer.sync(); Trace.addTimelineAnnotation("writer synced"); currentSequence = updateHighestSyncedSequence(currentSequence); } catch (IOException e) { LOG.error("Error syncing, request close of hlog ", e); t = e; } catch (Exception e) { LOG.warn("UNEXPECTED", e); t = e; } finally { takeSyncFuture.setSpan(scope.detach()); syncCount += releaseSyncFuture(takeSyncFuture, currentSequence, t); syncCount += releaseSyncFutures(currentSequence, t); if (t != null) { requestLogRoll(); } else checkLogRoll(); } postSync(System.nanoTime() - start, syncCount); } catch (InterruptedException e) { Thread.currentThread().interrupt(); } catch (Throwable t) { LOG.warn("UNEXPECTED, continuing", t); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/LruBlockCache.java:791 try { this.wait(1000 * 10); } catch (InterruptedException e) { } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.java:723 try { while (cacheEnabled && writerEnabled) { try { try { entries = getRAMQueueEntries(inputQueue, entries); synchronized (cacheWaitSignals[threadNO]) { cacheWaitSignals[threadNO].notifyAll(); } } catch (InterruptedException ie) { if (!cacheEnabled) break; } doDrain(entries); } catch (Exception ioe) { LOG.error("WriterThread encountered error", ioe); } } } catch (Throwable t) { LOG.warn("Failed doing drain", t); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java:587 try { readSelector.select(); while (adding) { this.wait(1000); } Iterator iter = readSelector.selectedKeys().iterator(); while (iter.hasNext()) { SelectionKey key = iter.next(); iter.remove(); if (key.isValid()) { if (key.isReadable()) { doRead(key); } } } } catch (InterruptedException e) { LOG.debug("Interrupted while sleeping"); return; } catch (IOException ex) { LOG.info(getName() + ": IOException in Reader", ex); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java:879 try { c.channel.register(writeSelector, SelectionKey.OP_WRITE, c); } catch (ClosedChannelException e) { } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java:999 try { key.interestOps(0); } catch (CancelledKeyException e) { LOG.warn("Exception while changing ops : " + e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java:1184 try { socket.setSendBufferSize(socketSendBufferSize); } catch (IOException e) { LOG.warn("Connection: unable to set socket send buffer size to " + socketSendBufferSize); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java:233 try { if (master.cpHost != null) { newValue = master.cpHost.preBalanceSwitch(newValue); } try { if (mode == BalanceSwitchMode.SYNC) { synchronized (master.balancer) { master.loadBalancerTracker.setBalancerOn(newValue); } } else { master.loadBalancerTracker.setBalancerOn(newValue); } } catch (KeeperException ke) { throw new IOException(ke); } LOG.info(master.getClientIdAuditPrefix() + " set balanceSwitch=" + newValue); if (master.cpHost != null) { master.cpHost.postBalanceSwitch(oldValue, newValue); } } catch (IOException ioe) { LOG.warn("Error flipping balance switch", ioe); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java:247 try { HRegion region = regionServer.getRegion(s.getRegionInfo().getRegionName()); if (region != null && region.getCoprocessorHost() != null) { region.getCoprocessorHost().preScannerClose(s); } s.close(); if (region != null && region.getCoprocessorHost() != null) { region.getCoprocessorHost().postScannerClose(s); } } catch (IOException e) { LOG.error("Closing scanner for " + s.getRegionInfo().getRegionNameAsString(), e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java:610 try { this.cpHost.preMasterInitialization(); } catch (IOException e) { LOG.error("Coprocessor preMasterInitialization() hook failed", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java:627 try { this.cpHost.postStartMaster(); } catch (IOException ioe) { LOG.error("Coprocessor postStartMaster() hook failed", ioe); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java:1009 try { this.cpHost.postBalance(rpCount < plans.size() ? plans.subList(0, rpCount) : plans); } catch (IOException ioe) { LOG.error("Error invoking master coprocessor postBalance()", ioe); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java:1520 try { byte[] bytes; try { bytes = ZKUtil.getData(this.zooKeeper, ZKUtil.joinZNode(this.zooKeeper.backupMasterAddressesZNode, s)); } catch (InterruptedException e) { throw new InterruptedIOException(); } if (bytes != null) { ServerName sn; try { sn = ServerName.parseFrom(bytes); } catch (DeserializationException e) { LOG.warn("Failed parse, skipping registering backup server", e); continue; } backupMasters.add(sn); } } catch (KeeperException e) { LOG.warn(this.zooKeeper.prefix("Unable to get information about backup servers"), e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java:930 try { deleteMyEphemeralNode(); } catch (KeeperException e) { LOG.warn("Failed deleting my ephemeral node", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java:1358 try { long multiplier = s.getCompactionCheckMultiplier(); assert multiplier > 0; if (iteration % multiplier != 0) continue; if (s.needsCompaction()) { this.instance.compactSplitThread.requestSystemCompaction(r, s, getName() + " requests compaction"); } else if (s.isMajorCompaction()) { if (majorCompactPriority == DEFAULT_PRIORITY || majorCompactPriority > r.getCompactPriority()) { this.instance.compactSplitThread.requestCompaction(r, s, getName() + " requests major compaction; use default priority", null); } else { this.instance.compactSplitThread.requestCompaction(r, s, getName() + " requests major compaction; use configured priority", this.majorCompactPriority, null); } } } catch (IOException e) { LOG.warn("Failed major compaction check on " + r, e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java:2888 try { long lastRecordedFlushedSequenceId = -1; String nodePath = ZKUtil.joinZNode(this.zooKeeper.recoveringRegionsZNode, region.getEncodedName()); byte[] data; try { data = ZKUtil.getData(zkw, nodePath); } catch (InterruptedException e) { throw new InterruptedIOException(); } if (data != null) { lastRecordedFlushedSequenceId = ZKSplitLog.parseLastFlushedSequenceIdFrom(data); } if (data == null || lastRecordedFlushedSequenceId < minSeqIdForLogReplay) { ZKUtil.setData(zkw, nodePath, ZKUtil.positionToByteArray(minSeqIdForLogReplay)); } if (previousRSName != null) { nodePath = ZKUtil.joinZNode(nodePath, previousRSName); ZKUtil.setData(zkw, nodePath, ZKUtil.regionSequenceIdsToByteArray(minSeqIdForLogReplay, maxSeqIdInStores)); LOG.debug("Update last flushed sequence id of region " + region.getEncodedName() + " for " + previousRSName); } else { LOG.warn("Can\'t find failed region server for recovering region " + region.getEncodedName()); } } catch (NoNodeException ignore) { LOG.debug("Region " + region.getEncodedName() + " must have completed recovery because its recovery znode has been removed", ignore); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ActiveMasterManager.java:220 try { clusterHasActiveMaster.wait(checkInterval); } catch (InterruptedException e) { LOG.debug("Interrupted waiting for master to die", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/BalancerChore.java:51 try { master.balance(); } catch (IOException e) { LOG.error("Failed to balance.", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/ClusterStatusChore.java:51 try { balancer.setClusterStatus(master.getClusterStatus()); } catch (InterruptedIOException e) { LOG.warn("Ignoring interruption", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ProcedureManagerHost.java:75 try { implClass = cl.loadClass(className); configured.add(loadInstance(implClass)); LOG.info("User procedure " + className + " was loaded successfully."); } catch (ClassNotFoundException e) { LOG.warn("Class " + className + " cannot be found. " + e.getMessage()); } catch (IOException e) { LOG.warn("Load procedure " + className + " failed. " + e.getMessage()); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/AuthenticationTokenSecretManager.java:110 try { this.zkWatcher.start(); this.leaderElector.start(); } catch (KeeperException ke) { LOG.error("Zookeeper initialization failed", ke); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/MemStoreFlusher.java:392 try { this.server.compactSplitThread.requestSystemCompaction(region, Thread.currentThread().getName()); } catch (IOException e) { LOG.error("Cache flush failed for region " + Bytes.toStringBinary(region.getRegionName()), RemoteExceptionHandler.checkIOException(e)); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoveringRegionWatcher.java:91 try { ZKUtil.getDataAndWatch(watcher, path); } catch (KeeperException e) { LOG.warn("Can\'t register watcher on znode " + path, e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java:196 try { mergePool.execute(new RegionMergeRequest(a, b, this.server, forcible)); if (LOG.isDebugEnabled()) { LOG.debug("Region merge requested for " + a + "," + b + ", forcible=" + forcible + ". " + this); } } catch (RejectedExecutionException ree) { LOG.warn("Could not execute merge for " + a + "," + b + ", forcible=" + forcible, ree); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java:228 try { this.splits.execute(new SplitRequest(r, midKey, this.server)); if (LOG.isDebugEnabled()) { LOG.debug("Split requested for " + r + ". " + this); } } catch (RejectedExecutionException ree) { LOG.info("Could not execute split for " + r, ree); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/zookeeper/ClusterStatusTracker.java:84 try { ZKUtil.deleteNode(watcher, watcher.clusterStateZNode); } catch (KeeperException.NoNodeException nne) { LOG.warn("Attempted to set cluster as down but already down, cluster " + "state node (" + watcher.clusterStateZNode + ") not found"); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/TakeSnapshotHandler.java:231 try { this.tableLock.release(); } catch (IOException ex) { LOG.warn("Could not release the table lock", ex); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/ZKSecretWatcher.java:90 try { Integer id = new Integer(keyId); secretManager.removeKey(id); } catch (NumberFormatException nfe) { LOG.error("Invalid znode name for key ID \'" + keyId + "\'", nfe); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/ZKSplitLogManagerCoordination.java:1001 try { getDataSetWatchSuccess(path, data, stat.getVersion()); } catch (DeserializationException e) { LOG.warn("Deserialization problem", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/HBaseInterClusterReplicationEndpoint.java:128 try { if (LOG.isTraceEnabled()) { LOG.trace(msg + ", sleeping " + sleepForRetries + " times " + sleepMultiplier); } Thread.sleep(this.sleepForRetries * sleepMultiplier); } catch (InterruptedException e) { LOG.debug("Interrupted while sleeping between retries"); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/HBaseReplicationEndpoint.java:73 try { reloadZkWatcher(); } catch (IOException io) { LOG.warn("Creation of ZookeeperWatcher failed for peer " + clusterKey, io); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java:79 try { this.labelsCache.refreshLabelsCache(data); } catch (IOException ioe) { LOG.error("Failed parsing data from labels table from zk", ioe); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/ZKVisibilityLabelWatcher.java:87 try { this.labelsCache.refreshUserAuthsCache(data); } catch (IOException ioe) { LOG.error("Failed parsing data from labels table from zk", ioe); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/Subprocedure.java:113 try { rpcs.sendMemberAborted(Subprocedure.this, ee); } catch (IOException e) { LOG.error("Can\'t reach controller, not propagating error", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.java:298 try { for (Map.Entry entry : scanMetrics.getMetricsMap().entrySet()) { Counter ct = (Counter)getCounter.invoke(context, HBASE_COUNTER_GROUP_NAME, entry.getKey()); ct.increment(entry.getValue()); } ((Counter)getCounter.invoke(context, HBASE_COUNTER_GROUP_NAME, "NUM_SCANNER_RESTARTS")).increment(numScannerRestarts); ((Counter)getCounter.invoke(context, HBASE_COUNTER_GROUP_NAME, "NUM_SCAN_RESULTS_STALE")).increment(numStale); } catch (Exception e) { LOG.debug("can\'t update counter." + StringUtils.stringifyException(e)); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceManager.java:426 try { this.executor.execute(transfer); } catch (RejectedExecutionException ex) { LOG.info("Cancelling the transfer of " + rsZnode + " because of " + ex.getMessage()); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceManager.java:593 try { ReplicationQueueInfo replicationQueueInfo = new ReplicationQueueInfo(peerId); String actualPeerId = replicationQueueInfo.getPeerId(); ReplicationPeer peer = replicationPeers.getPeer(actualPeerId); ReplicationPeerConfig peerConfig = null; try { peerConfig = replicationPeers.getReplicationPeerConfig(actualPeerId); } catch (ReplicationException ex) { LOG.warn("Received exception while getting replication peer config, skipping replay" + ex); } if (peer == null || peerConfig == null) { LOG.warn("Skipping failover for peer:" + actualPeerId + " of node" + rsZnode); continue; } ReplicationSourceInterface src = getReplicationSource(conf, fs, ReplicationSourceManager.this, this.rq, this.rp, stopper, peerId, this.clusterId, peerConfig, peer); if (!this.rp.getPeerIds().contains((src.getPeerClusterId()))) { src.terminate("Recovered queue doesn\'t belong to any current peer"); break; } oldsources.add(src); SortedSet hlogsSet = entry.getValue(); for (String hlog : hlogsSet) { src.enqueueLog(new Path(oldLogDir, hlog)); } src.startup(); hlogsByIdRecoveredQueues.put(peerId, hlogsSet); } catch (IOException e) { LOG.error("Failed creating a source", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/zookeeper/MiniZooKeeperCluster.java:375 try { Socket sock = new Socket("localhost", port); BufferedReader reader = null; try { OutputStream outstream = sock.getOutputStream(); outstream.write("stat".getBytes()); outstream.flush(); Reader isr = new InputStreamReader(sock.getInputStream()); reader = new BufferedReader(isr); String line = reader.readLine(); if (line != null && line.startsWith("Zookeeper version:")) { return true; } } finally { sock.close(); if (reader != null) { reader.close(); } } } catch (IOException e) { LOG.info("server localhost:" + port + " not up " + e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/FsDelegationToken.java:92 try { userToken.cancel(this.fs.getConf()); } catch (Exception e) { LOG.warn("Failed to cancel HDFS delegation token: " + userToken, e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSHDFSUtils.java:83 try { if (dfsUtilClazz == null) { dfsUtilClazz = Class.forName("org.apache.hadoop.hdfs.DFSUtil"); } if (getNNAddressesMethod == null) { getNNAddressesMethod = dfsUtilClazz.getMethod("getNNServiceRpcAddresses", Configuration.class); } Map> addressMap = (Map>)getNNAddressesMethod.invoke(null, conf); for (Map.Entry> entry : addressMap.entrySet()) { Map nnMap = entry.getValue(); for (Map.Entry e2 : nnMap.entrySet()) { InetSocketAddress addr = e2.getValue(); addresses.add(addr); } } } catch (Exception e) { LOG.warn("DFSUtil.getNNServiceRpcAddresses failed. serviceName=" + serviceName, e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormat.java:102 try { setHTable(new HTable(new Configuration(conf), tableName)); } catch (Exception e) { LOG.error(StringUtils.stringifyException(e)); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormat.java:158 try { scan = new Scan(); if (conf.get(SCAN_ROW_START) != null) { scan.setStartRow(Bytes.toBytes(conf.get(SCAN_ROW_START))); } if (conf.get(SCAN_ROW_STOP) != null) { scan.setStopRow(Bytes.toBytes(conf.get(SCAN_ROW_STOP))); } if (conf.get(SCAN_COLUMNS) != null) { addColumns(scan, conf.get(SCAN_COLUMNS)); } if (conf.get(SCAN_COLUMN_FAMILY) != null) { scan.addFamily(Bytes.toBytes(conf.get(SCAN_COLUMN_FAMILY))); } if (conf.get(SCAN_TIMESTAMP) != null) { scan.setTimeStamp(Long.parseLong(conf.get(SCAN_TIMESTAMP))); } if (conf.get(SCAN_TIMERANGE_START) != null && conf.get(SCAN_TIMERANGE_END) != null) { scan.setTimeRange(Long.parseLong(conf.get(SCAN_TIMERANGE_START)), Long.parseLong(conf.get(SCAN_TIMERANGE_END))); } if (conf.get(SCAN_MAXVERSIONS) != null) { scan.setMaxVersions(Integer.parseInt(conf.get(SCAN_MAXVERSIONS))); } if (conf.get(SCAN_CACHEDROWS) != null) { scan.setCaching(Integer.parseInt(conf.get(SCAN_CACHEDROWS))); } if (conf.get(SCAN_BATCHSIZE) != null) { scan.setBatch(Integer.parseInt(conf.get(SCAN_BATCHSIZE))); } scan.setCacheBlocks((conf.getBoolean(SCAN_CACHEBLOCKS, false))); } catch (Exception e) { LOG.error(StringUtils.stringifyException(e)); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.java:149 try { SnapshotFileCache.this.refreshCache(); } catch (IOException e) { LOG.warn("Failed to refresh snapshot hfile cache!", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/snapshot/SnapshotFileCache.java:291 try { SnapshotFileCache.this.refreshCache(); } catch (IOException e) { LOG.warn("Failed to refresh snapshot hfile cache!", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/TableEventHandler.java:157 try { this.tableLock.release(); } catch (IOException ex) { LOG.warn("Could not release the table lock", ex); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/metrics/PersistentMetricsTimeVaryingRate.java:81 try { mr.setMetric(getName() + "_num_ops", totalOps); } catch (Exception e) { LOG.info("pushMetric failed for " + getName() + "\n" + StringUtils.stringifyException(e)); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/TableAuthManager.java:120 try { this.zkperms.start(); } catch (KeeperException ke) { LOG.error("ZooKeeper initialization failed", ke); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/TableAuthManager.java:216 try { newCache = initGlobal(conf); for (Map.Entry entry : userPerms.entries()) { if (AccessControlLists.isGroupPrincipal(entry.getKey())) { newCache.putGroup(AccessControlLists.getGroupName(entry.getKey()), new Permission(entry.getValue().getActions())); } else { newCache.putUser(entry.getKey(), new Permission(entry.getValue().getActions())); } } globalCache = newCache; mtime++; } catch (IOException e) { LOG.error("Error occured while updating the global cache", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/ZKPermissionWatcher.java:110 try { byte[] data = ZKUtil.getDataAndWatch(watcher, path); refreshAuthManager(entry, data); } catch (KeeperException ke) { LOG.error("Error reading data from zookeeper for node " + entry, ke); watcher.abort("Zookeeper error getting data for node " + entry, ke); } catch (IOException ioe) { LOG.error("Error reading permissions writables", ioe); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/ZKPermissionWatcher.java:138 try { refreshAuthManager(entry, n.getData()); } catch (IOException ioe) { LOG.error("Failed parsing permissions for table \'" + entry + "\' from zk", ioe); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java:609 try { ByteArrayComparator comparator = new ByteArrayComparator(); List regions = MetaScanner.listAllRegions(getConf(), false); final RegionBoundariesInformation currentRegionBoundariesInformation = new RegionBoundariesInformation(); Path hbaseRoot = FSUtils.getRootDir(getConf()); for (HRegionInfo regionInfo : regions) { Path tableDir = FSUtils.getTableDir(hbaseRoot, regionInfo.getTable()); currentRegionBoundariesInformation.regionName = regionInfo.getRegionName(); Path path = new Path(tableDir, regionInfo.getEncodedName()); FileSystem fs = path.getFileSystem(getConf()); FileStatus[] files = fs.listStatus(path); byte[] storeFirstKey = null; byte[] storeLastKey = null; for (FileStatus file : files) { String fileName = file.getPath().toString(); fileName = fileName.substring(fileName.lastIndexOf("/") + 1); if (!fileName.startsWith(".") && !fileName.endsWith("recovered.edits")) { FileStatus[] storeFiles = fs.listStatus(file.getPath()); for (FileStatus storeFile : storeFiles) { HFile.Reader reader = HFile.createReader(fs, storeFile.getPath(), new CacheConfig(getConf()), getConf()); if ((reader.getFirstKey() != null) && ((storeFirstKey == null) || (comparator.compare(storeFirstKey, reader.getFirstKey()) > 0))) { storeFirstKey = reader.getFirstKey(); } if ((reader.getLastKey() != null) && ((storeLastKey == null) || (comparator.compare(storeLastKey, reader.getLastKey())) < 0)) { storeLastKey = reader.getLastKey(); } reader.close(); } } } currentRegionBoundariesInformation.metaFirstKey = regionInfo.getStartKey(); currentRegionBoundariesInformation.metaLastKey = regionInfo.getEndKey(); currentRegionBoundariesInformation.storesFirstKey = keyOnly(storeFirstKey); currentRegionBoundariesInformation.storesLastKey = keyOnly(storeLastKey); if (currentRegionBoundariesInformation.metaFirstKey.length == 0) currentRegionBoundariesInformation.metaFirstKey = null; if (currentRegionBoundariesInformation.metaLastKey.length == 0) currentRegionBoundariesInformation.metaLastKey = null; boolean valid = true; if ((currentRegionBoundariesInformation.storesFirstKey != null) && (currentRegionBoundariesInformation.metaFirstKey != null)) { valid = valid && comparator.compare(currentRegionBoundariesInformation.storesFirstKey, currentRegionBoundariesInformation.metaFirstKey) >= 0; } if ((currentRegionBoundariesInformation.storesLastKey != null) && (currentRegionBoundariesInformation.metaLastKey != null)) { valid = valid && comparator.compare(currentRegionBoundariesInformation.storesLastKey, currentRegionBoundariesInformation.metaLastKey) < 0; } if (!valid) { errors.reportError(ERROR_CODE.BOUNDARIES_ERROR, "Found issues with regions boundaries", tablesInfo.get(regionInfo.getTable())); LOG.warn("Region\'s boundaries not alligned between stores and META for:"); LOG.warn(currentRegionBoundariesInformation); } } } catch (IOException e) { LOG.error(e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java:924 try { f.get(); } catch (ExecutionException e) { LOG.warn("Failed to read .regioninfo file for region " + work.hbi.getRegionNameAsString(), e.getCause()); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java:1466 try { f.get(); } catch (ExecutionException e) { LOG.warn("Could not load region dir ", e.getCause()); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java:1553 try { f.get(); } catch (ExecutionException e) { LOG.warn("Could not process regionserver " + item.rsinfo.getHostAndPort(), e.getCause()); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java:1674 try { HBaseFsckRepair.closeRegionSilentlyAndWait(admin, rse.hsa, rse.hri); offline(rse.hri.getRegionName()); } catch (IOException ioe) { LOG.warn("Got exception when attempting to offline region " + Bytes.toString(rse.hri.getRegionName()), ioe); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java:2400 try { LOG.info("[" + thread + "] Offlining region: " + hi); offline(hi.getRegionName()); } catch (IOException ioe) { LOG.warn("[" + thread + "] Unable to offline region from master: " + hi + ". Just continuing... ", ioe); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java:2462 try { LOG.info("Offlining region: " + regionToSideline); offline(regionToSideline.getRegionName()); } catch (IOException ioe) { LOG.warn("Unable to offline region from master: " + regionToSideline + ". Just continuing... ", ioe); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java:249 try { ListMultimap perms = AccessControlLists.getPermissions(conf, entry); byte[] serialized = AccessControlLists.writePermissionsAsBytes(perms, conf); zkw.writeToZookeeper(entry, serialized); } catch (IOException ex) { LOG.error("Failed updating permissions mirror for \'" + Bytes.toString(entry) + "\'", ex); } Ding debug Ding debug WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java:2175 try { masterServices.checkTableModifiable(tableName); } catch (TableNotFoundException ex) { continue; } catch (TableNotDisabledException ex) { } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/ModifyTableHandler.java:131 try { Set newFamilies = this.htd.getFamiliesKeys(); MasterFileSystem mfs = this.masterServices.getMasterFileSystem(); for (byte[] familyName : oldFamilies) { if (!newFamilies.contains(familyName)) { LOG.debug("Removing family=" + Bytes.toString(familyName) + " from table=" + this.tableName); for (HRegionInfo hri : hris) { mfs.deleteFamilyFromFS(hri, familyName); } } } } catch (IOException e) { LOG.warn("Unable to remove on-disk directories for the removed families", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableInputFormat.java:62 try { setHTable(new HTable(HBaseConfiguration.create(job), tableNames[0].getName())); } catch (Exception e) { LOG.error(StringUtils.stringifyException(e)); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentVerificationReport.java:216 try { ServerName currentRS = currentAssignment.get(region); if (currentRS == null) { unAssignedRegionsList.add(region); continue; } Integer hostRegionCounter = serverToHostingRegionCounterMap.get(currentRS); if (hostRegionCounter == null) { hostRegionCounter = Integer.valueOf(0); } hostRegionCounter = hostRegionCounter.intValue() + 1; serverToHostingRegionCounterMap.put(currentRS, hostRegionCounter); List favoredNodes = favoredNodesAssignment.getFavoredNodes(region); if (favoredNodes == null || favoredNodes.size() != FavoredNodeAssignmentHelper.FAVORED_NODES_NUM) { regionsWithoutValidFavoredNodes.add(region); continue; } ServerName primaryRS = favoredNodes.get(FavoredNodesPlan.Position.PRIMARY.ordinal()); ServerName secondaryRS = favoredNodes.get(FavoredNodesPlan.Position.SECONDARY.ordinal()); ServerName tertiaryRS = favoredNodes.get(FavoredNodesPlan.Position.TERTIARY.ordinal()); Integer regionCounter = primaryRSToRegionCounterMap.get(primaryRS); if (regionCounter == null) { regionCounter = Integer.valueOf(0); } regionCounter = regionCounter.intValue() + 1; primaryRSToRegionCounterMap.put(primaryRS, regionCounter); Set secAndTerSet = primaryToSecTerRSMap.get(primaryRS); if (secAndTerSet == null) { secAndTerSet = new HashSet(); } secAndTerSet.add(secondaryRS); secAndTerSet.add(tertiaryRS); primaryToSecTerRSMap.put(primaryRS, secAndTerSet); FavoredNodesPlan.Position favoredNodePosition = FavoredNodesPlan.getFavoredServerPosition(favoredNodes, currentRS); if (favoredNodePosition == null) { nonFavoredAssignedRegionList.add(region); continue; } this.favoredNodes[favoredNodePosition.ordinal()]++; totalFavoredAssignments++; if (regionLocalityMap != null) { this.enforceLocality = true; Map regionDegreeLocalityMap = regionLocalityMap.get(region.getEncodedName()); if (regionDegreeLocalityMap == null) { continue; } for (FavoredNodesPlan.Position p : FavoredNodesPlan.Position.values()) { ServerName favoredNode = favoredNodes.get(p.ordinal()); Float locality = regionDegreeLocalityMap.get(favoredNode.getHostname()); if (locality != null) { this.favoredNodesLocalitySummary[p.ordinal()] += locality; } } Float actualLocality = regionDegreeLocalityMap.get(currentRS.getHostname()); if (actualLocality != null) { this.actualLocalitySummary += actualLocality; } } } catch (Exception e) { LOG.error("Cannot verify the region assignment for region " + ((region == null) ? " null " : region.getRegionNameAsString()) + "because of " + e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/AssignmentVerificationReport.java:383 try { List favoredNodes = plan.getFavoredNodes(region); if (favoredNodes == null || favoredNodes.size() != FavoredNodeAssignmentHelper.FAVORED_NODES_NUM) { regionsWithoutValidFavoredNodes.add(region); continue; } ServerName primaryRS = favoredNodes.get(FavoredNodesPlan.Position.PRIMARY.ordinal()); ServerName secondaryRS = favoredNodes.get(FavoredNodesPlan.Position.SECONDARY.ordinal()); ServerName tertiaryRS = favoredNodes.get(FavoredNodesPlan.Position.TERTIARY.ordinal()); Integer regionCounter = primaryRSToRegionCounterMap.get(primaryRS); if (regionCounter == null) { regionCounter = Integer.valueOf(0); } regionCounter = regionCounter.intValue() + 1; primaryRSToRegionCounterMap.put(primaryRS, regionCounter); Set secAndTerSet = primaryToSecTerRSMap.get(primaryRS); if (secAndTerSet == null) { secAndTerSet = new HashSet(); } secAndTerSet.add(secondaryRS); secAndTerSet.add(tertiaryRS); primaryToSecTerRSMap.put(primaryRS, secAndTerSet); } catch (Exception e) { LOG.error("Cannot verify the region assignment for region " + ((region == null) ? " null " : region.getRegionNameAsString()) + "because of " + e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/master/ReplicationLogCleaner.java:125 try { this.zkw = new ZooKeeperWatcher(conf, "replicationLogCleaner", null); this.replicationQueues = ReplicationFactory.getReplicationQueuesClient(zkw, conf, this); this.replicationQueues.init(); } catch (ReplicationException e) { LOG.error("Error while configuring " + this.getClass().getName(), e); } catch (IOException e) { LOG.error("Error while configuring " + this.getClass().getName(), e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/security/visibility/DefaultVisibilityLabelServiceImpl.java:86 try { StreamUtils.writeRawVInt32(dos, SYSTEM_LABEL_ORDINAL); } catch (IOException e) { } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/example/TableHFileArchiveTracker.java:85 try { addAndReWatchTable(path); } catch (KeeperException e) { LOG.warn("Couldn\'t read zookeeper data for table for path:" + path + ", not preserving a table.", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/example/TableHFileArchiveTracker.java:99 try { updateWatchedTables(); } catch (KeeperException e) { LOG.error("Failed to update tables to archive", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFilePrettyPrinter.java:227 try { processFile(fileName); } catch (IOException ex) { LOG.error("Error reading " + fileName, ex); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSRegionScanner.java:167 try { Map blockCountMap = new HashMap(); String tableName = regionPath.getParent().getName(); int totalBlkCount = 0; FileStatus[] cfList = fs.listStatus(regionPath); if (null == cfList) { return; } for (FileStatus cfStatus : cfList) { if (!cfStatus.isDirectory()) { continue; } if (cfStatus.getPath().getName().startsWith(".") || HConstants.HBASE_NON_USER_TABLE_DIRS.contains(cfStatus.getPath().getName())) { continue; } FileStatus[] storeFileLists = fs.listStatus(cfStatus.getPath()); if (null == storeFileLists) { continue; } for (FileStatus storeFile : storeFileLists) { BlockLocation[] blkLocations = fs.getFileBlockLocations(storeFile, 0, storeFile.getLen()); if (null == blkLocations) { continue; } totalBlkCount += blkLocations.length; for (BlockLocation blk : blkLocations) { for (String host : blk.getHosts()) { AtomicInteger count = blockCountMap.get(host); if (count == null) { count = new AtomicInteger(0); blockCountMap.put(host, count); } count.incrementAndGet(); } } } } if (regionToBestLocalityRSMapping != null) { int largestBlkCount = 0; String hostToRun = null; for (Map.Entry entry : blockCountMap.entrySet()) { String host = entry.getKey(); int tmp = entry.getValue().get(); if (tmp > largestBlkCount) { largestBlkCount = tmp; hostToRun = host; } } if (null == hostToRun) { return; } if (hostToRun.endsWith(".")) { hostToRun = hostToRun.substring(0, hostToRun.length() - 1); } String name = tableName + ":" + regionPath.getName(); synchronized (regionToBestLocalityRSMapping) { regionToBestLocalityRSMapping.put(name, hostToRun); } } if (regionDegreeLocalityMapping != null && totalBlkCount > 0) { Map hostLocalityMap = new HashMap(); for (Map.Entry entry : blockCountMap.entrySet()) { String host = entry.getKey(); if (host.endsWith(".")) { host = host.substring(0, host.length() - 1); } float locality = ((float)entry.getValue().get()) / totalBlkCount; hostLocalityMap.put(host, locality); } regionDegreeLocalityMapping.put(regionPath.getName(), hostLocalityMap); } } catch (IOException e) { LOG.warn("Problem scanning file system", e); } catch (RuntimeException e) { LOG.warn("Problem scanning file system", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/DisableTableHandler.java:152 try { this.tableLock.release(); } catch (IOException ex) { LOG.warn("Could not release the table lock", ex); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/DeleteTableHandler.java:168 try { masterServices.getTableLockManager().tableDeleted(tableName); } catch (IOException ex) { LOG.warn("Received exception from TableLockManager.tableDeleted:", ex); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/BulkReOpen.java:72 try { unassign(hris); } catch (Throwable t) { LOG.warn("Failed bulking re-open " + hris.size() + " region(s)", t); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.java:807 try { future.get(); } catch (Exception e) { LOG.warn("Got exception:" + e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java:487 try { if (!this.targetTableSet.isEmpty() && !this.targetTableSet.contains(table)) { continue; } genAssignmentPlan(table, assignmentSnapshot, regionLocalityMap, plan, USE_MUNKRES_FOR_PLACING_SECONDARY_AND_TERTIARY); } catch (Exception e) { LOG.error("Get some exceptions for placing primary region server" + "for table " + table + " because " + e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java:646 try { LOG.info("Start to update the hbase:meta with the new assignment plan"); Map> assignmentMap = plan.getAssignmentMap(); FavoredNodeAssignmentHelper.updateMetaWithFavoredNodesInfo(assignmentMap, conf); LOG.info("Updated the hbase:meta with the new assignment plan"); } catch (Exception e) { LOG.error("Failed to update hbase:meta with the new assignment" + "plan because " + e.getMessage()); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/EnableTableHandler.java:174 try { this.tableLock.release(); } catch (IOException ex) { LOG.warn("Could not release the table lock", ex); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/JMXListener.java:142 try { jmxCS = JMXConnectorServerFactory.newJMXConnectorServer(serviceUrl, jmxEnv, mbs); jmxCS.start(); LOG.info("ConnectorServer started!"); } catch (IOException e) { LOG.error("fail to start connector server!", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/metrics/MetricsRate.java:80 try { mr.setMetric(getName(), getPreviousIntervalValue()); } catch (Exception e) { LOG.info("pushMetric failed for " + getName() + "\n" + StringUtils.stringifyException(e)); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java:415 try { Threads.threadDumpingIsAlive(t); } catch (InterruptedException e) { LOG.debug("Interrupted", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java:426 try { Threads.threadDumpingIsAlive(t); } catch (InterruptedException e) { LOG.debug("Interrupted", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/backup/example/LongTermArchivingHFileCleaner.java:92 try { this.fs = FileSystem.get(conf); this.archiveTracker = TableHFileArchiveTracker.create(conf); this.archiveTracker.start(); } catch (KeeperException e) { LOG.error("Error while configuring " + this.getClass().getName(), e); } catch (IOException e) { LOG.error("Error while configuring " + this.getClass().getName(), e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/http/HtmlQuoting.java:110 try { quoteHtmlChars(buffer, bytes, 0, bytes.length); } catch (IOException ioe) { } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/Export.java:125 try { s.setBatch(batching); } catch (IncompatibleFilterException e) { LOG.error("Batching could not be set", e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/CreateTableHandler.java:168 try { assignmentManager.getTableStateManager().checkAndRemoveTableState(tableName, ZooKeeperProtos.Table.State.ENABLING, false); } catch (CoordinatedStateException e) { LOG.error("Got a keeper exception while removing the ENABLING table znode " + tableName, e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/CreateTableHandler.java:307 try { this.tableLock.release(); } catch (IOException ex) { LOG.warn("Could not release the table lock", ex); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MultithreadedTableMapper.java:294 try { mapper.run(subcontext); } catch (Throwable ie) { LOG.error("Problem in running map.", ie); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKSplitLog.java:114 try { fs.createNewFile(file); } catch (IOException e) { LOG.warn("Could not flag a log file as corrupted. Failed to create " + file, e); } WARNING: empty catch: /Users/ding/Research/software/hbase/hbase-0.99.0/hbase-server/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKSplitLog.java:219 try { byte[] data; try { data = ZKUtil.getData(zkw, nodePath); } catch (InterruptedException e) { throw new InterruptedIOException(); } if (data != null) { result = ZKUtil.parseRegionStoreSequenceIds(data); } } catch (KeeperException e) { throw new IOException("Cannot get lastFlushedSequenceId from ZooKeeper for server=" + serverName + "; region=" + encodedRegionName, e); } catch (DeserializationException e) { LOG.warn("Can\'t parse last flushed sequence Id from znode:" + nodePath, e); }