Error Message java.io.IOException: Unknown Job job_1487884137933_0001 at org.apache.hadoop.mapreduce.v2.hs.HistoryClientService$HSClientProtocolHandler.verifyAndGetJob(HistoryClientService.java:235) at org.apache.hadoop.mapreduce.v2.hs.HistoryClientService$HSClientProtocolHandler.getCounters(HistoryClientService.java:249) at org.apache.hadoop.mapreduce.v2.api.impl.pb.service.MRClientProtocolPBServiceImpl.getCounters(MRClientProtocolPBServiceImpl.java:159) at org.apache.hadoop.yarn.proto.MRClientProtocol$MRClientProtocolService$2.callBlockingMethod(MRClientProtocol.java:281) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:640) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2351) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2347) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1866) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2345) Stacktrace java.io.IOException: java.io.IOException: Unknown Job job_1487884137933_0001 at org.apache.hadoop.mapreduce.v2.hs.HistoryClientService$HSClientProtocolHandler.verifyAndGetJob(HistoryClientService.java:235) at org.apache.hadoop.mapreduce.v2.hs.HistoryClientService$HSClientProtocolHandler.getCounters(HistoryClientService.java:249) at org.apache.hadoop.mapreduce.v2.api.impl.pb.service.MRClientProtocolPBServiceImpl.getCounters(MRClientProtocolPBServiceImpl.java:159) at org.apache.hadoop.yarn.proto.MRClientProtocol$MRClientProtocolService$2.callBlockingMethod(MRClientProtocol.java:281) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:640) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2351) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2347) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1866) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2345) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1554) at org.apache.hadoop.ipc.Client.call(Client.java:1498) at org.apache.hadoop.ipc.Client.call(Client.java:1398) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:233) at com.sun.proxy.$Proxy106.getCounters(Unknown Source) at org.apache.hadoop.mapreduce.v2.api.impl.pb.client.MRClientProtocolPBClientImpl.getCounters(MRClientProtocolPBClientImpl.java:166) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:601) at org.apache.hadoop.mapred.ClientServiceDelegate.invoke(ClientServiceDelegate.java:325) at org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounters(ClientServiceDelegate.java:381) at org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunner.java:598) at org.apache.hadoop.mapreduce.Job$7.run(Job.java:761) at org.apache.hadoop.mapreduce.Job$7.run(Job.java:758) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1866) at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:758) at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1383) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1311) at org.apache.hadoop.hbase.mapreduce.TestMultithreadedTableMapper.runTestOnTable(TestMultithreadedTableMapper.java:158) at org.apache.hadoop.hbase.mapreduce.TestMultithreadedTableMapper.testMultithreadedTableMapper(TestMultithreadedTableMapper.java:134) Standard Output Formatting using clusterid: testClusterID Standard Error 2017-02-23 21:08:24,847 INFO [main] hbase.HBaseTestingUtility(1003): Starting up minicluster with 1 master(s) and 1 regionserver(s) and 1 datanode(s) 2017-02-23 21:08:24,875 INFO [main] hbase.HBaseTestingUtility(471): Created new mini-cluster data directory: /grid/0/nobody/workspace/build-support/SOURCES/hbase/hbase-server/target/test-data/755970a9-678d-47b5-8be6-c4196ca5ad78/dfscluster_c8bb8278-f8f7-409c-958d-33bfad21b06f, deleteOnExit=true 2017-02-23 21:08:24,875 INFO [main] hbase.HBaseTestingUtility(718): Setting test.cache.data to /grid/0/nobody/workspace/build-support/SOURCES/hbase/hbase-server/target/test-data/755970a9-678d-47b5-8be6-c4196ca5ad78/cache_data in system properties and HBase conf 2017-02-23 21:08:24,876 INFO [main] hbase.HBaseTestingUtility(718): Setting hadoop.tmp.dir to /grid/0/nobody/workspace/build-support/SOURCES/hbase/hbase-server/target/test-data/755970a9-678d-47b5-8be6-c4196ca5ad78/hadoop_tmp in system properties and HBase conf 2017-02-23 21:08:24,876 INFO [main] hbase.HBaseTestingUtility(718): Setting hadoop.log.dir to /grid/0/nobody/workspace/build-support/SOURCES/hbase/hbase-server/target/test-data/755970a9-678d-47b5-8be6-c4196ca5ad78/hadoop_logs in system properties and HBase conf 2017-02-23 21:08:24,877 INFO [main] hbase.HBaseTestingUtility(718): Setting mapreduce.cluster.local.dir to /grid/0/nobody/workspace/build-support/SOURCES/hbase/hbase-server/target/test-data/755970a9-678d-47b5-8be6-c4196ca5ad78/mapred_local in system properties and HBase conf 2017-02-23 21:08:24,878 INFO [main] hbase.HBaseTestingUtility(718): Setting mapreduce.cluster.temp.dir to /grid/0/nobody/workspace/build-support/SOURCES/hbase/hbase-server/target/test-data/755970a9-678d-47b5-8be6-c4196ca5ad78/mapred_temp in system properties and HBase conf 2017-02-23 21:08:24,878 INFO [main] hbase.HBaseTestingUtility(709): read short circuit is OFF 2017-02-23 21:08:25,020 WARN [main] util.NativeCodeLoader(62): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2017-02-23 21:08:25,515 DEBUG [main] fs.HFileSystem(225): The file system is not a DistributedFileSystem. Skipping on block location reordering 2017-02-23 21:08:26,373 WARN [main] impl.MetricsConfig(125): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2017-02-23 21:08:26,552 INFO [main] log.Slf4jLog(67): Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog 2017-02-23 21:08:26,621 INFO [main] log.Slf4jLog(67): jetty-6.1.26.hwx 2017-02-23 21:08:26,658 INFO [main] log.Slf4jLog(67): Extract jar:file:/grid/0/nobody/.m2/repository/org/apache/hadoop/hadoop-hdfs/2.7.3.2.6.0.0-SNAPSHOT/hadoop-hdfs-2.7.3.2.6.0.0-SNAPSHOT-tests.jar!/webapps/hdfs to /tmp/Jetty_localhost_42318_hdfs____1en6rq/webapp 2017-02-23 21:08:26,878 INFO [main] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:42318 2017-02-23 21:08:27,453 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 0, pendingReplications = 0. 2017-02-23 21:08:27,771 INFO [main] log.Slf4jLog(67): jetty-6.1.26.hwx 2017-02-23 21:08:27,774 INFO [main] log.Slf4jLog(67): Extract jar:file:/grid/0/nobody/.m2/repository/org/apache/hadoop/hadoop-hdfs/2.7.3.2.6.0.0-SNAPSHOT/hadoop-hdfs-2.7.3.2.6.0.0-SNAPSHOT-tests.jar!/webapps/datanode to /tmp/Jetty_localhost_34126_datanode____741997/webapp 2017-02-23 21:08:27,907 INFO [main] log.Slf4jLog(67): Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:34126 2017-02-23 21:08:29,284 INFO [Block report processor] blockmanagement.BlockManager(1978): BLOCK* processReport: from storage DS-ed1c569b-3206-458d-96a8-b1d0865c3068 node DatanodeRegistration(127.0.0.1:49360, datanodeUuid=8f4c05a6-910a-4bec-b882-ab3d5baeb877, infoPort=60252, infoSecurePort=0, ipcPort=47681, storageInfo=lv=-56;cid=testClusterID;nsid=1474348591;c=0), blocks: 0, hasStaleStorage: true, processing time: 3 msecs, invalidatedBlocks: 0 2017-02-23 21:08:29,285 INFO [Block report processor] blockmanagement.BlockManager(1978): BLOCK* processReport: from storage DS-2df1d367-df7c-42de-a4dd-3408850f4212 node DatanodeRegistration(127.0.0.1:49360, datanodeUuid=8f4c05a6-910a-4bec-b882-ab3d5baeb877, infoPort=60252, infoSecurePort=0, ipcPort=47681, storageInfo=lv=-56;cid=testClusterID;nsid=1474348591;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2017-02-23 21:08:29,471 INFO [main] zookeeper.MiniZooKeeperCluster(276): Started MiniZooKeeperCluster and ran successful 'stat' on client port=50332 2017-02-23 21:08:29,484 INFO [main] fs.HFileSystem(256): Added intercepting call to namenode#getBlockLocations so can do block reordering using class class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2017-02-23 21:08:29,488 INFO [main] fs.HFileSystem(256): Added intercepting call to namenode#getBlockLocations so can do block reordering using class class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2017-02-23 21:08:30,216 INFO [main] util.FSUtils(763): Created version file at hdfs://localhost:56543/user/nobody/test-data/2bead9b6-c8b1-477c-8f16-5f539ce27690 with version=8 2017-02-23 21:08:30,292 DEBUG [main] impl.BackupManager(160): Added region procedure manager: org.apache.hadoop.hbase.backup.regionserver.LogRollRegionServerProcedureManager 2017-02-23 21:08:30,456 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 0, pendingReplications = 0. 2017-02-23 21:08:30,481 INFO [main] client.ConnectionUtils(104): master/xxx/xxx:0 server-side HConnection retries=350 2017-02-23 21:08:30,493 INFO [main] ipc.SimpleRpcScheduler(196): Using fifo as user call queue, count=1 2017-02-23 21:08:30,502 INFO [main] ipc.RpcServer$Listener(608): master/xxx/xxx:0: started 10 reader(s). 2017-02-23 21:08:30,541 INFO [main] hfile.CacheConfig(530): Allocating LruBlockCache size=995.58 MB, blockSize=64 KB 2017-02-23 21:08:30,547 DEBUG [main] hfile.CacheConfig(544): Trying to use Internal l2 cache 2017-02-23 21:08:30,547 INFO [main] hfile.CacheConfig(278): blockCache=LruBlockCache{blockCount=0, currentSize=1071760, freeSize=1042864304, maxSize=1043936064, heapSize=1071760, minSize=991739264, minFactor=0.95, multiSize=495869632, multiFactor=0.5, singleSize=247934816, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false 2017-02-23 21:08:30,548 INFO [main] hfile.CacheConfig(278): blockCache=LruBlockCache{blockCount=0, currentSize=1071760, freeSize=1042864304, maxSize=1043936064, heapSize=1071760, minSize=991739264, minFactor=0.95, multiSize=495869632, multiFactor=0.5, singleSize=247934816, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false 2017-02-23 21:08:30,550 INFO [main] mob.MobFileCache(121): MobFileCache is initialized, and the cache size is 1000 2017-02-23 21:08:30,553 INFO [main] fs.HFileSystem(256): Added intercepting call to namenode#getBlockLocations so can do block reordering using class class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2017-02-23 21:08:30,559 INFO [main] fs.HFileSystem(256): Added intercepting call to namenode#getBlockLocations so can do block reordering using class class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2017-02-23 21:08:30,615 INFO [main] zookeeper.RecoverableZooKeeper(120): Process identifier=master:45182 connecting to ZooKeeper ensemble=localhost:50332 2017-02-23 21:08:30,672 DEBUG [main-EventThread] zookeeper.ZooKeeperWatcher(515): master:451820x0, quorum=localhost:50332, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2017-02-23 21:08:30,673 DEBUG [main-EventThread] zookeeper.ZooKeeperWatcher(578): master:45182-0x15a6ccdc6840000 connected 2017-02-23 21:08:30,883 DEBUG [main] zookeeper.ZKUtil(369): master:45182-0x15a6ccdc6840000, quorum=localhost:50332, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2017-02-23 21:08:30,884 DEBUG [main] zookeeper.ZKUtil(369): master:45182-0x15a6ccdc6840000, quorum=localhost:50332, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2017-02-23 21:08:30,886 INFO [RpcServer.listener,port=45182] ipc.RpcServer$Listener(739): RpcServer.listener,port=45182: starting 2017-02-23 21:08:30,887 INFO [RpcServer.responder] ipc.RpcServer$Responder(927): RpcServer.responder: starting 2017-02-23 21:08:30,887 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.default.handler=0,queue=0,port=45182 2017-02-23 21:08:30,888 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.default.handler=1,queue=0,port=45182 2017-02-23 21:08:30,888 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.default.handler=2,queue=0,port=45182 2017-02-23 21:08:30,888 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.default.handler=3,queue=0,port=45182 2017-02-23 21:08:30,888 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.default.handler=4,queue=0,port=45182 2017-02-23 21:08:30,889 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.default.handler=5,queue=0,port=45182 2017-02-23 21:08:30,889 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=0,queue=0,port=45182 2017-02-23 21:08:30,889 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=1,queue=1,port=45182 2017-02-23 21:08:30,889 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=2,queue=0,port=45182 2017-02-23 21:08:30,890 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=3,queue=1,port=45182 2017-02-23 21:08:30,890 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=4,queue=0,port=45182 2017-02-23 21:08:30,890 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=5,queue=1,port=45182 2017-02-23 21:08:30,891 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=6,queue=0,port=45182 2017-02-23 21:08:30,891 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=7,queue=1,port=45182 2017-02-23 21:08:30,891 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=8,queue=0,port=45182 2017-02-23 21:08:30,891 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=9,queue=1,port=45182 2017-02-23 21:08:30,892 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=10,queue=0,port=45182 2017-02-23 21:08:30,892 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=11,queue=1,port=45182 2017-02-23 21:08:30,892 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=12,queue=0,port=45182 2017-02-23 21:08:30,892 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=13,queue=1,port=45182 2017-02-23 21:08:30,892 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=14,queue=0,port=45182 2017-02-23 21:08:30,893 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=15,queue=1,port=45182 2017-02-23 21:08:30,893 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=16,queue=0,port=45182 2017-02-23 21:08:30,893 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=17,queue=1,port=45182 2017-02-23 21:08:30,893 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=18,queue=0,port=45182 2017-02-23 21:08:30,894 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=19,queue=1,port=45182 2017-02-23 21:08:30,894 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.replication.handler=0,queue=0,port=45182 2017-02-23 21:08:30,894 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.replication.handler=1,queue=0,port=45182 2017-02-23 21:08:30,894 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.replication.handler=2,queue=0,port=45182 2017-02-23 21:08:30,898 INFO [main] master.HMaster(429): hbase.rootdir=hdfs://localhost:56543/user/nobody/test-data/2bead9b6-c8b1-477c-8f16-5f539ce27690, hbase.cluster.distributed=false 2017-02-23 21:08:30,904 DEBUG [main] impl.BackupManager(136): Added log cleaner: org.apache.hadoop.hbase.backup.master.BackupLogCleaner 2017-02-23 21:08:30,904 DEBUG [main] impl.BackupManager(137): Added master procedure manager: org.apache.hadoop.hbase.backup.master.LogRollMasterProcedureManager 2017-02-23 21:08:30,904 DEBUG [main] impl.BackupManager(138): Added master observer: org.apache.hadoop.hbase.backup.master.BackupController 2017-02-23 21:08:30,912 INFO [main] master.HMaster(1897): Adding backup master ZNode /hbase/backup-masters/xxx,45182,1487884110529 2017-02-23 21:08:30,932 DEBUG [main] zookeeper.ZKUtil(367): master:45182-0x15a6ccdc6840000, quorum=localhost:50332, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/xxx,45182,1487884110529 2017-02-23 21:08:30,964 DEBUG [main-EventThread] zookeeper.ZooKeeperWatcher(515): master:45182-0x15a6ccdc6840000, quorum=localhost:50332, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2017-02-23 21:08:30,965 DEBUG [xxx:45182.activeMasterManager] zookeeper.ZKUtil(367): master:45182-0x15a6ccdc6840000, quorum=localhost:50332, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2017-02-23 21:08:30,966 INFO [xxx:45182.activeMasterManager] master.ActiveMasterManager(170): Deleting ZNode for /hbase/backup-masters/xxx,45182,1487884110529 from backup master directory 2017-02-23 21:08:30,968 DEBUG [main-EventThread] zookeeper.ZKUtil(367): master:45182-0x15a6ccdc6840000, quorum=localhost:50332, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2017-02-23 21:08:30,969 DEBUG [main-EventThread] master.ActiveMasterManager(126): A master is now available 2017-02-23 21:08:30,980 DEBUG [main] impl.BackupManager(160): Added region procedure manager: org.apache.hadoop.hbase.backup.regionserver.LogRollRegionServerProcedureManager 2017-02-23 21:08:30,981 INFO [main] client.ConnectionUtils(104): regionserver/xxx/xxx:0 server-side HConnection retries=350 2017-02-23 21:08:30,982 INFO [main] ipc.SimpleRpcScheduler(196): Using fifo as user call queue, count=1 2017-02-23 21:08:30,985 INFO [main] ipc.RpcServer$Listener(608): regionserver/xxx/xxx:0: started 10 reader(s). 2017-02-23 21:08:30,990 DEBUG [main-EventThread] zookeeper.ZooKeeperWatcher(515): master:45182-0x15a6ccdc6840000, quorum=localhost:50332, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/xxx,45182,1487884110529 2017-02-23 21:08:30,991 WARN [xxx:45182.activeMasterManager] hbase.ZNodeClearer(58): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2017-02-23 21:08:30,992 INFO [xxx:45182.activeMasterManager] master.ActiveMasterManager(179): Registered Active Master=xxx,45182,1487884110529 2017-02-23 21:08:30,993 INFO [main] hfile.CacheConfig(278): blockCache=LruBlockCache{blockCount=0, currentSize=1071760, freeSize=1042864304, maxSize=1043936064, heapSize=1071760, minSize=991739264, minFactor=0.95, multiSize=495869632, multiFactor=0.5, singleSize=247934816, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false 2017-02-23 21:08:30,993 INFO [main] hfile.CacheConfig(278): blockCache=LruBlockCache{blockCount=0, currentSize=1071760, freeSize=1042864304, maxSize=1043936064, heapSize=1071760, minSize=991739264, minFactor=0.95, multiSize=495869632, multiFactor=0.5, singleSize=247934816, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false 2017-02-23 21:08:30,998 INFO [main] fs.HFileSystem(256): Added intercepting call to namenode#getBlockLocations so can do block reordering using class class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2017-02-23 21:08:31,004 INFO [main] fs.HFileSystem(256): Added intercepting call to namenode#getBlockLocations so can do block reordering using class class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2017-02-23 21:08:31,009 INFO [main] zookeeper.RecoverableZooKeeper(120): Process identifier=regionserver:55713 connecting to ZooKeeper ensemble=localhost:50332 2017-02-23 21:08:31,033 DEBUG [main-EventThread] zookeeper.ZooKeeperWatcher(515): regionserver:557130x0, quorum=localhost:50332, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2017-02-23 21:08:31,033 DEBUG [main-EventThread] zookeeper.ZooKeeperWatcher(578): regionserver:55713-0x15a6ccdc6840001 connected 2017-02-23 21:08:31,034 DEBUG [main] zookeeper.ZKUtil(367): regionserver:55713-0x15a6ccdc6840001, quorum=localhost:50332, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2017-02-23 21:08:31,035 DEBUG [main] zookeeper.ZKUtil(369): regionserver:55713-0x15a6ccdc6840001, quorum=localhost:50332, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2017-02-23 21:08:31,036 INFO [RpcServer.responder] ipc.RpcServer$Responder(927): RpcServer.responder: starting 2017-02-23 21:08:31,036 INFO [RpcServer.listener,port=55713] ipc.RpcServer$Listener(739): RpcServer.listener,port=55713: starting 2017-02-23 21:08:31,036 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.default.handler=0,queue=0,port=55713 2017-02-23 21:08:31,037 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.default.handler=1,queue=0,port=55713 2017-02-23 21:08:31,037 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.default.handler=2,queue=0,port=55713 2017-02-23 21:08:31,037 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.default.handler=3,queue=0,port=55713 2017-02-23 21:08:31,037 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.default.handler=4,queue=0,port=55713 2017-02-23 21:08:31,038 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.default.handler=5,queue=0,port=55713 2017-02-23 21:08:31,038 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=0,queue=0,port=55713 2017-02-23 21:08:31,038 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=1,queue=1,port=55713 2017-02-23 21:08:31,039 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=2,queue=0,port=55713 2017-02-23 21:08:31,039 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=3,queue=1,port=55713 2017-02-23 21:08:31,039 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=4,queue=0,port=55713 2017-02-23 21:08:31,040 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=5,queue=1,port=55713 2017-02-23 21:08:31,040 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=6,queue=0,port=55713 2017-02-23 21:08:31,040 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=7,queue=1,port=55713 2017-02-23 21:08:31,040 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=8,queue=0,port=55713 2017-02-23 21:08:31,041 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=9,queue=1,port=55713 2017-02-23 21:08:31,041 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=10,queue=0,port=55713 2017-02-23 21:08:31,041 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=11,queue=1,port=55713 2017-02-23 21:08:31,042 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=12,queue=0,port=55713 2017-02-23 21:08:31,042 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=13,queue=1,port=55713 2017-02-23 21:08:31,042 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=14,queue=0,port=55713 2017-02-23 21:08:31,042 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=15,queue=1,port=55713 2017-02-23 21:08:31,043 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=16,queue=0,port=55713 2017-02-23 21:08:31,043 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=17,queue=1,port=55713 2017-02-23 21:08:31,043 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=18,queue=0,port=55713 2017-02-23 21:08:31,044 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.priority.handler=19,queue=1,port=55713 2017-02-23 21:08:31,044 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.replication.handler=0,queue=0,port=55713 2017-02-23 21:08:31,044 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.replication.handler=1,queue=0,port=55713 2017-02-23 21:08:31,045 DEBUG [main] ipc.RpcExecutor(129): Started RpcServer.FifoWFPBQ.replication.handler=2,queue=0,port=55713 2017-02-23 21:08:31,051 INFO [main] http.HttpRequestLog(69): Http request log for http.requests.regionserver is not defined 2017-02-23 21:08:31,053 INFO [main] http.HttpServer(852): Added global filter 'safety' (class=org.apache.hadoop.hbase.http.HttpServer$QuotingInputFilter) 2017-02-23 21:08:31,055 INFO [main] http.HttpServer(830): Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context regionserver 2017-02-23 21:08:31,056 INFO [main] http.HttpServer(837): Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context static 2017-02-23 21:08:31,056 INFO [main] http.HttpServer(837): Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs 2017-02-23 21:08:31,060 INFO [main] http.HttpServer(1083): Jetty bound to port 48260 2017-02-23 21:08:31,060 INFO [main] log.Slf4jLog(67): jetty-6.1.26.hwx 2017-02-23 21:08:31,094 DEBUG [xxx:45182.activeMasterManager] util.FSUtils(915): Created cluster ID file at hdfs://localhost:56543/user/nobody/test-data/2bead9b6-c8b1-477c-8f16-5f539ce27690/hbase.id with ID: 5f31e442-e547-44f7-9b8b-f35f0359ffd4 2017-02-23 21:08:31,119 INFO [main] log.Slf4jLog(67): Started SelectChannelConnector@0.0.0.0:48260 2017-02-23 21:08:31,133 INFO [xxx:45182.activeMasterManager] master.MasterFileSystem(560): BOOTSTRAP: creating hbase:meta region 2017-02-23 21:08:31,136 INFO [xxx:45182.activeMasterManager] regionserver.HRegion(6360): creating HRegion hbase:meta HTD == 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|'}, {NAME => 'info', DATA_BLOCK_ENCODING => 'NONE', BLOOMFILTER => 'NONE', REPLICATION_SCOPE => '0', COMPRESSION => 'NONE', VERSIONS => '10', TTL => 'FOREVER', MIN_VERSIONS => '0', CACHE_DATA_IN_L1 => 'true', KEEP_DELETED_CELLS => 'FALSE', BLOCKSIZE => '8192', IN_MEMORY => 'false', BLOCKCACHE => 'false'} RootDir = hdfs://localhost:56543/user/nobody/test-data/2bead9b6-c8b1-477c-8f16-5f539ce27690 Table name == hbase:meta 2017-02-23 21:08:31,190 INFO [RS:0;xxx:55713] zookeeper.RecoverableZooKeeper(120): Process identifier=hconnection-0x6716cca4 connecting to ZooKeeper ensemble=localhost:50332 2017-02-23 21:08:31,190 INFO [M:0;xxx:45182] zookeeper.RecoverableZooKeeper(120): Process identifier=hconnection-0x9d557dc connecting to ZooKeeper ensemble=localhost:50332 2017-02-23 21:08:31,366 DEBUG [RS:0;xxx:55713-EventThread] zookeeper.ZooKeeperWatcher(515): hconnection-0x6716cca40x0, quorum=localhost:50332, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2017-02-23 21:08:31,368 DEBUG [RS:0;xxx:55713-EventThread] zookeeper.ZooKeeperWatcher(578): hconnection-0x6716cca4-0x15a6ccdc6840002 connected 2017-02-23 21:08:31,372 DEBUG [M:0;xxx:45182-EventThread] zookeeper.ZooKeeperWatcher(515): hconnection-0x9d557dc0x0, quorum=localhost:50332, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2017-02-23 21:08:31,372 INFO [RS:0;xxx:55713] client.ZooKeeperRegistry(107): ClusterId read in ZooKeeper is null 2017-02-23 21:08:31,374 INFO [M:0;xxx:45182] client.ZooKeeperRegistry(107): ClusterId read in ZooKeeper is null 2017-02-23 21:08:31,373 DEBUG [M:0;xxx:45182-EventThread] zookeeper.ZooKeeperWatcher(578): hconnection-0x9d557dc-0x15a6ccdc6840003 connected 2017-02-23 21:08:31,374 DEBUG [M:0;xxx:45182] client.ConnectionManager$HConnectionImplementation(892): clusterid came back null, using default default-cluster 2017-02-23 21:08:31,374 DEBUG [RS:0;xxx:55713] client.ConnectionManager$HConnectionImplementation(892): clusterid came back null, using default default-cluster 2017-02-23 21:08:31,381 DEBUG [M:0;xxx:45182] ipc.AbstractRpcClient(105): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@84de77b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2017-02-23 21:08:31,381 DEBUG [RS:0;xxx:55713] ipc.AbstractRpcClient(105): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3298407f, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2017-02-23 21:08:31,578 INFO [xxx:45182.activeMasterManager] wal.WALFactory(145): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider 2017-02-23 21:08:31,599 INFO [xxx:45182.activeMasterManager] wal.FSHLog(581): WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hregion-45480475.default, suffix=, logDir=hdfs://localhost:56543/user/nobody/test-data/2bead9b6-c8b1-477c-8f16-5f539ce27690/WALs/hregion-45480475, archiveDir=hdfs://localhost:56543/user/nobody/test-data/2bead9b6-c8b1-477c-8f16-5f539ce27690/oldWALs 2017-02-23 21:08:31,622 INFO [xxx:45182.activeMasterManager] wal.FSHLog(1620): Slow sync cost: 10 ms, current pipeline: [] 2017-02-23 21:08:31,622 INFO [xxx:45182.activeMasterManager] wal.FSHLog(1048): New WAL /user/nobody/test-data/2bead9b6-c8b1-477c-8f16-5f539ce27690/WALs/hregion-45480475/hregion-45480475.default.1487884111599 2017-02-23 21:08:31,658 DEBUG [xxx:45182.activeMasterManager] regionserver.HRegion(797): Instantiated hbase:meta,,1.1588230740 2017-02-23 21:08:31,702 INFO [StoreOpener-1588230740-1] hfile.CacheConfig(278): blockCache=LruBlockCache{blockCount=0, currentSize=1071760, freeSize=1042864304, maxSize=1043936064, heapSize=1071760, minSize=991739264, minFactor=0.95, multiSize=495869632, multiFactor=0.5, singleSize=247934816, singleFactor=0.25}, cacheDataOnRead=false, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false 2017-02-23 21:08:31,710 INFO [StoreOpener-1588230740-1] compactions.CompactionConfiguration(133): size [134217728, 9223372036854775807); files [3, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, base window in milliseconds 21600000, windows per tier 4,incoming window min 6 2017-02-23 21:08:31,715 DEBUG [StoreOpener-1588230740-1] regionserver.HRegionFileSystem(202): No StoreFiles for: hdfs://localhost:56543/user/nobody/test-data/2bead9b6-c8b1-477c-8f16-5f539ce27690/data/hbase/meta/1588230740/info 2017-02-23 21:08:31,718 DEBUG [StoreOpener-1588230740-1] util.ChecksumType$2(70): org.apache.hadoop.util.PureJavaCrc32 available 2017-02-23 21:08:31,718 DEBUG [StoreOpener-1588230740-1] util.ChecksumType$3(113): org.apache.hadoop.util.PureJavaCrc32C available 2017-02-23 21:08:31,727 DEBUG [xxx:45182.activeMasterManager] regionserver.HRegion(4092): Found 0 recovered edits file(s) under hdfs://localhost:56543/user/nobody/test-data/2bead9b6-c8b1-477c-8f16-5f539ce27690/data/hbase/meta/1588230740 2017-02-23 21:08:31,749 DEBUG [xxx:45182.activeMasterManager] wal.WALSplitter(720): Wrote region seqId=hdfs://localhost:56543/user/nobody/test-data/2bead9b6-c8b1-477c-8f16-5f539ce27690/data/hbase/meta/1588230740/recovered.edits/2.seqid to file, newSeqId=2, maxSeqId=0 2017-02-23 21:08:31,750 INFO [xxx:45182.activeMasterManager] regionserver.HRegion(924): Onlined 1588230740; next sequenceid=2 2017-02-23 21:08:31,750 DEBUG [xxx:45182.activeMasterManager] regionserver.HRegion(1453): Closing hbase:meta,,1.1588230740: disabling compactions & flushes 2017-02-23 21:08:31,750 DEBUG [xxx:45182.activeMasterManager] regionserver.HRegion(1480): Updates disabled for region hbase:meta,,1.1588230740 2017-02-23 21:08:31,751 INFO [StoreCloserThread-hbase:meta,,1.1588230740-1] regionserver.HStore(917): Closed info 2017-02-23 21:08:31,752 INFO [xxx:45182.activeMasterManager] regionserver.HRegion(1581): Closed hbase:meta,,1.1588230740 2017-02-23 21:08:31,752 DEBUG [xxx:45182.activeMasterManager] wal.FSHLog(1245): Closing WAL writer in /user/nobody/test-data/2bead9b6-c8b1-477c-8f16-5f539ce27690/WALs/hregion-45480475 2017-02-23 21:08:32,176 DEBUG [xxx:45182.activeMasterManager] wal.FSHLog(1203): Moved 1 WAL file(s) to /user/nobody/test-data/2bead9b6-c8b1-477c-8f16-5f539ce27690/oldWALs 2017-02-23 21:08:32,177 INFO [xxx:45182.activeMasterManager] wal.FSHLog(1206): Closed WAL: FSHLog hregion-45480475.default:(num 1487884111599) 2017-02-23 21:08:32,223 DEBUG [xxx:45182.activeMasterManager] util.FSTableDescriptors(660): Wrote descriptor into: hdfs://localhost:56543/user/nobody/test-data/2bead9b6-c8b1-477c-8f16-5f539ce27690/data/hbase/meta/.tabledesc/.tableinfo.0000000001 2017-02-23 21:08:32,248 INFO [xxx:45182.activeMasterManager] fs.HFileSystem(256): Added intercepting call to namenode#getBlockLocations so can do block reordering using class class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2017-02-23 21:08:32,252 DEBUG [xxx:45182.activeMasterManager] coordination.ZKSplitLogManagerCoordination(886): Distributed log replay=false 2017-02-23 21:08:32,256 INFO [xxx:45182.activeMasterManager] coordination.ZKSplitLogManagerCoordination(599): Found 0 orphan tasks and 0 rescan nodes 2017-02-23 21:08:32,257 DEBUG [xxx:45182.activeMasterManager] util.FSTableDescriptors(208): Fetching table descriptors from the filesystem. 2017-02-23 21:08:32,300 INFO [xxx:45182.activeMasterManager] zookeeper.RecoverableZooKeeper(120): Process identifier=hconnection-0x1cdd604 connecting to ZooKeeper ensemble=localhost:50332 2017-02-23 21:08:32,326 DEBUG [xxx:45182.activeMasterManager-EventThread] zookeeper.ZooKeeperWatcher(515): hconnection-0x1cdd6040x0, quorum=localhost:50332, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2017-02-23 21:08:32,326 DEBUG [xxx:45182.activeMasterManager-EventThread] zookeeper.ZooKeeperWatcher(578): hconnection-0x1cdd604-0x15a6ccdc6840004 connected 2017-02-23 21:08:32,328 DEBUG [xxx:45182.activeMasterManager] ipc.AbstractRpcClient(105): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@27b0f6a, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2017-02-23 21:08:32,343 INFO [xxx:45182.activeMasterManager] balancer.StochasticLoadBalancer(155): loading config 2017-02-23 21:08:32,353 DEBUG [xxx:45182.activeMasterManager] zookeeper.ZKUtil(369): master:45182-0x15a6ccdc6840000, quorum=localhost:50332, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/balancer 2017-02-23 21:08:32,355 DEBUG [xxx:45182.activeMasterManager] zookeeper.ZKUtil(369): master:45182-0x15a6ccdc6840000, quorum=localhost:50332, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/normalizer 2017-02-23 21:08:32,395 DEBUG [xxx:45182.activeMasterManager] zookeeper.ZKUtil(369): master:45182-0x15a6ccdc6840000, quorum=localhost:50332, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/switch/split 2017-02-23 21:08:32,396 DEBUG [xxx:45182.activeMasterManager] zookeeper.ZKUtil(369): master:45182-0x15a6ccdc6840000, quorum=localhost:50332, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/switch/merge 2017-02-23 21:08:32,436 DEBUG [main-EventThread] zookeeper.ZooKeeperWatcher(515): regionserver:55713-0x15a6ccdc6840001, quorum=localhost:50332, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2017-02-23 21:08:32,436 DEBUG [main-EventThread] zookeeper.ZooKeeperWatcher(515): master:45182-0x15a6ccdc6840000, quorum=localhost:50332, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2017-02-23 21:08:32,438 INFO [xxx:45182.activeMasterManager] master.HMaster(673): Server active/primary master=xxx,45182,1487884110529, sessionid=0x15a6ccdc6840000, setting cluster-up flag (Was=false) 2017-02-23 21:08:32,441 INFO [RS:0;xxx:55713] regionserver.HRegionServer(807): ClusterId : 5f31e442-e547-44f7-9b8b-f35f0359ffd4 2017-02-23 21:08:32,441 INFO [M:0;xxx:45182] regionserver.HRegionServer(807): ClusterId : 5f31e442-e547-44f7-9b8b-f35f0359ffd4 2017-02-23 21:08:32,444 INFO [RS:0;xxx:55713] procedure.ProcedureManagerHost(71): User procedure org.apache.hadoop.hbase.backup.regionserver.LogRollRegionServerProcedureManager was loaded successfully. 2017-02-23 21:08:32,447 DEBUG [RS:0;xxx:55713] procedure.RegionServerProcedureManagerHost(43): Procedure backup-proc is initializing 2017-02-23 21:08:32,448 INFO [xxx:45182.activeMasterManager] procedure.ProcedureManagerHost(71): User procedure org.apache.hadoop.hbase.backup.master.LogRollMasterProcedureManager was loaded successfully. 2017-02-23 21:08:32,645 DEBUG [RS:0;xxx:55713] procedure.RegionServerProcedureManagerHost(45): Procedure backup-proc is initialized 2017-02-23 21:08:32,646 DEBUG [RS:0;xxx:55713] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot is initializing 2017-02-23 21:08:32,667 INFO [xxx:45182.activeMasterManager] procedure.ZKProcedureUtil(270): Clearing all procedure znodes: /hbase/online-snapshot/acquired /hbase/online-snapshot/reached /hbase/online-snapshot/abort 2017-02-23 21:08:32,680 DEBUG [RS:0;xxx:55713] zookeeper.RecoverableZooKeeper(584): Node /hbase/online-snapshot/acquired already exists 2017-02-23 21:08:32,682 DEBUG [xxx:45182.activeMasterManager] procedure.ZKProcedureCoordinatorRpcs(238): Starting the controller for procedure member:xxx,45182,1487884110529 2017-02-23 21:08:32,682 DEBUG [RS:0;xxx:55713] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot is initialized 2017-02-23 21:08:32,682 DEBUG [RS:0;xxx:55713] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc is initializing 2017-02-23 21:08:32,718 DEBUG [xxx:45182.activeMasterManager] zookeeper.RecoverableZooKeeper(584): Node /hbase/rolllog-proc/acquired already exists 2017-02-23 21:08:32,754 INFO [xxx:45182.activeMasterManager] procedure.ZKProcedureUtil(270): Clearing all procedure znodes: /hbase/rolllog-proc/acquired /hbase/rolllog-proc/reached /hbase/rolllog-proc/abort 2017-02-23 21:08:32,779 DEBUG [xxx:45182.activeMasterManager] procedure.ZKProcedureCoordinatorRpcs(238): Starting the controller for procedure member:xxx,45182,1487884110529 2017-02-23 21:08:32,878 DEBUG [RS:0;xxx:55713] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc is initialized 2017-02-23 21:08:32,883 INFO [RS:0;xxx:55713] regionserver.MemStoreFlusher(125): globalMemStoreLimit=995.6 M, globalMemStoreLimitLowMark=945.8 M, maxHeap=2.4 G 2017-02-23 21:08:32,890 INFO [RS:0;xxx:55713] regionserver.HRegionServer$CompactionChecker(1633): CompactionChecker runs every 1sec 2017-02-23 21:08:32,908 DEBUG [RS:0;xxx:55713] ipc.AbstractRpcClient(105): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@173b2842, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=xxx/xxx:0 2017-02-23 21:08:32,912 DEBUG [RS:0;xxx:55713] regionserver.ShutdownHook(87): Installed shutdown hook thread: Shutdownhook:RS:0;xxx:55713 2017-02-23 21:08:32,932 DEBUG [xxx:45182.activeMasterManager] zookeeper.RecoverableZooKeeper(584): Node /hbase/flush-table-proc/acquired already exists 2017-02-23 21:08:32,948 DEBUG [main-EventThread] zookeeper.ZooKeeperWatcher(515): master:45182-0x15a6ccdc6840000, quorum=localhost:50332, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2017-02-23 21:08:32,949 DEBUG [RS:0;xxx:55713] zookeeper.ZKUtil(367): regionserver:55713-0x15a6ccdc6840001, quorum=localhost:50332, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/xxx,55713,1487884110992 2017-02-23 21:08:32,949 INFO [xxx:45182.activeMasterManager] procedure.ZKProcedureUtil(270): Clearing all procedure znodes: /hbase/flush-table-proc/acquired /hbase/flush-table-proc/reached /hbase/flush-table-proc/abort 2017-02-23 21:08:32,951 DEBUG [main-EventThread] zookeeper.ZKUtil(367): master:45182-0x15a6ccdc6840000, quorum=localhost:50332, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/xxx,55713,1487884110992 2017-02-23 21:08:32,953 DEBUG [main-EventThread] zookeeper.RegionServerTracker(93): Added tracking of RS /hbase/rs/xxx,55713,1487884110992 2017-02-23 21:08:32,953 DEBUG [xxx:45182.activeMasterManager] procedure.ZKProcedureCoordinatorRpcs(238): Starting the controller for procedure member:xxx,45182,1487884110529 2017-02-23 21:08:32,958 INFO [RS:0;xxx:55713] regionserver.RegionServerCoprocessorHost(66): System coprocessor loading is enabled 2017-02-23 21:08:32,958 INFO [RS:0;xxx:55713] regionserver.RegionServerCoprocessorHost(67): Table coprocessor loading is enabled 2017-02-23 21:08:32,961 INFO [RS:0;xxx:55713] regionserver.HRegionServer(2395): reportForDuty to master=xxx,45182,1487884110529 with port=55713, startcode=1487884110992 2017-02-23 21:08:32,987 INFO [xxx:45182.activeMasterManager] master.MasterCoprocessorHost(92): System coprocessor loading is enabled 2017-02-23 21:08:32,995 INFO [xxx:45182.activeMasterManager] coprocessor.CoprocessorHost(160): System coprocessor org.apache.hadoop.hbase.backup.master.BackupController was loaded successfully with priority (536870911). 2017-02-23 21:08:32,997 DEBUG [xxx:45182.activeMasterManager] executor.ExecutorService(100): Starting executor service name=MASTER_OPEN_REGION-xxx:45182, corePoolSize=5, maxPoolSize=5 2017-02-23 21:08:32,997 DEBUG [xxx:45182.activeMasterManager] executor.ExecutorService(100): Starting executor service name=MASTER_CLOSE_REGION-xxx:45182, corePoolSize=5, maxPoolSize=5 2017-02-23 21:08:32,998 DEBUG [xxx:45182.activeMasterManager] executor.ExecutorService(100): Starting executor service name=MASTER_SERVER_OPERATIONS-xxx:45182, corePoolSize=5, maxPoolSize=5 2017-02-23 21:08:32,998 DEBUG [xxx:45182.activeMasterManager] executor.ExecutorService(100): Starting executor service name=MASTER_META_SERVER_OPERATIONS-xxx:45182, corePoolSize=5, maxPoolSize=5 2017-02-23 21:08:32,998 DEBUG [xxx:45182.activeMasterManager] executor.ExecutorService(100): Starting executor service name=M_LOG_REPLAY_OPS-xxx:45182, corePoolSize=10, maxPoolSize=10 2017-02-23 21:08:32,998 DEBUG [xxx:45182.activeMasterManager] executor.ExecutorService(100): Starting executor service name=MASTER_TABLE_OPERATIONS-xxx:45182, corePoolSize=1, maxPoolSize=1 2017-02-23 21:08:33,015 INFO [xxx:45182.activeMasterManager] procedure2.ProcedureExecutor(436): Starting procedure executor threads=33 2017-02-23 21:08:33,017 INFO [xxx:45182.activeMasterManager] wal.WALProcedureStore(318): Starting WAL Procedure Store lease recovery 2017-02-23 21:08:33,019 WARN [xxx:45182.activeMasterManager] wal.WALProcedureStore(983): Log directory not found: File hdfs://localhost:56543/user/nobody/test-data/2bead9b6-c8b1-477c-8f16-5f539ce27690/MasterProcWALs does not exist. 2017-02-23 21:08:33,029 DEBUG [xxx:45182.activeMasterManager] wal.WALProcedureStore(861): Roll new state log: 1 2017-02-23 21:08:33,032 INFO [xxx:45182.activeMasterManager] wal.WALProcedureStore(347): Lease acquired for flushLogId: 1 2017-02-23 21:08:33,032 DEBUG [xxx:45182.activeMasterManager] wal.WALProcedureStore(364): No state logs to replay. 2017-02-23 21:08:33,041 DEBUG [ProcedureExecutorTimeout] procedure2.ProcedureExecutor$CompletedProcedureCleaner(162): No completed procedures to cleanup. 2017-02-23 21:08:33,042 DEBUG [xxx:45182.activeMasterManager] cleaner.CleanerChore(91): initialize cleaner=org.apache.hadoop.hbase.backup.master.BackupLogCleaner 2017-02-23 21:08:33,043 DEBUG [xxx:45182.activeMasterManager] cleaner.CleanerChore(91): initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2017-02-23 21:08:33,045 INFO [xxx:45182.activeMasterManager] zookeeper.RecoverableZooKeeper(120): Process identifier=replicationLogCleaner connecting to ZooKeeper ensemble=localhost:50332 2017-02-23 21:08:33,057 DEBUG [RS:0;xxx:55713] ipc.RpcClientImpl$Connection(345): Use SIMPLE authentication for service RegionServerStatusService, sasl=false 2017-02-23 21:08:33,072 DEBUG [xxx:45182.activeMasterManager-EventThread] zookeeper.ZooKeeperWatcher(515): replicationLogCleaner0x0, quorum=localhost:50332, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2017-02-23 21:08:33,072 DEBUG [xxx:45182.activeMasterManager-EventThread] zookeeper.ZooKeeperWatcher(578): replicationLogCleaner-0x15a6ccdc6840005 connected 2017-02-23 21:08:33,081 DEBUG [RS:0;xxx:55713] ipc.RpcClientImpl$Connection(710): Connecting to xxx/xxx:45182 2017-02-23 21:08:33,087 DEBUG [RpcServer.listener,port=45182] ipc.RpcServer$Listener(850): RpcServer.listener,port=45182: connection from xxx:44618; # active connections: 1 2017-02-23 21:08:33,095 INFO [RpcServer.reader=1,bindAddress=xxx,port=45182] ipc.RpcServer$Connection(1683): Connection from xxx port: 44618 with version info: version: "1.1.2.2.6.0.0-SNAPSHOT" url: "git://xxx/grid/0/nobody/workspace/build-support/SOURCES/hbase" revision: "87b4908be2ff0ee4c12025175a32057af7018bae" user: "nobody" date: "Thu Feb 23 20:49:35 UTC 2017" src_checksum: "dcba0c744a09baccb7e88330156a1cf6" 2017-02-23 21:08:33,106 DEBUG [RpcServer.FifoWFPBQ.priority.handler=19,queue=1,port=45182] ipc.CallRunner(115): RpcServer.FifoWFPBQ.priority.handler=19,queue=1,port=45182: callId: 0 service: RegionServerStatusService methodName: RegionServerStartup size: 45 connection: xxx:44618 org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:2626) at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:370) at org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:10444) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2141) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:187) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:167) 2017-02-23 21:08:33,116 DEBUG [xxx:45182.activeMasterManager] cleaner.CleanerChore(91): initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2017-02-23 21:08:33,120 DEBUG [RS:0;xxx:55713] regionserver.HRegionServer(2414): Master is not running yet 2017-02-23 21:08:33,120 DEBUG [xxx:45182.activeMasterManager] cleaner.CleanerChore(91): initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotLogCleaner 2017-02-23 21:08:33,120 WARN [RS:0;xxx:55713] regionserver.HRegionServer(938): reportForDuty failed; sleeping and then retrying. 2017-02-23 21:08:33,122 DEBUG [xxx:45182.activeMasterManager] cleaner.CleanerChore(91): initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2017-02-23 21:08:33,124 DEBUG [xxx:45182.activeMasterManager] cleaner.CleanerChore(91): initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2017-02-23 21:08:33,124 DEBUG [xxx:45182.activeMasterManager] cleaner.CleanerChore(91): initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2017-02-23 21:08:33,125 INFO [xxx ...[truncated 2845120 chars]... atedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:46:10,305 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:46:13,318 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:46:13,333 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:46:16,333 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:46:16,334 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:46:19,335 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:46:19,372 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:46:22,373 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:46:22,374 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:46:25,375 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:46:25,375 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:46:28,376 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:46:28,403 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:46:31,414 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:46:31,415 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:46:34,416 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 1 blocks at priority level 0; Total=1 Reset bookmarks? true 2017-02-23 21:46:34,448 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:46:37,487 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:46:37,518 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:46:40,519 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:46:40,519 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:46:43,535 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:46:43,566 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:46:46,567 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:46:46,567 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:46:49,588 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:46:49,589 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:46:52,595 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:46:52,595 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:46:55,596 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:46:55,596 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:46:58,608 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:46:58,609 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:01,610 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 1 blocks at priority level 0; Total=1 Reset bookmarks? true 2017-02-23 21:47:01,624 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:04,637 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:47:04,637 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:07,638 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:47:07,655 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:10,656 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:47:10,657 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:13,657 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:47:13,675 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:16,676 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:47:16,695 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:19,696 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:47:19,712 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:22,714 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:47:22,716 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:25,728 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:47:25,729 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:28,743 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 1 blocks at priority level 0; Total=1 Reset bookmarks? true 2017-02-23 21:47:28,754 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:31,788 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:47:31,813 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:34,841 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:47:34,842 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:37,842 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:47:37,846 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:40,862 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:47:40,881 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:43,893 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:47:43,894 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:46,895 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:47:46,895 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:49,896 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:47:49,897 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:52,897 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:47:52,917 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:55,940 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 1 blocks at priority level 0; Total=1 Reset bookmarks? true 2017-02-23 21:47:55,941 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:47:58,941 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:47:58,942 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:01,942 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:48:01,943 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:04,943 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:48:04,944 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:07,945 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:48:07,946 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:10,946 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:48:10,972 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:14,032 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:48:14,064 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:17,065 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:48:17,329 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:20,330 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:48:20,330 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:23,343 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 1 blocks at priority level 0; Total=1 Reset bookmarks? true 2017-02-23 21:48:23,343 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:26,344 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:48:26,345 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:29,345 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:48:29,346 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:32,351 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:48:32,351 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:35,352 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:48:35,353 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:38,354 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:48:38,384 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:41,155 WARN [HBase-Metrics2-1] impl.MetricsConfig(125): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2017-02-23 21:48:41,384 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:48:41,385 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:44,385 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:48:44,386 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:47,399 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:48:47,418 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:50,418 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 1 blocks at priority level 0; Total=1 Reset bookmarks? true 2017-02-23 21:48:50,456 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:53,456 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:48:53,457 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:56,475 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:48:56,475 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:48:59,476 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:48:59,476 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:02,477 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:49:02,477 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:05,478 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:49:05,479 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:08,480 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:49:08,498 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:11,506 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:49:11,508 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:14,508 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:49:14,509 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:17,510 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 1 blocks at priority level 0; Total=1 Reset bookmarks? true 2017-02-23 21:49:17,510 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:20,511 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:49:20,542 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:23,560 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:49:23,619 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:26,629 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:49:26,629 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:29,644 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:49:29,667 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:32,680 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:49:32,681 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:35,681 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:49:35,708 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:38,732 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:49:38,737 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:42,133 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:49:42,148 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:45,172 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 1 blocks at priority level 0; Total=1 Reset bookmarks? true 2017-02-23 21:49:45,175 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:48,175 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:49:48,195 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:51,195 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:49:51,225 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:54,226 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:49:54,227 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:49:57,227 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:49:57,255 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:00,273 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:50:00,274 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:03,293 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:50:03,352 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:06,375 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:50:06,449 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:09,462 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:50:09,489 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:12,517 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 1 blocks at priority level 0; Total=1 Reset bookmarks? true 2017-02-23 21:50:12,558 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:15,590 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:50:15,591 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:18,620 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:50:18,704 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:21,704 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:50:21,705 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:24,721 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:50:24,771 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:27,832 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:50:27,839 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:30,880 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:50:30,944 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:33,953 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:50:34,069 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:37,097 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:50:37,141 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:40,165 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 1 blocks at priority level 0; Total=1 Reset bookmarks? true 2017-02-23 21:50:40,212 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:43,213 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:50:43,227 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:46,227 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:50:46,300 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:49,380 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:50:49,409 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:49,653 INFO [Socket Reader #1 for port 53615] ipc.Server$Connection(1573): Auth successful for appattempt_1487884137933_0001_000001 (auth:SIMPLE) 2017-02-23 21:50:49,747 INFO [Socket Reader #1 for port 53615] ipc.Server$Connection(1573): Auth successful for appattempt_1487884137933_0001_000001 (auth:SIMPLE) 2017-02-23 21:50:49,986 INFO [Socket Reader #1 for port 53615] ipc.Server$Connection(1573): Auth successful for appattempt_1487884137933_0001_000001 (auth:SIMPLE) 2017-02-23 21:50:50,131 INFO [Socket Reader #1 for port 53615] ipc.Server$Connection(1573): Auth successful for appattempt_1487884137933_0001_000001 (auth:SIMPLE) 2017-02-23 21:50:50,615 WARN [ContainersLauncher #2] nodemanager.DefaultContainerExecutor(249): Exit code from container container_1487884137933_0001_01_000025 is : 143 2017-02-23 21:50:52,274 WARN [ContainersLauncher #0] nodemanager.DefaultContainerExecutor(249): Exit code from container container_1487884137933_0001_01_000022 is : 143 2017-02-23 21:50:52,481 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:50:52,533 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:53,657 INFO [Socket Reader #1 for port 54853] ipc.Server$Connection(1573): Auth successful for appattempt_1487884137933_0001_000001 (auth:SIMPLE) 2017-02-23 21:50:53,712 INFO [Socket Reader #1 for port 54853] ipc.Server$Connection(1573): Auth successful for appattempt_1487884137933_0001_000001 (auth:SIMPLE) 2017-02-23 21:50:53,841 WARN [ContainersLauncher #2] nodemanager.DefaultContainerExecutor(249): Exit code from container container_1487884137933_0001_01_000026 is : 143 2017-02-23 21:50:55,534 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:50:55,537 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:50:55,570 WARN [ContainersLauncher #3] nodemanager.DefaultContainerExecutor(249): Exit code from container container_1487884137933_0001_01_000023 is : 143 2017-02-23 21:50:57,258 WARN [ContainersLauncher #1] nodemanager.DefaultContainerExecutor(249): Exit code from container container_1487884137933_0001_01_000027 is : 143 2017-02-23 21:50:58,553 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.UnderReplicatedBlocks(395): chooseUnderReplicatedBlocks selected 2 blocks at priority level 0; Total=2 Reset bookmarks? false 2017-02-23 21:50:58,586 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 17, pendingReplications = 0. 2017-02-23 21:51:00,213 WARN [ContainersLauncher #1] nodemanager.DefaultContainerExecutor(249): Exit code from container container_1487884137933_0001_01_000024 is : 143 2017-02-23 21:51:01,620 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 0, pendingReplications = 0. 2017-02-23 21:51:02,749 WARN [ContainersLauncher #0] nodemanager.DefaultContainerExecutor(249): Exit code from container container_1487884137933_0001_01_000001 is : 1 2017-02-23 21:51:02,791 WARN [ContainersLauncher #0] nodemanager.DefaultContainerExecutor(255): Exception from container-launch with container ID: container_1487884137933_0001_01_000001 and exit code: 1 ExitCodeException exitCode=1: at org.apache.hadoop.util.Shell.runCommand(Shell.java:933) at org.apache.hadoop.util.Shell.run(Shell.java:844) at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:1123) at org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:237) at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:317) at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:83) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) at java.util.concurrent.FutureTask.run(FutureTask.java:166) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:722) 2017-02-23 21:51:03,156 WARN [ContainersLauncher #0] launcher.ContainerLaunch(365): Container exited with a non-zero exit code 1 2017-02-23 21:51:04,631 INFO [org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$ReplicationMonitor@5d1456e] blockmanagement.BlockManager(1594): BLOCK* neededReplications = 0, pendingReplications = 0. 2017-02-23 21:51:04,843 ERROR [IPC Server handler 0 on 43607] hs.CachedHistoryStorage(196): Unable to find job job_1487884137933_0001 2017-02-23 21:51:05,095 ERROR [IPC Server handler 8 on 43607] hs.CachedHistoryStorage(196): Unable to find job job_1487884137933_0001 2017-02-23 21:51:05,378 ERROR [IPC Server handler 1 on 43607] hs.CachedHistoryStorage(196): Unable to find job job_1487884137933_0001 2017-02-23 21:51:05,764 DEBUG [main] ipc.RpcClientImpl(1115): Stopping rpc client 2017-02-23 21:51:06,052 WARN [AsyncDispatcher event handler] nodemanager.NMAuditLogger(150): USER=nobody OPERATION=Container Finished - Failed TARGET=ContainerImpl RESULT=FAILURE DESCRIPTION=Container failed with state: EXITED_WITH_FAILURE APPID=application_1487884137933_0001 CONTAINERID=container_1487884137933_0001_01_000001