hbase (2.1.2, 4 regionservers) hbase pe \ > --nomapred \ > --oneCon=true \ > --valueSize=8 \ > --compress=SNAPPY \ > --rows=200000 \ > --presplit=10 \ > --columns=101 \ > --autoFlush=true \ > --multiPut=5000 \ > --writeToWAL=false \ > --table=test_for_randomWriteRead \ > randomWrite 1 SLF4J: Class path contains multiple SLF4J bindings. SLF4J: Found binding in [jar:file:/data/hadoop3/hadoop-3.1.2/share/hadoop/common/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found binding in [jar:file:/data/hadoop3/hbase-2.1.2/lib/client-facing-thirdparty/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] 2020-12-11 14:45:20,929 INFO [main] hbase.PerformanceEvaluation: RandomWriteTest test run options={"addColumns":true,"autoFlush":true,"blockEncoding":"NONE","blockSize":65536,"bloomType":"ROW","bufferSize":2097152,"caching":30,"cmdName":"randomWrite","columns":101,"compression":"SNAPPY","connCount":-1,"cycles":1,"families":1,"filterAll":false,"flushCommits":true,"inMemoryCF":false,"inMemoryCompaction":"NONE","measureAfter":0,"multiGet":0,"multiPut":5000,"noOfTags":1,"nomapred":true,"numClientThreads":1,"oneCon":true,"perClientRunRows":200000,"period":104857,"presplitRegions":10,"randomSleep":0,"replicas":1,"reportLatency":false,"sampleRate":1.0,"size":0.0,"splitPolicy":null,"startRow":0,"tableName":"test_for_randomWriteRead","totalRows":200000,"traceRate":0.0,"useTags":false,"valueRandom":false,"valueSize":8,"valueZipf":false,"writeToWAL":false} 2020-12-11 14:45:21,228 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677] zookeeper.ZooKeeper: Client environment:zookeeper.version=3.4.10-39d3a4f269333c922ed3db283be479f9deacaa0f, built on 03/23/2017 10:13 GMT 2020-12-11 14:45:21,228 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677] zookeeper.ZooKeeper: Client environment:host.name=cdh129130 2020-12-11 14:45:21,229 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677] zookeeper.ZooKeeper: Client environment:java.version=1.8.0_192 2020-12-11 14:45:21,229 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677] zookeeper.ZooKeeper: Client environment:java.vendor=Oracle Corporation 2020-12-11 14:45:21,229 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677] zookeeper.ZooKeeper: Client environment:java.home=/usr/java/jdk1.8.0_192/jre 2020-12-11 14:45:21,229 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677] zookeeper.ZooKeeper: op-3.1.2/share/hadoop/yarn/hadoop-yarn-common-3.1.2.jar:/data/hadoop3/hadoop-3.1.2/share/hadoop/yarn/hadoop-yarn-applications-distributedshell-3.1.2.jar:/data/hadoop3/hadoop-3.1.2/share/hadoop/yarn/hadoop-yarn-registry-3.1.2.jar:/data/hadoop3/hadoop-3.1.2/share/hadoop/yarn/hadoop-yarn-server-nodemanager-3.1.2.jar:/data/hadoop3/hadoop-3.1.2/share/hadoop/yarn/hadoop-yarn-server-tests-3.1.2.jar:/data/hadoop3/hadoop-3.1.2/share/hadoop/yarn/hadoop-yarn-server-sharedcachemanager-3.1.2.jar:/data/hadoop3/hadoop-3.1.2/share/hadoop/yarn/hadoop-yarn-server-common-3.1.2.jar:/data/hadoop3/hadoop-3.1.2/share/hadoop/yarn/hadoop-yarn-applications-unmanaged-am-launcher-3.1.2.jar:/data/hadoop3/hadoop-3.1.2/share/hadoop/yarn/hadoop-yarn-server-resourcemanager-3.1.2.jar:/data/hadoop3/hadoop-3.1.2/share/hadoop/yarn/hadoop-yarn-server-web-proxy-3.1.2.jar:/data/ozone-1.0.0/share/ozone/lib/hadoop-ozone-filesystem-hadoop3-1.0.0.jar:/data/hadoop3/hbase-2.1.2/lib/client-facing-thirdparty/slf4j-log4j12-1.7.25.jar 2020-12-11 14:45:21,229 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677] zookeeper.ZooKeeper: Client environment:java.library.path=/data/hadoop3/hadoop-3.1.2/lib/native 2020-12-11 14:45:21,230 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677] zookeeper.ZooKeeper: Client environment:java.io.tmpdir=/tmp 2020-12-11 14:45:21,230 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677] zookeeper.ZooKeeper: Client environment:java.compiler= 2020-12-11 14:45:21,230 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677] zookeeper.ZooKeeper: Client environment:os.name=Linux 2020-12-11 14:45:21,230 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677] zookeeper.ZooKeeper: Client environment:os.arch=amd64 2020-12-11 14:45:21,230 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677] zookeeper.ZooKeeper: Client environment:os.version=2.6.32-573.el6.x86_64 2020-12-11 14:45:21,230 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677] zookeeper.ZooKeeper: Client environment:user.name=hadoop3 2020-12-11 14:45:21,230 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677] zookeeper.ZooKeeper: Client environment:user.home=/home/hadoop3 2020-12-11 14:45:21,230 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677] zookeeper.ZooKeeper: Client environment:user.dir=/home/hadoop3 2020-12-11 14:45:21,233 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677] zookeeper.ZooKeeper: Initiating client connection, connectString=192.168.128.131:2281 sessionTimeout=90000 watcher=org.apache.hadoop.hbase.zookeeper.ReadOnlyZKClient$$Lambda$12/642447544@7a0916c4 2020-12-11 14:45:21,266 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677-SendThread(192.168.128.131:2281)] zookeeper.ClientCnxn: Opening socket connection to server 192.168.128.131/192.168.128.131:2281. Will not attempt to authenticate using SASL (unknown error) 2020-12-11 14:45:21,275 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677-SendThread(192.168.128.131:2281)] zookeeper.ClientCnxn: Socket connection established to 192.168.128.131/192.168.128.131:2281, initiating session 2020-12-11 14:45:21,293 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677-SendThread(192.168.128.131:2281)] zookeeper.ClientCnxn: Session establishment complete on server 192.168.128.131/192.168.128.131:2281, sessionid = 0x1764ab4c2680115, negotiated timeout = 40000 2020-12-11 14:45:22,513 INFO [main] client.HBaseAdmin: Started disable of test_for_randomWriteRead 2020-12-11 14:45:24,805 INFO [main] client.HBaseAdmin: Operation: DISABLE, Table Name: default:test_for_randomWriteRead, procId: 277 completed 2020-12-11 14:45:25,253 INFO [main] client.HBaseAdmin: Operation: DELETE, Table Name: default:test_for_randomWriteRead, procId: 288 completed 2020-12-11 14:45:26,522 INFO [main] client.HBaseAdmin: Operation: CREATE, Table Name: default:test_for_randomWriteRead, procId: 289 completed 2020-12-11 14:45:26,522 INFO [main] hbase.PerformanceEvaluation: Table 'test_for_randomWriteRead', {NAME => 'info0', VERSIONS => '1', EVICT_BLOCKS_ON_CLOSE => 'false', NEW_VERSION_BEHAVIOR => 'false', KEEP_DELETED_CELLS => 'FALSE', CACHE_DATA_ON_WRITE => 'false', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', CACHE_INDEX_ON_WRITE => 'false', IN_MEMORY => 'false', CACHE_BLOOMS_ON_WRITE => 'false', PREFETCH_BLOCKS_ON_OPEN => 'false', COMPRESSION => 'SNAPPY', BLOCKCACHE => 'true', BLOCKSIZE => '65536', METADATA => {'IN_MEMORY_COMPACTION' => 'NONE'}} created 2020-12-11 14:45:26,526 INFO [main] client.ConnectionImplementation: Closing master protocol: MasterService 2020-12-11 14:45:26,532 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x665e9289] zookeeper.ZooKeeper: Initiating client connection, connectString=192.168.128.131:2281 sessionTimeout=90000 watcher=org.apache.hadoop.hbase.zookeeper.ReadOnlyZKClient$$Lambda$12/642447544@7a0916c4 2020-12-11 14:45:26,534 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x665e9289-SendThread(192.168.128.131:2281)] zookeeper.ClientCnxn: Opening socket connection to server 192.168.128.131/192.168.128.131:2281. Will not attempt to authenticate using SASL (unknown error) 2020-12-11 14:45:26,535 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x665e9289-SendThread(192.168.128.131:2281)] zookeeper.ClientCnxn: Socket connection established to 192.168.128.131/192.168.128.131:2281, initiating session 2020-12-11 14:45:26,539 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677] zookeeper.ZooKeeper: Session: 0x1764ab4c2680115 closed 2020-12-11 14:45:26,542 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x6caf0677-EventThread] zookeeper.ClientCnxn: EventThread shut down for session: 0x1764ab4c2680115 2020-12-11 14:45:26,548 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x665e9289-SendThread(192.168.128.131:2281)] zookeeper.ClientCnxn: Session establishment complete on server 192.168.128.131/192.168.128.131:2281, sessionid = 0x1764ab4c2680116, negotiated timeout = 40000 2020-12-11 14:45:26,552 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x7d3430a7] zookeeper.ZooKeeper: Initiating client connection, connectString=192.168.128.131:2281 sessionTimeout=90000 watcher=org.apache.hadoop.hbase.zookeeper.ReadOnlyZKClient$$Lambda$12/642447544@7a0916c4 2020-12-11 14:45:26,553 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x7d3430a7-SendThread(192.168.128.131:2281)] zookeeper.ClientCnxn: Opening socket connection to server 192.168.128.131/192.168.128.131:2281. Will not attempt to authenticate using SASL (unknown error) 2020-12-11 14:45:26,554 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x7d3430a7-SendThread(192.168.128.131:2281)] zookeeper.ClientCnxn: Socket connection established to 192.168.128.131/192.168.128.131:2281, initiating session 2020-12-11 14:45:26,565 INFO [ReadOnlyZKClient-192.168.128.131:2281@0x7d3430a7-SendThread(192.168.128.131:2281)] zookeeper.ClientCnxn: Session establishment complete on server 192.168.128.131/192.168.128.131:2281, sessionid = 0x1764ab4c2680117, negotiated timeout = 40000 2020-12-11 14:45:26,581 INFO [main] hbase.PerformanceEvaluation: Created 1 connections for 1 threads 2020-12-11 14:45:26,584 INFO [TestClient-0] hbase.PerformanceEvaluation: Start class org.apache.hadoop.hbase.PerformanceEvaluation$RandomWriteTest at offset 0 for 200000 rows 2020-12-11 14:45:26,591 INFO [TestClient-0] hbase.PerformanceEvaluation: Sampling 1 every 1 out of 200000 total rows. 2020-12-11 14:45:26,591 INFO [TestClient-0] hbase.PerformanceEvaluation: MultiPut enabled. Sending PUTs in batches of 5000. 2020-12-11 14:45:26,775 INFO [TestClient-0] hbase.PerformanceEvaluation: Timed test starting in thread TestClient-0 2020-12-11 14:45:32,388 INFO [hconnection-0x10dab199-shared-pool7-t3] client.AsyncRequestFutureImpl: 80000,1607669125287.6b7d586c7de07f2db08b41dd0a2cca6b.:info0 Above parallelPutToStoreThreadLimit(10) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:1063) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicBatchOp(RSRpcServices.java:966) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:929) at org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2681) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:42014) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:413) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:130) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:324) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:304) on cdh129139,16020,1607656079881, tracking started null, retrying after=2011ms, operationsToReplay=737 2020-12-11 14:45:32,456 INFO [hconnection-0x10dab199-shared-pool7-t4] client.AsyncRequestFutureImpl: 80000,1607669125287.5662eaedaaa9cf68705c2ad0e9cdfa83.:info0 Above parallelPutToStoreThreadLimit(10) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:1063) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicBatchOp(RSRpcServices.java:966) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:929) at org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2681) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:42014) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:413) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:130) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:324) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:304) on cdh129144,16020,1607656080173, tracking started null, retrying after=2019ms, operationsToReplay=765 2020-12-11 14:45:32,706 INFO [hconnection-0x10dab199-shared-pool7-t1] client.AsyncRequestFutureImpl: 0000,1607669125287.f4ae72e3f742988d4d5bbce36d1fdd48.:info0 Above parallelPutToStoreThreadLimit(10) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:1063) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicBatchOp(RSRpcServices.java:966) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:929) at org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2681) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:42014) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:413) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:130) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:324) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:304) on cdh129136,16020,1607656080098, tracking started null, retrying after=2013ms, operationsToReplay=1118 2020-12-11 14:45:32,902 INFO [hconnection-0x10dab199-shared-pool7-t2] client.AsyncRequestFutureImpl: 0000,1607669125287.a0c9c062ba049a96f2bfe7a054535637.:info0 Above parallelPutToStoreThreadLimit(10) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:1063) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicBatchOp(RSRpcServices.java:966) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:929) at org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2681) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:42014) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:413) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:130) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:324) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:304) on cdh129135,16020,1607656079823, tracking started null, retrying after=2012ms, operationsToReplay=1180 2020-12-11 14:45:34,567 INFO [hconnection-0x10dab199-shared-pool7-t3] client.AsyncRequestFutureImpl: 80000,1607669125287.6b7d586c7de07f2db08b41dd0a2cca6b.:info0 Above parallelPutToStoreThreadLimit(10) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:1063) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicBatchOp(RSRpcServices.java:966) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:929) at org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2681) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:42014) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:413) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:130) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:324) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:304) on cdh129139,16020,1607656079881, tracking started null, retrying after=4025ms, operationsToReplay=697 2020-12-11 14:45:34,640 INFO [hconnection-0x10dab199-shared-pool7-t4] client.AsyncRequestFutureImpl: 80000,1607669125287.5662eaedaaa9cf68705c2ad0e9cdfa83.:info0 Above parallelPutToStoreThreadLimit(10) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:1063) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicBatchOp(RSRpcServices.java:966) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:929) at org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2681) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:42014) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:413) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:130) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:324) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:304) on cdh129144,16020,1607656080173, tracking started null, retrying after=4026ms, operationsToReplay=725 2020-12-11 14:45:34,965 INFO [hconnection-0x10dab199-shared-pool7-t1] client.AsyncRequestFutureImpl: 0000,1607669125287.f4ae72e3f742988d4d5bbce36d1fdd48.:info0 Above parallelPutToStoreThreadLimit(10) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:1063) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicBatchOp(RSRpcServices.java:966) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:929) at org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2681) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:42014) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:413) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:130) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:324) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:304) on cdh129136,16020,1607656080098, tracking started null, retrying after=4028ms, operationsToReplay=1058 2020-12-11 14:45:35,188 INFO [hconnection-0x10dab199-shared-pool7-t2] client.AsyncRequestFutureImpl: 0000,1607669125287.a0c9c062ba049a96f2bfe7a054535637.:info0 Above parallelPutToStoreThreadLimit(10) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:1063) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicBatchOp(RSRpcServices.java:966) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:929) at org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2681) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:42014) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:413) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:130) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:324) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:304) on cdh129135,16020,1607656079823, tracking started null, retrying after=4031ms, operationsToReplay=1120 2020-12-11 14:45:37,806 INFO [TestClient-0] client.AsyncRequestFutureImpl: #2, waiting for 3600 actions to finish on table: test_for_randomWriteRead 2020-12-11 14:45:38,751 INFO [hconnection-0x10dab199-shared-pool7-t3] client.AsyncRequestFutureImpl: 0000,1607669125287.6b7d586c7de07f2db08b41dd0a2cca6b.:info0 Above parallelPutToStoreThreadLimit(10) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:1063) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicBatchOp(RSRpcServices.java:966) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:929) at org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2681) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:42014) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:413) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:130) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:324) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:304) on cdh129139,16020,1607656079881, tracking started null, retrying after=10082ms, operationsToReplay=657 2020-12-11 14:45:38,828 INFO [hconnection-0x10dab199-shared-pool7-t4] client.AsyncRequestFutureImpl: 0000,1607669125287.5662eaedaaa9cf68705c2ad0e9cdfa83.:info0 Above parallelPutToStoreThreadLimit(10) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:1063) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicBatchOp(RSRpcServices.java:966) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:929) at org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2681) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:42014) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:413) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:130) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:324) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:304) on cdh129144,16020,1607656080173, tracking started null, retrying after=10076ms, operationsToReplay=685 2020-12-11 14:45:39,481 INFO [hconnection-0x10dab199-shared-pool7-t1] client.AsyncRequestFutureImpl: 0000,1607669125287.f4ae72e3f742988d4d5bbce36d1fdd48.:info0 Above parallelPutToStoreThreadLimit(10) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:1063) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicBatchOp(RSRpcServices.java:966) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:929) at org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2681) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:42014) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:413) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:130) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:324) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:304) on cdh129136,16020,1607656080098, tracking started null, retrying after=10022ms, operationsToReplay=998 2020-12-11 14:45:39,715 INFO [hconnection-0x10dab199-shared-pool7-t2] client.AsyncRequestFutureImpl: 000,1607669125287.a0c9c062ba049a96f2bfe7a054535637.:info0 Above parallelPutToStoreThreadLimit(10) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:1063) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicBatchOp(RSRpcServices.java:966) at org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:929) at org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2681) at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:42014) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:413) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:130) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:324) at org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:304) on cdh129135,16020,1607656079823, tracking started null, retrying after=10057ms, operationsToReplay=1060