Strange, I checked the code, it seems we get NPE on this line

https://github.com/apache/hbase/blob/4d7ce1aac724fbf09e526fc422b5a11e530c32f0/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterRpcServices.java#L2872

Could you please confirm that you connect to the correct active master
which is hanging? It seems that you are connecting the backup
master...

Thanks.

张铎(Duo Zhang) <palomino...@gmail.com> 于2024年4月23日周二 15:31写道:
>
> Ah, NPE usually means a code bug, then there is no simple way to fix
> it, need to take a deep look on the code :(
>
> Sorry.
>
> Udo Offermann <udo.offerm...@zfabrik.de> 于2024年4月22日周一 15:32写道:
> >
> > Unfortunately not.
> > I’ve found the node hosting the meta region and was able to run hack 
> > scheduleRecoveries using hbase-operator-tools-1.2.0.
> > The tool however stops with an NPE:
> >
> > 09:22:00.532 [main] WARN  org.apache.hadoop.util.NativeCodeLoader - Unable 
> > to load native-hadoop library for your platform... using builtin-java 
> > classes where applicable
> > 09:22:00.703 [main] INFO  org.apache.hadoop.conf.Configuration.deprecation 
> > - hbase.client.pause.cqtbe is deprecated. Instead, use 
> > hbase.client.pause.server.overloaded
> > 09:22:00.765 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Client 
> > environment:zookeeper.version=3.8.3-6ad6d364c7c0bcf0de452d54ebefa3058098ab56,
> >  built on 2023-10-05 10:34 UTC
> > 09:22:00.765 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Client 
> > environment:host.name=HBaseMaster.gmd9.intern
> > 09:22:00.765 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Client 
> > environment:java.version=1.8.0_402
> > 09:22:00.766 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Client 
> > environment:java.vendor=Red Hat, Inc.
> > 09:22:00.766 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Client 
> > environment:java.home=/usr/lib/jvm/java-1.8.0-openjdk-1.8.0.402.b06-2.el8.x86_64/jre
> > 09:22:00.766 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Client 
> > environment:java.class.path=hbase-operator-tools-1.2.0/hbase-hbck2/hbase-hbck2-1.2.0.jar:hbase/conf:/opt/seritrack/tt/jdk/lib/tools.jar:/opt/seritrack/tt/nosql/hbase:/opt/seritrack/tt/nosql/hbase/lib/shaded-clients/hbase-shaded-mapreduce-2.5.7.jar:/opt/seritrack/tt/nosql/hbase/lib/client-facing-thirdparty/audience-annotations-0.13.0.jar:/opt/seritrack/tt/nosql/hbase/lib/client-facing-thirdparty/commons-logging-1.2.jar:/opt/seritrack/tt/nosql/hbase/lib/client-facing-thirdparty/htrace-core4-4.1.0-incubating.jar:/opt/seritrack/tt/nosql/hbase/lib/client-facing-thirdparty/jcl-over-slf4j-1.7.33.jar:/opt/seritrack/tt/nosql/hbase/lib/client-facing-thirdparty/jul-to-slf4j-1.7.33.jar:/opt/seritrack/tt/nosql/hbase/lib/client-facing-thirdparty/opentelemetry-api-1.15.0.jar:/opt/seritrack/tt/nosql/hbase/lib/client-facing-thirdparty/opentelemetry-context-1.15.0.jar:/opt/seritrack/tt/nosql/hbase/lib/client-facing-thirdparty/opentelemetry-semconv-1.15.0-alpha.jar:/opt/seritrack/tt/nosql/hbase/lib/client-facing-thirdparty/slf4j-api-1.7.33.jar:/opt/seritrack/tt/nosql/hbase/lib/shaded-clients/hbase-shaded-client-2.5.7.jar:/opt/seritrack/tt/nosql/pl_nosql_ext/libs/pl_nosql_ext-3.0.jar:/opt/seritrack/tt/nosql/hbase/lib/client-facing-thirdparty/log4j-1.2-api-2.17.2.jar:/opt/seritrack/tt/nosql/hbase/lib/client-facing-thirdparty/log4j-api-2.17.2.jar:/opt/seritrack/tt/nosql/hbase/lib/client-facing-thirdparty/log4j-core-2.17.2.jar:/opt/seritrack/tt/nosql/hbase/lib/client-facing-thirdparty/log4j-slf4j-impl-2.17.2.jar:/opt/seritrack/tt/prometheus_exporters/jmx_exporter/jmx_prometheus_javaagent.jar
> > 09:22:00.766 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Client 
> > environment:java.library.path=/opt/seritrack/tt/nosql/hadoop/lib/native
> > 09:22:00.766 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Client 
> > environment:java.io.tmpdir=/tmp
> > 09:22:00.766 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Client 
> > environment:java.compiler=<NA>
> > 09:22:00.766 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Client 
> > environment:os.name=Linux
> > 09:22:00.766 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Client 
> > environment:os.arch=amd64
> > 09:22:00.766 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Client 
> > environment:os.version=4.18.0-513.18.1.el8_9.x86_64
> > 09:22:00.766 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Client 
> > environment:user.name=seritrack
> > 09:22:00.766 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Client 
> > environment:user.home=/opt/seritrack
> > 09:22:00.766 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Client 
> > environment:user.dir=/opt/seritrack/tt/nosql_3.0
> > 09:22:00.766 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Client 
> > environment:os.memory.free=275MB
> > 09:22:00.766 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Client 
> > environment:os.memory.max=2966MB
> > 09:22:00.766 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Client 
> > environment:os.memory.total=361MB
> > 09:22:00.771 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ZooKeeper - Initiating 
> > client connection, connectString=HBaseMaster:2181 sessionTimeout=90000 
> > watcher=org.apache.hadoop.hbase.zookeeper.ReadOnlyZKClient$$Lambda$45/1091799416@aed32c5
> > 09:22:00.774 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.common.X509Util - 
> > Setting -D jdk.tls.rejectClientInitiatedRenegotiation=true to disable 
> > client-initiated TLS renegotiation
> > 09:22:00.777 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ClientCnxnSocket - 
> > jute.maxbuffer value is 1048575 Bytes
> > 09:22:00.785 [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f] INFO  
> > org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ClientCnxn - 
> > zookeeper.request.timeout value is 0. feature enabled=false
> > 09:22:00.793 
> > [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f-SendThread(HBaseMaster:2181)] 
> > INFO  org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ClientCnxn - 
> > Opening socket connection to server HBaseMaster/10.21.204.230:2181.
> > 09:22:00.793 
> > [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f-SendThread(HBaseMaster:2181)] 
> > INFO  org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ClientCnxn - SASL 
> > config status: Will not attempt to authenticate using SASL (unknown error)
> > 09:22:00.797 
> > [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f-SendThread(HBaseMaster:2181)] 
> > INFO  org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ClientCnxn - 
> > Socket connection established, initiating session, client: 
> > /10.21.204.230:41072, server: HBaseMaster/10.21.204.230:2181
> > 09:22:00.801 
> > [ReadOnlyZKClient-HBaseMaster:2181@0x7d9f158f-SendThread(HBaseMaster:2181)] 
> > INFO  org.apache.hadoop.hbase.shaded.org.apache.zookeeper.ClientCnxn - 
> > Session establishment complete on server HBaseMaster/10.21.204.230:2181, 
> > session id = 0x10009a4f379001e, negotiated timeout = 90000
> > -1
> > Exception in thread "main" java.io.IOException: 
> > org.apache.hbase.thirdparty.com.google.protobuf.ServiceException: 
> > org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(java.io.IOException): 
> > java.io.IOException
> >         at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:479)
> >         at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124)
> >         at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102)
> >         at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > Caused by: java.lang.NullPointerException
> >         at 
> > org.apache.hadoop.hbase.master.MasterRpcServices.shouldSubmitSCP(MasterRpcServices.java:2872)
> >         at 
> > org.apache.hadoop.hbase.master.MasterRpcServices.scheduleServerCrashProcedure(MasterRpcServices.java:2600)
> >         at 
> > org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$HbckService$2.callBlockingMethod(MasterProtos.java)
> >         at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:415)
> >         ... 3 more
> >
> >         at 
> > org.apache.hadoop.hbase.client.HBaseHbck.scheduleServerCrashProcedures(HBaseHbck.java:198)
> >         at 
> > org.apache.hadoop.hbase.client.Hbck.scheduleServerCrashProcedure(Hbck.java:128)
> >         at org.apache.hbase.HBCK2.scheduleRecoveries(HBCK2.java:418)
> >         at org.apache.hbase.HBCK2.doCommandLine(HBCK2.java:960)
> >         at org.apache.hbase.HBCK2.run(HBCK2.java:830)
> >         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)
> >         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90)
> >         at org.apache.hbase.HBCK2.main(HBCK2.java:1145)
> > Caused by: 
> > org.apache.hbase.thirdparty.com.google.protobuf.ServiceException: 
> > org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(java.io.IOException): 
> > java.io.IOException
> >         at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:479)
> >         at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124)
> >         at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102)
> >         at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > Caused by: java.lang.NullPointerException
> >         at 
> > org.apache.hadoop.hbase.master.MasterRpcServices.shouldSubmitSCP(MasterRpcServices.java:2872)
> >         at 
> > org.apache.hadoop.hbase.master.MasterRpcServices.scheduleServerCrashProcedure(MasterRpcServices.java:2600)
> >         at 
> > org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$HbckService$2.callBlockingMethod(MasterProtos.java)
> >         at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:415)
> >         ... 3 more
> >
> >         at 
> > org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:340)
> >         at 
> > org.apache.hadoop.hbase.ipc.AbstractRpcClient.access$200(AbstractRpcClient.java:92)
> >         at 
> > org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:595)
> >         at 
> > org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$HbckService$BlockingStub.scheduleServerCrashProcedure(MasterProtos.java)
> >         at 
> > org.apache.hadoop.hbase.client.HBaseHbck.scheduleServerCrashProcedures(HBaseHbck.java:190)
> >         ... 7 more
> > Caused by: 
> > org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(java.io.IOException): 
> > java.io.IOException
> >         at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:479)
> >         at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124)
> >         at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:102)
> >         at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > Caused by: java.lang.NullPointerException
> >         at 
> > org.apache.hadoop.hbase.master.MasterRpcServices.shouldSubmitSCP(MasterRpcServices.java:2872)
> >         at 
> > org.apache.hadoop.hbase.master.MasterRpcServices.scheduleServerCrashProcedure(MasterRpcServices.java:2600)
> >         at 
> > org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos$HbckService$2.callBlockingMethod(MasterProtos.java)
> >         at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:415)
> >         ... 3 more
> >
> >         at 
> > org.apache.hadoop.hbase.ipc.AbstractRpcClient.onCallFinished(AbstractRpcClient.java:388)
> >         at 
> > org.apache.hadoop.hbase.ipc.AbstractRpcClient.access$100(AbstractRpcClient.java:92)
> >         at 
> > org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:425)
> >         at 
> > org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:420)
> >         at org.apache.hadoop.hbase.ipc.Call.callComplete(Call.java:114)
> >         at org.apache.hadoop.hbase.ipc.Call.setException(Call.java:129)
> >         at 
> > org.apache.hadoop.hbase.ipc.NettyRpcDuplexHandler.readResponse(NettyRpcDuplexHandler.java:199)
> >         at 
> > org.apache.hadoop.hbase.ipc.NettyRpcDuplexHandler.channelRead(NettyRpcDuplexHandler.java:220)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:442)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:346)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:318)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:286)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:442)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:440)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
> >         at 
> > org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
> >         at java.lang.Thread.run(Thread.java:750)
> >
> >
> >
> >
> > > Am 20.04.2024 um 15:53 schrieb 张铎(Duo Zhang) <palomino...@gmail.com>:
> > >
> > > OK, it was waitForMetaOnline.
> > >
> > > Maybe the problem is that you do have some correct procedures before
> > > upgrading, like ServerCrashProcedure, but then you delete all the
> > > procedure wals so the ServerCrashProcedure is also gone, so meta can
> > > never be online.
> > >
> > > Please check the /hbase/meta-region-server znode on zookeeper, dump
> > > its content, it is protobuf based but anyway, you could see the
> > > encoded server name which hosts meta region.
> > >
> > > Then use HBCK2, to schedule a SCP for this region server, to see if it
> > > can fix the problem.
> > >
> > > https://github.com/apache/hbase-operator-tools/blob/master/hbase-hbck2/README.md
> > >
> > > This is the document for HBCK2, you should use the scheduleRecoveries 
> > > command.
> > >
> > > Hope this could fix your problem.
> > >
> > > Thread 92 (master/masterserver:16000:becomeActiveMaster):
> > >  State: TIMED_WAITING
> > >  Blocked count: 165
> > >  Waited count: 404
> > >  Stack:
> > >    java.lang.Thread.sleep(Native Method)
> > >    org.apache.hadoop.hbase.util.Threads.sleep(Threads.java:125)
> > >    
> > > org.apache.hadoop.hbase.master.HMaster.isRegionOnline(HMaster.java:1358)
> > >
> > > org.apache.hadoop.hbase.master.HMaster.waitForMetaOnline(HMaster.java:1328)
> > >
> > > org.apache.hadoop.hbase.master.HMaster.finishActiveMasterInitialization(HMaster.java:1069)
> > >
> > > org.apache.hadoop.hbase.master.HMaster.startActiveMasterManager(HMaster.java:2405)
> > >    org.apache.hadoop.hbase.master.HMaster.lambda$null$0(HMaster.java:565)
> > >
> > > org.apache.hadoop.hbase.master.HMaster$$Lambda$265/1598878738.run(Unknown
> > > Source)
> > >    org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187)
> > >    org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177)
> > >    org.apache.hadoop.hbase.master.HMaster.lambda$run$1(HMaster.java:562)
> > >
> > > org.apache.hadoop.hbase.master.HMaster$$Lambda$264/1129144214.run(Unknown
> > > Source)
> > >    java.lang.Thread.run(Thread.java:750)
> > >
> > > Udo Offermann <udo.offerm...@zfabrik.de 
> > > <mailto:udo.offerm...@zfabrik.de>> 于2024年4月20日周六 21:13写道:
> > >>
> > >> Master status for masterserver.gmd9.intern,16000,1713515965162 as of Fri
> > >> Apr 19 10:55:22 CEST 2024
> > >>
> > >>
> > >> Version Info:
> > >> ===========================================================
> > >> HBase 2.5.7
> > >> Source code repository
> > >> git://buildbox.localdomain/home/apurtell/tmp/RM/hbase
> > >> revision=6788f98356dd70b4a7ff766ea7a8298e022e7b95
> > >> Compiled by apurtell on Thu Dec 14 15:59:16 PST 2023
> > >> From source with checksum
> > >> 1501d7fdf72398791ee335a229d099fc972cea7c2a952da7622eb087ddf975361f107cbbbee5d0ad6f603466e9afa1f4fd242ffccbd4371eb0b56059bb3b5402
> > >> Hadoop 2.10.2
> > >> Source code repository Unknown
> > >> revision=965fd380006fa78b2315668fbc7eb432e1d8200f
> > >> Compiled by ubuntu on 2022-05-25T00:12Z
> > >>
> > >>
> > >> Tasks:
> > >> ===========================================================
> > >> Task: Master startup
> > >> Status: RUNNING:Starting assignment manager
> > >> Running for 954s
> > >>
> > >> Task: Flushing master:store,,1.1595e783b53d99cd5eef43b6debb2682.
> > >> Status: COMPLETE:Flush successful flush 
> > >> result:CANNOT_FLUSH_MEMSTORE_EMPTY,
> > >> failureReason:Nothing to flush,flush seq id14
> > >> Completed 49s ago
> > >> Ran for 0s
> > >>
> > >> Task: RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=16000
> > >> Status: WAITING:Waiting for a call
> > >> Running for 951s
> > >>
> > >> Task: RpcServer.priority.RWQ.Fifo.write.handler=1,queue=0,port=16000
> > >> Status: WAITING:Waiting for a call
> > >> Running for 951s
> > >>
> > >>
> > >>
> > >> Servers:
> > >> ===========================================================
> > >> servername1ct.gmd9.intern,16020,1713514863737: requestsPerSecond=0.0,
> > >> numberOfOnlineRegions=0, usedHeapMB=37.0MB, maxHeapMB=2966.0MB,
> > >> numberOfStores=0, numberOfStorefiles=0, storeRefCount=0,
> > >> maxCompactedStoreFileRefCount=0, storefileUncompressedSizeMB=0,
> > >> storefileSizeMB=0, memstoreSizeMB=0, readRequestsCount=0,
> > >> filteredReadRequestsCount=0, writeRequestsCount=0, rootIndexSizeKB=0,
> > >> totalStaticIndexSizeKB=0, totalStaticBloomSizeKB=0, totalCompactingKVs=0,
> > >> currentCompactedKVs=0, compactionProgressPct=NaN, coprocessors=[]
> > >> servername2ct.gmd9.intern,16020,1713514925960: requestsPerSecond=0.0,
> > >> numberOfOnlineRegions=0, usedHeapMB=20.0MB, maxHeapMB=2966.0MB,
> > >> numberOfStores=0, numberOfStorefiles=0, storeRefCount=0,
> > >> maxCompactedStoreFileRefCount=0, storefileUncompressedSizeMB=0,
> > >> storefileSizeMB=0, memstoreSizeMB=0, readRequestsCount=0,
> > >> filteredReadRequestsCount=0, writeRequestsCount=0, rootIndexSizeKB=0,
> > >> totalStaticIndexSizeKB=0, totalStaticBloomSizeKB=0, totalCompactingKVs=0,
> > >> currentCompactedKVs=0, compactionProgressPct=NaN, coprocessors=[]
> > >> servername3ct.gmd9.intern,16020,1713514937151: requestsPerSecond=0.0,
> > >> numberOfOnlineRegions=0, usedHeapMB=67.0MB, maxHeapMB=2966.0MB,
> > >> numberOfStores=0, numberOfStorefiles=0, storeRefCount=0,
> > >> maxCompactedStoreFileRefCount=0, storefileUncompressedSizeMB=0,
> > >> storefileSizeMB=0, memstoreSizeMB=0, readRequestsCount=0,
> > >> filteredReadRequestsCount=0, writeRequestsCount=0, rootIndexSizeKB=0,
> > >> totalStaticIndexSizeKB=0, totalStaticBloomSizeKB=0, totalCompactingKVs=0,
> > >> currentCompactedKVs=0, compactionProgressPct=NaN, coprocessors=[]
> > >> servername4ct.gmd9.intern,16020,1713514968019: requestsPerSecond=0.0,
> > >> numberOfOnlineRegions=0, usedHeapMB=24.0MB, maxHeapMB=2966.0MB,
> > >> numberOfStores=0, numberOfStorefiles=0, storeRefCount=0,
> > >> maxCompactedStoreFileRefCount=0, storefileUncompressedSizeMB=0,
> > >> storefileSizeMB=0, memstoreSizeMB=0, readRequestsCount=0,
> > >> filteredReadRequestsCount=0, writeRequestsCount=0, rootIndexSizeKB=0,
> > >> totalStaticIndexSizeKB=0, totalStaticBloomSizeKB=0, totalCompactingKVs=0,
> > >> currentCompactedKVs=0, compactionProgressPct=NaN, coprocessors=[]
> > >> servername5ct.gmd9.intern,16020,1713514979294: requestsPerSecond=0.0,
> > >> numberOfOnlineRegions=0, usedHeapMB=58.0MB, maxHeapMB=2966.0MB,
> > >> numberOfStores=0, numberOfStorefiles=0, storeRefCount=0,
> > >> maxCompactedStoreFileRefCount=0, storefileUncompressedSizeMB=0,
> > >> storefileSizeMB=0, memstoreSizeMB=0, readRequestsCount=0,
> > >> filteredReadRequestsCount=0, writeRequestsCount=0, rootIndexSizeKB=0,
> > >> totalStaticIndexSizeKB=0, totalStaticBloomSizeKB=0, totalCompactingKVs=0,
> > >> currentCompactedKVs=0, compactionProgressPct=NaN, coprocessors=[]
> > >> servername6ct.gmd9.intern,16020,1713514994770: requestsPerSecond=0.0,
> > >> numberOfOnlineRegions=0, usedHeapMB=31.0MB, maxHeapMB=2966.0MB,
> > >> numberOfStores=0, numberOfStorefiles=0, storeRefCount=0,
> > >> maxCompactedStoreFileRefCount=0, storefileUncompressedSizeMB=0,
> > >> storefileSizeMB=0, memstoreSizeMB=0, readRequestsCount=0,
> > >> filteredReadRequestsCount=0, writeRequestsCount=0, rootIndexSizeKB=0,
> > >> totalStaticIndexSizeKB=0, totalStaticBloomSizeKB=0, totalCompactingKVs=0,
> > >> currentCompactedKVs=0, compactionProgressPct=NaN, coprocessors=[]
> > >>
> > >>
> > >> Regions-in-transition:
> > >> ===========================================================
> > >>
> > >>
> > >> Executors:
> > >> ===========================================================
> > >>  Status for executor:
> > >> Executor-4-MASTER_META_SERVER_OPERATIONS-master/masterserver:16000
> > >>  =======================================
> > >>  0 events queued, 0 running
> > >>  Status for executor:
> > >> Executor-6-MASTER_SNAPSHOT_OPERATIONS-master/masterserver:16000
> > >>  =======================================
> > >>  0 events queued, 0 running
> > >>  Status for executor:
> > >> Executor-3-MASTER_SERVER_OPERATIONS-master/masterserver:16000
> > >>  =======================================
> > >>  0 events queued, 0 running
> > >>  Status for executor: 
> > >> Executor-5-M_LOG_REPLAY_OPS-master/masterserver:16000
> > >>  =======================================
> > >>  0 events queued, 0 running
> > >>  Status for executor:
> > >> Executor-2-MASTER_CLOSE_REGION-master/masterserver:16000
> > >>  =======================================
> > >>  0 events queued, 0 running
> > >>  Status for executor:
> > >> Executor-7-MASTER_MERGE_OPERATIONS-master/masterserver:16000
> > >>  =======================================
> > >>  0 events queued, 0 running
> > >>  Status for executor:
> > >> Executor-8-MASTER_TABLE_OPERATIONS-master/masterserver:16000
> > >>  =======================================
> > >>  0 events queued, 0 running
> > >>  Status for executor:
> > >> Executor-1-MASTER_OPEN_REGION-master/masterserver:16000
> > >>  =======================================
> > >>  0 events queued, 0 running
> > >>
> > >>
> > >> Stacks:
> > >> ===========================================================
> > >> Process Thread Dump:
> > >> 131 active threads
> > >> Thread 186 (WAL-Archive-0):
> > >>  State: WAITING
> > >>  Blocked count: 5
> > >>  Waited count: 11
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@42f44d41
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 185 (Close-WAL-Writer-0):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 2
> > >>  Waited count: 6
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:460)
> > >>
> > >> java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:362)
> > >>    java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:941)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 152 (Session-Scheduler-3bc4ef12-1):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093)
> > >>
> > >> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 151
> > >> (master/masterserver:16000:becomeActiveMaster-HFileCleaner.small.0-1713515973400):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@58626ec5
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.PriorityBlockingQueue.take(PriorityBlockingQueue.java:549)
> > >>
> > >> org.apache.hadoop.hbase.master.cleaner.HFileCleaner.consumerLoop(HFileCleaner.java:285)
> > >>
> > >> org.apache.hadoop.hbase.master.cleaner.HFileCleaner$2.run(HFileCleaner.java:269)
> > >> Thread 150
> > >> (master/masterserver:16000:becomeActiveMaster-HFileCleaner.large.0-1713515973400):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@18916420
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>    
> > >> org.apache.hadoop.hbase.util.StealJobQueue.take(StealJobQueue.java:101)
> > >>
> > >> org.apache.hadoop.hbase.master.cleaner.HFileCleaner.consumerLoop(HFileCleaner.java:285)
> > >>
> > >> org.apache.hadoop.hbase.master.cleaner.HFileCleaner$1.run(HFileCleaner.java:254)
> > >> Thread 149 (snapshot-hfile-cleaner-cache-refresher):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 4
> > >>  Waited count: 11
> > >>  Stack:
> > >>    java.lang.Object.wait(Native Method)
> > >>    java.util.TimerThread.mainLoop(Timer.java:552)
> > >>    java.util.TimerThread.run(Timer.java:505)
> > >> Thread 148 (master/masterserver:16000.Chore.1):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 2
> > >>  Waited count: 10
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093)
> > >>
> > >> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 147 (OldWALsCleaner-1):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7a6a3b7e
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>
> > >> org.apache.hadoop.hbase.master.cleaner.LogCleaner.deleteFile(LogCleaner.java:172)
> > >>
> > >> org.apache.hadoop.hbase.master.cleaner.LogCleaner.lambda$createOldWalsCleaner$1(LogCleaner.java:152)
> > >>
> > >> org.apache.hadoop.hbase.master.cleaner.LogCleaner$$Lambda$494/556458560.run(Unknown
> > >> Source)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 146 (OldWALsCleaner-0):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7a6a3b7e
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>
> > >> org.apache.hadoop.hbase.master.cleaner.LogCleaner.deleteFile(LogCleaner.java:172)
> > >>
> > >> org.apache.hadoop.hbase.master.cleaner.LogCleaner.lambda$createOldWalsCleaner$1(LogCleaner.java:152)
> > >>
> > >> org.apache.hadoop.hbase.master.cleaner.LogCleaner$$Lambda$494/556458560.run(Unknown
> > >> Source)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 139 (PEWorker-16):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 16
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:165)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:147)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2113)
> > >> Thread 138 (PEWorker-15):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 16
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:165)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:147)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2113)
> > >> Thread 137 (PEWorker-14):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 16
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:165)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:147)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2113)
> > >> Thread 136 (PEWorker-13):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 16
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:165)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:147)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2113)
> > >> Thread 135 (PEWorker-12):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 16
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:165)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:147)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2113)
> > >> Thread 134 (PEWorker-11):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 17
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:165)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:147)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2113)
> > >> Thread 133 (PEWorker-10):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 16
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:165)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:147)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2113)
> > >> Thread 132 (PEWorker-9):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 17
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:165)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:147)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2113)
> > >> Thread 131 (PEWorker-8):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 18
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:165)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:147)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2113)
> > >> Thread 130 (PEWorker-7):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 16
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:165)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:147)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2113)
> > >> Thread 129 (PEWorker-6):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 18
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:165)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:147)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2113)
> > >> Thread 128 (PEWorker-5):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 18
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:165)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:147)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2113)
> > >> Thread 127 (PEWorker-4):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 17
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:165)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:147)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2113)
> > >> Thread 126 (PEWorker-3):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 18
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:165)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:147)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2113)
> > >> Thread 125 (PEWorker-2):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 21
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:165)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:147)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2113)
> > >> Thread 124 (PEWorker-1):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 16
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:165)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.AbstractProcedureScheduler.poll(AbstractProcedureScheduler.java:147)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.ProcedureExecutor$WorkerThread.run(ProcedureExecutor.java:2113)
> > >> Thread 123 (WorkerMonitor):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 191
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>    java.util.concurrent.DelayQueue.poll(DelayQueue.java:273)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.util.DelayedUtil.takeWithoutInterrupt(DelayedUtil.java:81)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.TimeoutExecutorThread.run(TimeoutExecutorThread.java:56)
> > >> Thread 122 (ProcExecTimeout):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 64
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>    java.util.concurrent.DelayQueue.poll(DelayQueue.java:268)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.util.DelayedUtil.takeWithoutInterrupt(DelayedUtil.java:81)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.TimeoutExecutorThread.run(TimeoutExecutorThread.java:56)
> > >> Thread 145 (ActiveMasterInitializationMonitor-1713515973319):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 2
> > >>  Stack:
> > >>    java.lang.Thread.sleep(Native Method)
> > >>
> > >> org.apache.hadoop.hbase.master.MasterInitializationMonitor.run(MasterInitializationMonitor.java:63)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 143 (SnapshotHandlerChoreCleaner):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 95
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093)
> > >>
> > >> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 142 (normalizer-worker-0):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1ac17b12
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> org.apache.hadoop.hbase.master.normalizer.RegionNormalizerWorkQueue.take(RegionNormalizerWorkQueue.java:146)
> > >>
> > >> org.apache.hadoop.hbase.master.normalizer.RegionNormalizerWorker.run(RegionNormalizerWorker.java:191)
> > >>    
> > >> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> > >>    java.util.concurrent.FutureTask.run(FutureTask.java:266)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 141 (masterserver:16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20c47452
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> org.apache.hadoop.hbase.master.assignment.AssignmentManager.waitOnAssignQueue(AssignmentManager.java:2195)
> > >>
> > >> org.apache.hadoop.hbase.master.assignment.AssignmentManager.processAssignQueue(AssignmentManager.java:2217)
> > >>
> > >> org.apache.hadoop.hbase.master.assignment.AssignmentManager.access$600(AssignmentManager.java:109)
> > >>
> > >> org.apache.hadoop.hbase.master.assignment.AssignmentManager$1.run(AssignmentManager.java:2157)
> > >> Thread 140 (ProcedureDispatcherTimeoutThread):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 48
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>    java.util.concurrent.DelayQueue.poll(DelayQueue.java:259)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.util.DelayedUtil.takeWithoutInterrupt(DelayedUtil.java:81)
> > >>
> > >> org.apache.hadoop.hbase.procedure2.RemoteProcedureDispatcher$TimeoutExecutorThread.run(RemoteProcedureDispatcher.java:320)
> > >> Thread 121 (Idle-Rpc-Conn-Sweeper-pool-0):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 8
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093)
> > >>
> > >> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 119 (master:store-Flusher):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 3
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2163)
> > >>
> > >> org.apache.hadoop.hbase.master.region.MasterRegionFlusherAndCompactor.flushLoop(MasterRegionFlusherAndCompactor.java:193)
> > >>
> > >> org.apache.hadoop.hbase.master.region.MasterRegionFlusherAndCompactor$$Lambda$433/1001275970.run(Unknown
> > >> Source)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 114 (AsyncFSWAL-0-hdfs://masterserver:9000/hbase/MasterData):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 3
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@4a093bfa
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 113 (Connector-Scheduler-5ec5ea63-1):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 36
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093)
> > >>
> > >> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 112 (prometheus-http-1-1):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 64
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f7dcfc6
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 111 (RS-EventLoopGroup-1-7):
> > >>  State: RUNNABLE
> > >>  Blocked count: 1
> > >>  Waited count: 26
> > >>  Stack:
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait0(Native
> > >> Method)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:182)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWait(EpollEventLoop.java:312)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:376)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 110 (RS-EventLoopGroup-1-6):
> > >>  State: RUNNABLE
> > >>  Blocked count: 0
> > >>  Waited count: 29
> > >>  Stack:
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native
> > >> Method)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:209)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:202)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:316)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:373)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 109 (RS-EventLoopGroup-1-5):
> > >>  State: RUNNABLE
> > >>  Blocked count: 0
> > >>  Waited count: 54
> > >>  Stack:
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait0(Native
> > >> Method)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:182)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWait(EpollEventLoop.java:312)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:376)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 108 (RS-EventLoopGroup-1-4):
> > >>  State: RUNNABLE
> > >>  Blocked count: 98
> > >>  Waited count: 38
> > >>  Stack:
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native
> > >> Method)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:209)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:202)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:316)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:373)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 107 (RS-EventLoopGroup-1-3):
> > >>  State: RUNNABLE
> > >>  Blocked count: 130
> > >>  Waited count: 39
> > >>  Stack:
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native
> > >> Method)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:209)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:202)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:316)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:373)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 106 (RS-EventLoopGroup-1-2):
> > >>  State: RUNNABLE
> > >>  Blocked count: 51
> > >>  Waited count: 7
> > >>  Stack:
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native
> > >> Method)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:209)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:202)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:316)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:373)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 105 (master:store-WAL-Roller):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 3
> > >>  Waited count: 104
> > >>  Stack:
> > >>    java.lang.Object.wait(Native Method)
> > >>
> > >> org.apache.hadoop.hbase.wal.AbstractWALRoller.run(AbstractWALRoller.java:179)
> > >> Thread 103 (LeaseRenewer:seritrack@masterserver:9000):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 31
> > >>  Waited count: 1016
> > >>  Stack:
> > >>    java.lang.Thread.sleep(Native Method)
> > >>
> > >> org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:412)
> > >>
> > >> org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$600(LeaseRenewer.java:76)
> > >>
> > >> org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:308)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 100 (org.apache.hadoop.hdfs.PeerCache@2af0ac32):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 319
> > >>  Stack:
> > >>    java.lang.Thread.sleep(Native Method)
> > >>    org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253)
> > >>    org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46)
> > >>    org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 99 (IPC Parameter Sending Thread #0):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 1
> > >>  Waited count: 204
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:460)
> > >>
> > >> java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:362)
> > >>    java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:941)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1073)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 97 (master/masterserver:16000:becomeActiveMaster-MemStoreChunkPool
> > >> Statistics):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 4
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093)
> > >>
> > >> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 95 (master/masterserver:16000:becomeActiveMaster-MemStoreChunkPool
> > >> Statistics):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 4
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093)
> > >>
> > >> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 93 (Monitor thread for TaskMonitor):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 96
> > >>  Stack:
> > >>    java.lang.Thread.sleep(Native Method)
> > >>
> > >> org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 92 (master/masterserver:16000:becomeActiveMaster):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 165
> > >>  Waited count: 404
> > >>  Stack:
> > >>    java.lang.Thread.sleep(Native Method)
> > >>    org.apache.hadoop.hbase.util.Threads.sleep(Threads.java:125)
> > >>    
> > >> org.apache.hadoop.hbase.master.HMaster.isRegionOnline(HMaster.java:1358)
> > >>
> > >> org.apache.hadoop.hbase.master.HMaster.waitForMetaOnline(HMaster.java:1328)
> > >>
> > >> org.apache.hadoop.hbase.master.HMaster.finishActiveMasterInitialization(HMaster.java:1069)
> > >>
> > >> org.apache.hadoop.hbase.master.HMaster.startActiveMasterManager(HMaster.java:2405)
> > >>    org.apache.hadoop.hbase.master.HMaster.lambda$null$0(HMaster.java:565)
> > >>
> > >> org.apache.hadoop.hbase.master.HMaster$$Lambda$265/1598878738.run(Unknown
> > >> Source)
> > >>    org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187)
> > >>    org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177)
> > >>    org.apache.hadoop.hbase.master.HMaster.lambda$run$1(HMaster.java:562)
> > >>
> > >> org.apache.hadoop.hbase.master.HMaster$$Lambda$264/1129144214.run(Unknown
> > >> Source)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 19 (master/masterserver:16000):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 5
> > >>  Waited count: 322
> > >>  Stack:
> > >>    java.lang.Object.wait(Native Method)
> > >>    org.apache.hadoop.hbase.util.Sleeper.sleep(Sleeper.java:81)
> > >>    org.apache.hadoop.hbase.util.Sleeper.sleep(Sleeper.java:64)
> > >>
> > >> org.apache.hadoop.hbase.master.HMaster.waitForMasterActive(HMaster.java:677)
> > >>
> > >> org.apache.hadoop.hbase.regionserver.HRegionServer.initializeZooKeeper(HRegionServer.java:999)
> > >>
> > >> org.apache.hadoop.hbase.regionserver.HRegionServer.preRegistrationInitialization(HRegionServer.java:942)
> > >>
> > >> org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:1048)
> > >>    org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:578)
> > >> Thread 91 (Session-HouseKeeper-513b52af-1):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 2
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093)
> > >>
> > >> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 90 (qtp85435056-90):
> > >>  State: RUNNABLE
> > >>  Blocked count: 0
> > >>  Waited count: 24
> > >>  Stack:
> > >>    sun.management.ThreadImpl.getThreadInfo1(Native Method)
> > >>    sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:185)
> > >>    sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:144)
> > >>
> > >> org.apache.hadoop.hbase.util.ReflectionUtils.printThreadInfo(ReflectionUtils.java:181)
> > >>    org.apache.hadoop.hbase.util.Threads.printThreadInfo(Threads.java:186)
> > >>
> > >> org.apache.hadoop.hbase.master.http.MasterDumpServlet.doGet(MasterDumpServlet.java:86)
> > >>    javax.servlet.http.HttpServlet.service(HttpServlet.java:687)
> > >>    javax.servlet.http.HttpServlet.service(HttpServlet.java:790)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:799)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.servlet.ServletHandler$ChainEnd.doFilter(ServletHandler.java:1656)
> > >>
> > >> org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter.doFilter(StaticUserWebFilter.java:117)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.servlet.FilterHolder.doFilter(FilterHolder.java:193)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.servlet.ServletHandler$Chain.doFilter(ServletHandler.java:1626)
> > >>
> > >> org.apache.hadoop.hbase.http.SecurityHeadersFilter.doFilter(SecurityHeadersFilter.java:65)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.servlet.FilterHolder.doFilter(FilterHolder.java:193)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.servlet.ServletHandler$Chain.doFilter(ServletHandler.java:1626)
> > >>
> > >> org.apache.hadoop.hbase.http.ClickjackingPreventionFilter.doFilter(ClickjackingPreventionFilter.java:49)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.servlet.FilterHolder.doFilter(FilterHolder.java:193)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.servlet.ServletHandler$Chain.doFilter(ServletHandler.java:1626)
> > >>
> > >> org.apache.hadoop.hbase.http.HttpServer$QuotingInputFilter.doFilter(HttpServer.java:1521)
> > >> Thread 89 (qtp85435056-89):
> > >>  State: RUNNABLE
> > >>  Blocked count: 2
> > >>  Waited count: 26
> > >>  Stack:
> > >>    sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
> > >>    sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269)
> > >>    sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93)
> > >>    sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86)
> > >>    sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97)
> > >>    sun.nio.ch.SelectorImpl.select(SelectorImpl.java:101)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.run(EatWhatYouKill.java:131)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:409)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 88 (qtp85435056-88):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 2
> > >>  Waited count: 24
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 87 (qtp85435056-87):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 8
> > >>  Waited count: 24
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:460)
> > >>
> > >> java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:362)
> > >>    java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:941)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.reservedWait(ReservedThreadExecutor.java:324)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:399)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 86 (qtp85435056-86):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 3
> > >>  Waited count: 27
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 85 (qtp85435056-85):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 3
> > >>  Waited count: 26
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 84 
> > >> (qtp85435056-84-acceptor-0@acc4c8d-ServerConnector@5ec5ea63{HTTP/1.1,
> > >> (http/1.1)}{0.0.0.0:16010}):
> > >>  State: RUNNABLE
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Stack:
> > >>    sun.nio.ch.ServerSocketChannelImpl.accept0(Native Method)
> > >>
> > >> sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:421)
> > >>
> > >> sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:249)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 83 (qtp85435056-83):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 6
> > >>  Waited count: 21
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974)
> > >>
> > >> org.apache.hbase.thirdparty.org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 81 
> > >> (RpcServer.metaPriority.FPBQ.Fifo.handler=0,queue=0,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@52fbca80
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 80 (RpcServer.replication.FPBQ.Fifo.handler=2,queue=0,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@2a42337c
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 79 (RpcServer.replication.FPBQ.Fifo.handler=1,queue=0,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@550383c4
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 78 (RpcServer.replication.FPBQ.Fifo.handler=0,queue=0,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@cdb4b1f
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 77 
> > >> (RpcServer.priority.RWQ.Fifo.read.handler=19,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@783fce
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 76 
> > >> (RpcServer.priority.RWQ.Fifo.read.handler=18,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@783fce
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 75 
> > >> (RpcServer.priority.RWQ.Fifo.read.handler=17,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@783fce
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 74 
> > >> (RpcServer.priority.RWQ.Fifo.read.handler=16,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@783fce
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 73 
> > >> (RpcServer.priority.RWQ.Fifo.read.handler=15,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@783fce
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 72 
> > >> (RpcServer.priority.RWQ.Fifo.read.handler=14,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@783fce
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 71 
> > >> (RpcServer.priority.RWQ.Fifo.read.handler=13,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@783fce
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 70 
> > >> (RpcServer.priority.RWQ.Fifo.read.handler=12,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@783fce
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 69 
> > >> (RpcServer.priority.RWQ.Fifo.read.handler=11,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@783fce
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 68 
> > >> (RpcServer.priority.RWQ.Fifo.read.handler=10,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@783fce
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 67 
> > >> (RpcServer.priority.RWQ.Fifo.read.handler=9,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@783fce
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 66 
> > >> (RpcServer.priority.RWQ.Fifo.read.handler=8,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@783fce
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 65 
> > >> (RpcServer.priority.RWQ.Fifo.read.handler=7,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@783fce
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 64 
> > >> (RpcServer.priority.RWQ.Fifo.read.handler=6,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@783fce
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 63 
> > >> (RpcServer.priority.RWQ.Fifo.read.handler=5,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@783fce
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 62 
> > >> (RpcServer.priority.RWQ.Fifo.read.handler=4,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@783fce
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 61 
> > >> (RpcServer.priority.RWQ.Fifo.read.handler=3,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@783fce
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 60 
> > >> (RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@783fce
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 59 
> > >> (RpcServer.priority.RWQ.Fifo.write.handler=1,queue=0,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 35
> > >>  Waited count: 837
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3b9a256e
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 58 
> > >> (RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 25
> > >>  Waited count: 819
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3b9a256e
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    
> > >> org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:68)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 57 (RpcServer.default.FPBQ.Fifo.handler=29,queue=2,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@724e8c7a
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 56 (RpcServer.default.FPBQ.Fifo.handler=28,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@249bb29f
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 55 (RpcServer.default.FPBQ.Fifo.handler=27,queue=0,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@655c7f8d
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 54 (RpcServer.default.FPBQ.Fifo.handler=26,queue=2,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@f87971d
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 53 (RpcServer.default.FPBQ.Fifo.handler=25,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@9ace7cb
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 52 (RpcServer.default.FPBQ.Fifo.handler=24,queue=0,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@705aa32b
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 51 (RpcServer.default.FPBQ.Fifo.handler=23,queue=2,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@767c0aba
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 50 (RpcServer.default.FPBQ.Fifo.handler=22,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@45704417
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 49 (RpcServer.default.FPBQ.Fifo.handler=21,queue=0,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@2b8c61cd
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 48 (RpcServer.default.FPBQ.Fifo.handler=20,queue=2,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@eeadc6c
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 47 (RpcServer.default.FPBQ.Fifo.handler=19,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@18552ed1
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 46 (RpcServer.default.FPBQ.Fifo.handler=18,queue=0,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@5fc29130
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 45 (RpcServer.default.FPBQ.Fifo.handler=17,queue=2,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@1708110c
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 44 (RpcServer.default.FPBQ.Fifo.handler=16,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@4f59a63d
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 43 (RpcServer.default.FPBQ.Fifo.handler=15,queue=0,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@7e4b2aa1
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 42 (RpcServer.default.FPBQ.Fifo.handler=14,queue=2,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@354b7f49
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 41 (RpcServer.default.FPBQ.Fifo.handler=13,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@629c889d
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 40 (RpcServer.default.FPBQ.Fifo.handler=12,queue=0,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@6076a2bd
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 39 (RpcServer.default.FPBQ.Fifo.handler=11,queue=2,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@474f8230
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 38 (RpcServer.default.FPBQ.Fifo.handler=10,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@29901ca7
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 37 (RpcServer.default.FPBQ.Fifo.handler=9,queue=0,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@2fceb168
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 36 (RpcServer.default.FPBQ.Fifo.handler=8,queue=2,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@7af89d66
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 35 (RpcServer.default.FPBQ.Fifo.handler=7,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@411dce0b
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 34 (RpcServer.default.FPBQ.Fifo.handler=6,queue=0,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@684187ab
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 33 (RpcServer.default.FPBQ.Fifo.handler=5,queue=2,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@325e2e3a
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 32 (RpcServer.default.FPBQ.Fifo.handler=4,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@cf386ba
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 31 (RpcServer.default.FPBQ.Fifo.handler=3,queue=0,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@23b17cb9
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 30 (RpcServer.default.FPBQ.Fifo.handler=2,queue=2,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@4729dbbb
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 29 (RpcServer.default.FPBQ.Fifo.handler=1,queue=1,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@445129a
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 28 (RpcServer.default.FPBQ.Fifo.handler=0,queue=0,port=16000):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 1
> > >>  Waiting on java.util.concurrent.Semaphore$NonfairSync@141392fe
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> > >>    java.util.concurrent.Semaphore.acquire(Semaphore.java:312)
> > >>
> > >> org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55)
> > >>    org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:82)
> > >> Thread 27 (zk-event-processor-pool-0):
> > >>  State: WAITING
> > >>  Blocked count: 10
> > >>  Waited count: 16
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3cef7071
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 26 (main-EventThread):
> > >>  State: WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 8
> > >>  Waiting on
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@35929054
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
> > >>
> > >> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
> > >>    org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:549)
> > >> Thread 25 (main-SendThread(masterserver:2181)):
> > >>  State: RUNNABLE
> > >>  Blocked count: 3
> > >>  Waited count: 0
> > >>  Stack:
> > >>    sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
> > >>    sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269)
> > >>    sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93)
> > >>    sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86)
> > >>    sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97)
> > >>
> > >> org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:332)
> > >>    org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1289)
> > >> Thread 23
> > >> (org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner):
> > >>  State: WAITING
> > >>  Blocked count: 3
> > >>  Waited count: 4
> > >>  Waiting on java.lang.ref.ReferenceQueue$Lock@19ef85be
> > >>  Stack:
> > >>    java.lang.Object.wait(Native Method)
> > >>    java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:144)
> > >>    java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:165)
> > >>
> > >> org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:3712)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 22 (RS-EventLoopGroup-1-1):
> > >>  State: RUNNABLE
> > >>  Blocked count: 21
> > >>  Waited count: 0
> > >>  Stack:
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native
> > >> Method)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:209)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:202)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:316)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:373)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
> > >>
> > >> org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 21 (HBase-Metrics2-1):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 96
> > >>  Stack:
> > >>    sun.misc.Unsafe.park(Native Method)
> > >>    java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
> > >>
> > >> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2078)
> > >>
> > >> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1093)
> > >>
> > >> java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
> > >>
> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 20 (Timer for 'HBase' metrics system):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 96
> > >>  Stack:
> > >>    java.lang.Object.wait(Native Method)
> > >>    java.util.TimerThread.mainLoop(Timer.java:552)
> > >>    java.util.TimerThread.run(Timer.java:505)
> > >> Thread 16 (RMI TCP Accept-0):
> > >>  State: RUNNABLE
> > >>  Blocked count: 0
> > >>  Waited count: 0
> > >>  Stack:
> > >>    java.net.PlainSocketImpl.socketAccept(Native Method)
> > >>
> > >> java.net.AbstractPlainSocketImpl.accept(AbstractPlainSocketImpl.java:409)
> > >>    java.net.ServerSocket.implAccept(ServerSocket.java:560)
> > >>    java.net.ServerSocket.accept(ServerSocket.java:528)
> > >>
> > >> sun.management.jmxremote.LocalRMIServerSocketFactory$1.accept(LocalRMIServerSocketFactory.java:52)
> > >>
> > >> sun.rmi.transport.tcp.TCPTransport$AcceptLoop.executeAcceptLoop(TCPTransport.java:405)
> > >>    
> > >> sun.rmi.transport.tcp.TCPTransport$AcceptLoop.run(TCPTransport.java:377)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 15 (RMI TCP Accept-10101):
> > >>  State: RUNNABLE
> > >>  Blocked count: 0
> > >>  Waited count: 0
> > >>  Stack:
> > >>    java.net.PlainSocketImpl.socketAccept(Native Method)
> > >>
> > >> java.net.AbstractPlainSocketImpl.accept(AbstractPlainSocketImpl.java:409)
> > >>    java.net.ServerSocket.implAccept(ServerSocket.java:560)
> > >>    java.net.ServerSocket.accept(ServerSocket.java:528)
> > >>
> > >> sun.rmi.transport.tcp.TCPTransport$AcceptLoop.executeAcceptLoop(TCPTransport.java:405)
> > >>    
> > >> sun.rmi.transport.tcp.TCPTransport$AcceptLoop.run(TCPTransport.java:377)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 10 (Thread-3):
> > >>  State: RUNNABLE
> > >>  Blocked count: 2
> > >>  Waited count: 0
> > >>  Stack:
> > >>    sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
> > >>    sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269)
> > >>    sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93)
> > >>    sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86)
> > >>    sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97)
> > >>    sun.net.httpserver.ServerImpl$Dispatcher.run(ServerImpl.java:453)
> > >>    java.lang.Thread.run(Thread.java:750)
> > >> Thread 8 (req-rsp-timeout-task):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 1
> > >>  Waited count: 960
> > >>  Stack:
> > >>    java.lang.Object.wait(Native Method)
> > >>    java.util.TimerThread.mainLoop(Timer.java:552)
> > >>    java.util.TimerThread.run(Timer.java:505)
> > >> Thread 7 (idle-timeout-task):
> > >>  State: TIMED_WAITING
> > >>  Blocked count: 0
> > >>  Waited count: 96
> > >>  Stack:
> > >>    java.lang.Object.wait(Native Method)
> > >>    java.util.TimerThread.mainLoop(Timer.java:552)
> > >>    java.util.TimerThread.run(Timer.java:505)
> > >> Thread 5 (Signal Dispatcher):
> > >>  State: RUNNABLE
> > >>  Blocked count: 0
> > >>  Waited count: 0
> > >>  Stack:
> > >> Thread 3 (Finalizer):
> > >>  State: WAITING
> > >>  Blocked count: 92
> > >>  Waited count: 24
> > >>  Waiting on java.lang.ref.ReferenceQueue$Lock@27c20538
> > >>  Stack:
> > >>    java.lang.Object.wait(Native Method)
> > >>    java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:144)
> > >>    java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:165)
> > >>    java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:188)
> > >> Thread 2 (Reference Handler):
> > >>  State: WAITING
> > >>  Blocked count: 52
> > >>  Waited count: 23
> > >>  Waiting on java.lang.ref.Reference$Lock@72d818d1
> > >>  Stack:
> > >>    java.lang.Object.wait(Native Method)
> > >>    java.lang.Object.wait(Object.java:502)
> > >>    java.lang.ref.Reference.tryHandlePending(Reference.java:191)
> > >>    java.lang.ref.Reference$ReferenceHandler.run(Reference.java:153)
> > >> Thread 1 (main):
> > >>  State: WAITING
> > >>  Blocked count: 20
> > >>  Waited count: 19
> > >>  Waiting on org.apache.hadoop.hbase.master.HMaster@4772c3a0
> > >>  Stack:
> > >>    java.lang.Object.wait(Native Method)
> > >>    java.lang.Thread.join(Thread.java:1257)
> > >>    java.lang.Thread.join(Thread.java:1331)
> > >>
> > >> org.apache.hadoop.hbase.master.HMasterCommandLine.startMaster(HMasterCommandLine.java:254)
> > >>
> > >> org.apache.hadoop.hbase.master.HMasterCommandLine.run(HMasterCommandLine.java:147)
> > >>    org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)
> > >>
> > >> org.apache.hadoop.hbase.util.ServerCommandLine.doMain(ServerCommandLine.java:140)
> > >>    org.apache.hadoop.hbase.master.HMaster.main(HMaster.java:3311)
> > >>
> > >>
> > >>
> > >> Am Sa., 20. Apr. 2024 um 15:11 Uhr schrieb 张铎(Duo Zhang) <
> > >> palomino...@gmail.com>:
> > >>
> > >>> Just post it somewhere so we can check it.
> > >>>
> > >>> Udo Offermann <udo.offerm...@zfabrik.de> 于2024年4月20日周六 20:25写道:
> > >>>>
> > >>>> I do have the dump File from the web ui. I can sende it all or you Tell
> > >>> me
> > >>>> threads you are interessted in. Fortunately they all have meaningfull
> > >>> named.
> > >>>>
> > >>>> 张铎(Duo Zhang) <palomino...@gmail.com> schrieb am Sa., 20. Apr. 2024,
> > >>> 14:13:
> > >>>>
> > >>>>> What is the jstack result for HMaster while hanging? Wait on the
> > >>>>> namespace table online or meta table online?
> > >>>>>
> > >>>>> Udo Offermann <udo.offerm...@zfabrik.de> 于2024年4月20日周六 19:43写道:
> > >>>>>>
> > >>>>>> Hello everyone,
> > >>>>>>
> > >>>>>> We are upgrading our Hadoop/HBase cluster from Hadoop 2.8.5 & HBase
> > >>> 2.2.5
> > >>>>>> to Hadoop 3.3.6 & HBase 2.5.7
> > >>>>>>
> > >>>>>> The Hadoop upgrade worked well, but unfortunately we have problems
> > >>> with
> > >>>>> the
> > >>>>>> Hbase upgrade, because the master hangs on startup inside the
> > >>> „Starting
> > >>>>>> assignment manger“ task.
> > >>>>>>
> > >>>>>> After 15 minutes the following message appears in the log file:
> > >>>>>>
> > >>>>>> Master failed to complete initialization after 900000ms. Please
> > >>>>>> consider submitting a bug report including a thread dump of this
> > >>>>>> process.
> > >>>>>>
> > >>>>>>
> > >>>>>> We face the same problem as Adam a couple of weeks ago: "Rolling
> > >>> upgrade
> > >>>>>> from HBase 2.2.2 to 2.5.8 [typo corrected]: There are 2336 corrupted
> > >>>>>> procedures“ and we fixed it in the same way by deleting the
> > >>>>>> MasterProcWALs-folder
> > >>>>>> in HDFS.
> > >>>>>>
> > >>>>>> I can provide HMaster dump and a dump of one data nodes!
> > >>>>>>
> > >>>>>> How can we proceed with the upgrade?
> > >>>>>>
> > >>>>>> Thanks and best regards
> > >>>>>> Udo
> > >>>>>
> > >>>
> > >>
> > >>
> > >> --
> > >> Udo Offermann
> > >>
> > >> udo.offerm...@zfabrik.de
> > >>
> > >> ZFabrik <http://www.zfabrik.de/>
> > >> Blog <http://www.z2-environment.net/blog>
> > >> Z2-Environment <http://www.z2-environment.eu/>
> > >> Z2 Wiki <http://redmine.z2-environment.net/>
> > >> T: +49 6227 3984255
> > >> F: +49 6227 3984254
> > >> M: +49 1781891820
> > >>
> > >> *ZFabrik Software GmbH & Co. KG*
> > >> Lammstrasse 2, 69190 Walldorf
> > >> Handelsregister: Amtsgericht Mannheim HRA 702598
> > >> Persönlich haftende Gesellschafterin: ZFabrik Verwaltungs GmbH, Sitz
> > >> Walldorf
> > >> Geschäftsführer: Dr. H. Blohm u. Udo Offermann
> > >> Handelsregister: Amtsgericht Mannheim HRB 723699
> >

Reply via email to