Hi, 

Just for completeness sake: I followed the advice of Gyula and now everything 
seems to work. 
I changed my HashSet to managed MapState of Flink and changed my LinkedList to 
a custom class of mine which is recognized by Flink as POJOType and thus don't 
use Kryo for serialization. This new version survived a few stress tests 
already. Seems really to be some kryo / race condition which is at best avoided 
by avoiding kryo :) 

Best regards 
Theo 


Von: "Gyula Fóra" <gyf...@apache.org> 
An: "Theo Diefenthal" <theo.diefent...@scoop-software.de> 
CC: "user" <user@flink.apache.org> 
Gesendet: Freitag, 29. November 2019 10:53:20 
Betreff: Re: ArrayIndexOutOfBoundException on checkpoint creation 

Hi Theo! 

I have not seen this error before however I have encountered many strange 
things when using Kryo for serialization. From the stack trace it seems that 
this might indeed be a Kryo related issue. 

I am not sure what it is but what I would try is to change the state 
serializers to a non Kryo variants. 

When you create the state you can specify the TypeInformation object for the 
state instead of the class. You have 2 states: 

For LinkedList: 
Use the ListTypeInfo class and change your state type to List 

For HashSet: 
There is no Set Type info built into Flink but you can replace it with a Map 
and use the MapTypeInfo to try it out, before implementing a custom type info 

Also make sure that your MyPOJO type is a nice flink supported pojo. All public 
non-final fields + empty constructor. 

Let me know if this makes sense :) 

Cheers, 
Gyula 

On Wed, Nov 27, 2019 at 2:18 PM [ mailto:theo.diefent...@scoop-software.de | 
theo.diefent...@scoop-software.de ] < [ 
mailto:theo.diefent...@scoop-software.de | theo.diefent...@scoop-software.de ] 
> wrote: 


Sorry, I forgot to mention the environment. 
We use Flink 1.9.1 on a cloudera cdh6. 3.1 cluster (with Hadoop 3.0.0 but using 
Flink shaded 2.8.3-7. Might this be a problem? As it seems to arise from kryo, 
I doubt it) 
Our flink is configured as default. Our job uses FsStateBackend and exactly 
once processing with Kafka source and sink. 
Best regardsTheo 
-------- Ursprüngliche Nachricht -------- 
Betreff: ArrayIndexOutOfBoundException on checkpoint creation 
Von: Theo Diefenthal 
An: user 
Cc: 


Hi, 

We have a pipeline with a custom ProcessFunction and state (see [1], 
implemented as suggested by Fabian with a ValueState<LinkedList<MyPOJO> and 
ValueState<HashSet<String>>) 
The behavior of that function works fine in our unittests and with low load in 
our test environment (100.000 records per minute). On the production 
environment, we observe reproduceable crashes like the attached one. 
Any idea on why this out of bound could be caused? Every time we read the state 
and modify it, we are certain that an .update() was called: 

2019-11-26T11:26:55+01:00 host19 java.lang.Exception: Could not materialize 
checkpoint 7 for operator our_operator) (4/8). 
2019-11-26T11:26:55+01:00 host19 at 
org.apache.flink.streaming.runtime.tasks.StreamTask$AsyncCheckpointRunnable.handleExecutionException(StreamTask.java:1100)
 
2019-11-26T11:26:55+01:00 host19 at 
org.apache.flink.streaming.runtime.tasks.StreamTask$AsyncCheckpointRunnable.run(StreamTask.java:1042)
 
2019-11-26T11:26:55+01:00 host19 at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) 
2019-11-26T11:26:55+01:00 host19 at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) 
2019-11-26T11:26:55+01:00 host19 at java.lang.Thread.run(Thread.java:745) 
2019-11-26T11:26:55+01:00 host19 Caused by: 
java.util.concurrent.ExecutionException: 
java.lang.ArrayIndexOutOfBoundsException: 67108864 
2019-11-26T11:26:55+01:00 host19 at 
java.util.concurrent.FutureTask.report(FutureTask.java:122) 
2019-11-26T11:26:55+01:00 host19 at 
java.util.concurrent.FutureTask.get(FutureTask.java:192) 
2019-11-26T11:26:55+01:00 host19 at 
org.apache.flink.runtime.concurrent.FutureUtils.runIfNotDoneAndGet(FutureUtils.java:450)
 
2019-11-26T11:26:55+01:00 host19 at 
org.apache.flink.streaming.api.operators.OperatorSnapshotFinalizer.<init>(OperatorSnapshotFinalizer.java:47)
 
2019-11-26T11:26:55+01:00 host19 at 
org.apache.flink.streaming.runtime.tasks.StreamTask$AsyncCheckpointRunnable.run(StreamTask.java:1011)
 
2019-11-26T11:26:55+01:00 host19 ... 3 more 
2019-11-26T11:26:55+01:00 host19 Caused by: 
java.lang.ArrayIndexOutOfBoundsException: 67108864 
2019-11-26T11:26:55+01:00 host19 at 
com.esotericsoftware.kryo.util.IdentityObjectIntMap.clear(IdentityObjectIntMap.java:364)
 
2019-11-26T11:26:55+01:00 host19 at 
com.esotericsoftware.kryo.util.MapReferenceResolver.reset(MapReferenceResolver.java:47)
 
2019-11-26T11:26:55+01:00 host19 at 
com.esotericsoftware.kryo.Kryo.reset(Kryo.java:836) 
2019-11-26T11:26:55+01:00 host19 at 
com.esotericsoftware.kryo.Kryo.writeClassAndObject(Kryo.java:601) 
2019-11-26T11:26:55+01:00 host19 at 
org.apache.flink.api.java.typeutils.runtime.kryo.KryoSerializer.serialize(KryoSerializer.java:305)
 
2019-11-26T11:26:55+01:00 host19 at 
org.apache.flink.runtime.state.heap.CopyOnWriteStateMapSnapshot.writeState(CopyOnWriteStateMapSnapshot.java:116)
 
2019-11-26T11:26:55+01:00 host19 at 
org.apache.flink.runtime.state.heap.AbstractStateTableSnapshot.writeStateInKeyGroup(AbstractStateTableSnapshot.java:121)
 
2019-11-26T11:26:55+01:00 host19 at 
org.apache.flink.runtime.state.heap.CopyOnWriteStateTableSnapshot.writeStateInKeyGroup(CopyOnWriteStateTableSnapshot.java:37)
 
2019-11-26T11:26:55+01:00 host19 at 
org.apache.flink.runtime.state.heap.HeapSnapshotStrategy$1.callInternal(HeapSnapshotStrategy.java:191)
 
2019-11-26T11:26:55+01:00 host19 at 
org.apache.flink.runtime.state.heap.HeapSnapshotStrategy$1.callInternal(HeapSnapshotStrategy.java:158)
 
2019-11-26T11:26:55+01:00 host19 at 
org.apache.flink.runtime.state.AsyncSnapshotCallable.call(AsyncSnapshotCallable.java:75)
 
2019-11-26T11:26:55+01:00 host19 at 
java.util.concurrent.FutureTask.run(FutureTask.java:266) 
2019-11-26T11:26:55+01:00 host19 at 
org.apache.flink.runtime.concurrent.FutureUtils.runIfNotDoneAndGet(FutureUtils.java:447)
 
2019-11-26T11:26:55+01:00 host19 ... 5 more 
2019-11-26T11:26:55+01:00 host18 WARN org.apache.hadoop.hdfs.DataStreamer - 
DataStreamer Exception 
2019-11-26T11:26:55+01:00 host18 java.io.FileNotFoundException: File does not 
exist: 
/.../STATE/CHECKPOINTS/0a2e111b3a800aae0d3b49f33e0db6f3/chk-7/3da2a0a4-f5ef-4e8c-bc1a-9fe892cb0b18
 (inode 577546140) Holder DFSClient_NONMAPREDUCE_-1714419242_95 does not have 
any open files. 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:2782)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.analyzeFileState(FSDirWriteFileOp.java:599)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.validateAddBlock(FSDirWriteFileOp.java:171)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2661)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:872)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:550)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:869) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:815) 
2019-11-26T11:26:55+01:00 host18 at 
java.security.AccessController.doPrivileged(Native Method) 
2019-11-26T11:26:55+01:00 host18 at 
javax.security.auth.Subject.doAs(Subject.java:422) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.Server$Handler.run(Server.java:2675) 
2019-11-26T11:26:55+01:00 host18 
2019-11-26T11:26:55+01:00 host18 at 
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) 
2019-11-26T11:26:55+01:00 host18 at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
 
2019-11-26T11:26:55+01:00 host18 at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
 
2019-11-26T11:26:55+01:00 host18 at 
java.lang.reflect.Constructor.newInstance(Constructor.java:423) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:121)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:88)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.DataStreamer.locateFollowingBlock(DataStreamer.java:1842)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.DataStreamer.nextBlockOutputStream(DataStreamer.java:1638)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:704) 
2019-11-26T11:26:55+01:00 host18 Caused by: 
org.apache.hadoop.ipc.RemoteException( [ http://java.io/ | java.io ] 
.FileNotFoundException): File does not exist: 
/.../STATE/INGESTOR/CHECKPOINTS/0a2e111b3a800aae0d3b49f33e0db6f3/chk-7/3da2a0a4-f5ef-4e8c-bc1a-9fe892cb0b18
 (inode 577546140) Holder DFSClient_NONMAPREDUCE_-1714419242_95 does not have 
any open files. 

2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:2782)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.analyzeFileState(FSDirWriteFileOp.java:599)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.validateAddBlock(FSDirWriteFileOp.java:171)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2661)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:872)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:550)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:869) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:815) 
2019-11-26T11:26:55+01:00 host18 at 
java.security.AccessController.doPrivileged(Native Method) 
2019-11-26T11:26:55+01:00 host18 at 
javax.security.auth.Subject.doAs(Subject.java:422) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.Server$Handler.run(Server.java:2675) 
2019-11-26T11:26:55+01:00 host18 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1489) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.Client.call(Client.java:1435) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.Client.call(Client.java:1345) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:227)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
 
2019-11-26T11:26:55+01:00 host18 at com.sun.proxy.$Proxy19.addBlock(Unknown 
Source) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:444)
 
2019-11-26T11:26:55+01:00 host18 at 
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 
2019-11-26T11:26:55+01:00 host18 at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
2019-11-26T11:26:55+01:00 host18 at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 
2019-11-26T11:26:55+01:00 host18 at 
java.lang.reflect.Method.invoke(Method.java:498) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:409)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:163)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:155)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:346)
 
2019-11-26T11:26:55+01:00 host18 at com.sun.proxy.$Proxy20.addBlock(Unknown 
Source) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.DataStreamer.locateFollowingBlock(DataStreamer.java:1838)
 
2019-11-26T11:26:55+01:00 host18 ... 2 more 
2019-11-26T11:26:55+01:00 host18 WARN 
org.apache.flink.runtime.state.filesystem.FsCheckpointStreamFactory - Could not 
close the state stream for 
hdfs:/.../STATE/INGESTOR/CHECKPOINTS/0a2e111b3a800aae0d3b49f33e0db6f3/chk-7/3da2a0a4-f5ef-4e8c-bc1a-9fe892cb0b18.
 
2019-11-26T11:26:55+01:00 host18 java.io.FileNotFoundException: File does not 
exist: 
/.../STATE/INGESTOR/CHECKPOINTS/0a2e111b3a800aae0d3b49f33e0db6f3/chk-7/3da2a0a4-f5ef-4e8c-bc1a-9fe892cb0b18
 (inode 577546140) Holder DFSClient_NONMAPREDUCE_-1714419242_95 does not have 
any open files. 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:2782)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.analyzeFileState(FSDirWriteFileOp.java:599)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.validateAddBlock(FSDirWriteFileOp.java:171)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2661)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:872)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:550)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:869) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:815) 
2019-11-26T11:26:55+01:00 host18 at 
java.security.AccessController.doPrivileged(Native Method) 
2019-11-26T11:26:55+01:00 host18 at 
javax.security.auth.Subject.doAs(Subject.java:422) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.Server$Handler.run(Server.java:2675) 
2019-11-26T11:26:55+01:00 host18 
2019-11-26T11:26:55+01:00 host18 at 
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) 
2019-11-26T11:26:55+01:00 host18 at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
 
2019-11-26T11:26:55+01:00 host18 at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
 

2019-11-26T11:26:55+01:00 host18 at 
java.lang.reflect.Constructor.newInstance(Constructor.java:423) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:121)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:88)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.DataStreamer.locateFollowingBlock(DataStreamer.java:1842)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.DataStreamer.nextBlockOutputStream(DataStreamer.java:1638)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:704) 
2019-11-26T11:26:55+01:00 host18 Caused by: 
org.apache.hadoop.ipc.RemoteException( [ http://java.io/ | java.io ] 
.FileNotFoundException): File does not exist: 
/.../STATE/INGESTOR/CHECKPOINTS/0a2e111b3a800aae0d3b49f33e0db6f3/chk-7/3da2a0a4-f5ef-4e8c-bc1a-9fe892cb0b18
 (inode 577546140) Holder DFSClient_NONMAPREDUCE_-1714419242_95 does not have 
any open files. 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:2782)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.analyzeFileState(FSDirWriteFileOp.java:599)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.validateAddBlock(FSDirWriteFileOp.java:171)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2661)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:872)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:550)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:869) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:815) 
2019-11-26T11:26:55+01:00 host18 at 
java.security.AccessController.doPrivileged(Native Method) 
2019-11-26T11:26:55+01:00 host18 at 
javax.security.auth.Subject.doAs(Subject.java:422) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.Server$Handler.run(Server.java:2675) 
2019-11-26T11:26:55+01:00 host18 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1489) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.Client.call(Client.java:1435) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.Client.call(Client.java:1345) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:227)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
 
2019-11-26T11:26:55+01:00 host18 at com.sun.proxy.$Proxy19.addBlock(Unknown 
Source) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:444)
 
2019-11-26T11:26:55+01:00 host18 at 
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 
2019-11-26T11:26:55+01:00 host18 at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
2019-11-26T11:26:55+01:00 host18 at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 
2019-11-26T11:26:55+01:00 host18 at 
java.lang.reflect.Method.invoke(Method.java:498) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:409)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:163)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:155)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:346)
 
2019-11-26T11:26:55+01:00 host18 at com.sun.proxy.$Proxy20.addBlock(Unknown 
Source) 
2019-11-26T11:26:55+01:00 host18 at 
org.apache.hadoop.hdfs.DataStreamer.locateFollowingBlock(DataStreamer.java:1838)
 
2019-11-26T11:26:55+01:00 host18 ... 2 more 
2019-11-26T11:29:50+01:00 host19 INFO org.apache.flink.yarn.YarnResourceManager 
- The heartbeat of TaskManager with id 
container_e137_1574681304655_2382_01_000007 timed out. 

Best regards 
Theo 

[1] [ 
http://mail-archives.apache.org/mod_mbox/flink-user/201909.mbox/%3c1340607597.13576306.1568278936983.javamail.zim...@scoop-software.de%3E
 | 
http://mail-archives.apache.org/mod_mbox/flink-user/201909.mbox/%3c1340607597.13576306.1568278936983.javamail.zim...@scoop-software.de%3E
 ] 





-- 
SCOOP Software GmbH - Gut Maarhausen - Eiler Straße 3 P - D-51107 Köln 
Theo Diefenthal 

T +49 221 801916-196 - F +49 221 801916-17 - M +49 160 90506575 
theo.diefent...@scoop-software.de - www.scoop-software.de 
Sitz der Gesellschaft: Köln, Handelsregister: Köln, 
Handelsregisternummer: HRB 36625 
Geschäftsführung: Dr. Oleg Balovnev, Frank Heinen, 
Martin Müller-Rohde, Dr. Wolfgang Reddig, Roland Scheel 

Reply via email to