Update: I deleted the /flink/recovery folder on HDFS and even then I get
the same Exception after the next checkpoint.
2016-11-21 21:51 GMT+01:00 static-max :
> Hi Stephan,
>
> it's not a problem, but makes finding other errors on my NameNode
> complicated as I have this er
behavior?
>
> Greetings,
> Stephan
>
>
> On Tue, Oct 11, 2016 at 5:25 PM, static-max
> wrote:
>
>> Hi,
>>
>> I get many (multiple times per minute) errors in my Namenode HDFS logfile:
>>
>> 2016-10-11 17:17:07,596 INF
Hi,
I get a ton of these messages in my Job Manager's logfile. This makes Flink
unstable, as I cannot list or cancel/stop the jobs.
I run Flink in YARN under a default Horton HDP 2.5 installation. HDP sets
the hard and soft limit of open files to 32768 for the user "yarn" that
runs the Flink JVMs,
for the error on Google.
>
> Did you check the GC times also for Flink? Is your Flink job doing any
> heavy tasks (like maintaining large windows, or other operations involving
> a lot of heap space?)
>
> Regards,
> Robert
>
>
> On Tue, Oct 11, 2016 at 10:51 AM, stati
Hi,
I get many (multiple times per minute) errors in my Namenode HDFS logfile:
2016-10-11 17:17:07,596 INFO ipc.Server (Server.java:logException(2401)) -
IPC Server handler 295 on 8020, call
org.apache.hadoop.hdfs.protocol.ClientProtocol.delete from datanode1:34872
Call#2361 Retry#0
org.apache.h
Hi,
I have a low throughput job (approx. 1000 messager per Minute), that
consumes from Kafka und writes directly to HDFS. After an hour or so, I get
the following warnings in the Task Manager log:
2016-10-10 01:59:44,635 WARN org.apache.hadoop.hdfs.DFSClient
- Slow ReadProcessor
I went back to commit 2afc092461cf68cf0f3c26a3ab4c58a7bd68cf71 on MASTER,
seems to work.
2016-10-05 15:48 GMT+02:00 static-max :
> Hi Stephan,
>
> thanks for the quick answer! I try to go to an older revision,
>
> Best,
> Max
>
> 2016-10-05 12:10 GMT+02:00 Stephan E
Hi Stephan,
thanks for the quick answer! I try to go to an older revision,
Best,
Max
2016-10-05 12:10 GMT+02:00 Stephan Ewen :
> Hi!
>
> The master has a temporary regression due to the Work In Progress for the
> "changing parallelism of savepoints" feature.
> We'll try and complete the change