Here is the exception and error log:

2018-05-29 14:41:04,762 WARN  org.apache.flink.runtime.taskmanager.Task         
            - Task 'over: (PARTITION BY: uid, ORDER BY: proctime, RANGEBETWEEN 
86400000 PRECEDI
NG AND CURRENT ROW, select: (id, uid, proctime, group_concat($7) AS w0$o0)) -> 
select:
(id, uid, proctime, w0$o0 AS EXPR$3) -> to: Row -> Flat Map -> Filter -> Sink: 
Unnamed (10/1
5)' did not react to cancelling signal for 30 seconds, but is stuck in method:
 
org.apache.flink.streaming.runtime.io.StreamInputProcessor.processInput(StreamInputProcessor.java:199)
org.apache.flink.streaming.runtime.tasks.OneInputStreamTask.run(OneInputStreamTask.java:103)
org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:306)
org.apache.flink.runtime.taskmanager.Task.run(Task.java:703)
java.lang.Thread.run(Thread.java:748)

...
...

2018-05-29 14:43:34,663 ERROR org.apache.flink.runtime.taskmanager.Task         
            - Task did not exit gracefully within 180 + seconds.
2018-05-29 14:43:34,663 ERROR 
org.apache.flink.runtime.taskexecutor.TaskExecutor            - Task did not 
exit gracefully within 180 + seconds.
2018-05-29 14:43:34,663 ERROR 
org.apache.flink.runtime.taskexecutor.TaskManagerRunner       - Fatal error 
occurred while executing the TaskManager. Shutting it down...
2018-05-29 14:43:34,666 INFO  
org.apache.flink.runtime.taskexecutor.TaskExecutor            - Stopping 
TaskExecutor akka.tcp://flink@fds-hadoop-prod04-mp:35187/user/taskmanager_0.
2018-05-29 14:43:34,669 INFO  
org.apache.flink.runtime.state.TaskExecutorLocalStateStoresManager  - Shutting 
down TaskExecutorLocalStateStoresManager.


________________________________
From: Yan Zhou [FDS Science] <yz...@coupang.com>
Sent: Tuesday, May 29, 2018 11:04:52 PM
To: user@flink.apache.org
Subject: Task did not exit gracefully and lost TaskManager


Hi,

when I stop my flink application in standalone cluster, one of the tasks can 
NOT exit gracefully. And the task managers are lost(or detached?). I can't see 
them in the web ui. However, the task managers are still running in the slave 
servers.


What could be the possible cause? My application run Over-window aggregation on 
a datastream table, the results are written into a custom mysql sink with 
org.apache.commons.dbcp.BasicDataSource. Close methods are called for the 
PreparedStatement,  Connection, and BasicDataSource within 
AbstractRichFunction::close()


Is it because of mysql jdbc doesn't handle interrupt propertly? Should I call 
PreparedStatement::cancel()? I find a similar issue here[1].  Thank you for 
your help.



[1] : 
https://stackoverflow.com/questions/40127228/flink-cannot-cancel-a-running-job-streaming


Best

Yan

[https://cdn.sstatic.net/Sites/stackoverflow/img/apple-touch-i...@2.png?v=73d79a89bded]<https://stackoverflow.com/questions/40127228/flink-cannot-cancel-a-running-job-streaming>

Flink : cannot cancel a running job (streaming) - Stack 
...<https://stackoverflow.com/questions/40127228/flink-cannot-cancel-a-running-job-streaming>
stackoverflow.com
I want to run a streaming job. When I try to run it locally using 
start-clusted.sh and the Flink Web Interface, I have no problem. However, I am 
currently trying to run my job using Flink on YARN


Reply via email to