Hello everyone,

I have a question concerning stopping Flink streaming processes that run
in a detached Yarn session.

Here's what we do: We start a Yarn session via
yarn-session.sh -n 8 -d -jm 4096 -tm 10000 -s 10 -qu flink_queue

Then, we start our Flink streaming application via
flink run -p 65 -c SomeClass some.jar > /dev/null 2>&1  &

The problem occurs when we stop the application.
If we stop the Flink application with
flink cancel <JOB_ID>
and then kill the yarn application with
yarn application -kill <APPLICATION_ID>
everything is fine.
But what we expected was that when we only kill the yarn application
without specifically canceling the Flink job before, the Flink job will
stay lingering on the machine and use resources until it is killed
manually via its process id.

One thing that we tried was to stop using ephemeral ports for the
application-manager, namely we set yarn.application-master.port
specifically to some port number, but the problem remains: Killing the
yarn application does not kill the corresponding Flink job.

Does anyone have an idea about this? Any help is greatly appreciated :-)
By the way, our application reads data from a Kafka queue and writes it
into HDFS, maybe this is also important to know.

Thank you and best regards

Konstantin
-- 
Konstantin Gregor * konstantin.gre...@tngtech.com
TNG Technology Consulting GmbH, Betastr. 13a, 85774 Unterföhring
Geschäftsführer: Henrik Klagges, Christoph Stock, Dr. Robert Dahlke
Sitz: Unterföhring * Amtsgericht München * HRB 135082

Reply via email to