Hi Theo, I want to interrupt/cancel my current job as it has produced the desired results even though it runs infinitely, and the next one requires full resources.
Due to some technical issue we cannot access the web UI so just working with the CLI, for now. I found a less crude way by running the command ./bin/flink cancel <job id> specified by the commands listed here: https://ci.apache.org/projects/flink/flink-docs-stable/ops/cli.html ------------ Hello Vino, Thank you! That's exactly what's happening. Is there any way to force it write to a specific .out of a TaskManager? Best Regards, Komal On Mon, 25 Nov 2019 at 11:10, vino yang <yanghua1...@gmail.com> wrote: > Hi Komal, > > Since you use the Flink standalone deployment mode, the tasks of the jobs > which print information to the STDOUT may randomly deploy in any task > manager of the cluster. Did you check other Task Managers out file? > > Best, > Vino > > Komal Mariam <komal.mar...@gmail.com> 于2019年11月22日周五 下午6:59写道: > >> Dear all, >> >> Thank you for your help regarding my previous queries. Unfortunately, I'm >> stuck with another one and will really appreciate your input. >> >> I can't seem to produce any outputs in "flink-taskexecutor-0.out" from my >> second job after submitting the first one in my 3-node-flink standalone >> cluster. >> >> Say I want to test out two jobs sequentially. (I do not want to run them >> concurrently/in parallel). >> >> After submitting "job1.jar " via command line, I press "Ctrl + C" to exit >> from it (as it runs infinitely). After that I >> try to submit a second jar file having the same properties (group-id, >> topic, etc) with the only difference being the query written in main >> function. >> >> The first job produces relevant outputs in "flink-taskexecutor-0.out" but >> the second one doesn't. >> >> The only way I can see the output produced is if I restart the cluster >> after job1 and then submit job2 as it produces another .out file. >> >> But I want to submit 2 jobs sequentially and see their outputs without >> having to restart my cluster. Is there any way to do this? >> >> Additional info: >> For both jobs I'm using DataStream API and I have set: >> StreamExecutionEnvironment env = >> StreamExecutionEnvironment.getExecutionEnvironment(); >> >> Best Regards, >> Komal >> >