Dear all, Thank you for your help regarding my previous queries. Unfortunately, I'm stuck with another one and will really appreciate your input.
I can't seem to produce any outputs in "flink-taskexecutor-0.out" from my second job after submitting the first one in my 3-node-flink standalone cluster. Say I want to test out two jobs sequentially. (I do not want to run them concurrently/in parallel). After submitting "job1.jar " via command line, I press "Ctrl + C" to exit from it (as it runs infinitely). After that I try to submit a second jar file having the same properties (group-id, topic, etc) with the only difference being the query written in main function. The first job produces relevant outputs in "flink-taskexecutor-0.out" but the second one doesn't. The only way I can see the output produced is if I restart the cluster after job1 and then submit job2 as it produces another .out file. But I want to submit 2 jobs sequentially and see their outputs without having to restart my cluster. Is there any way to do this? Additional info: For both jobs I'm using DataStream API and I have set: StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment(); Best Regards, Komal