Hello
I have set up a cluster and added taskmanagers manually with bin/taskmanager.sh 
start.
I noticed that if i have 5 task managers with one slot each and start a job 
with -p5, then if i stop a taskmanager the job will fail even if there are 4 
more taskmanagers.

Is this expected (I turned off restart policy)?
So the way to ensure continuous operation of a single "job" is to have e.g. 10 
TM and deploy 10 job instances to fill each of 10 slots?
Or if I have a job that does require -p3 for example, I should always have at 
least 3 TM alive?

Many thanks!
-Rob

Reply via email to