Hi Deepak! The "slaves" file is only used by the SSH script to start a standalone cluster.
As Max said, TaskManagers register dynamically at the JobManager. Discovery works via: - config in non-HA mode - ZooKeeper in HA mode On Wed, Feb 17, 2016 at 10:11 AM, Maximilian Michels <m...@apache.org> wrote: > Hi Deepak, > > The job manager doesn't have to know about task managers. They will > simply register at the job manager using the provided configuration. > In HA mode, they will lookup the currently leading job manager first > and then connect to it. The job manager can then assign work. > > Cheers, > Max > > On Tue, Feb 16, 2016 at 10:41 PM, Deepak Jha <dkjhan...@gmail.com> wrote: > > Hi All, > > I have a question on scaling-up/scaling-down flink cluster. > > As per the documentation, in order to scale-up the cluster, I can add a > new > > taskmanager on the fly and jobmanager can assign work to it. Assuming, I > > have Flink HA , so in the event of master JobManager failure, how is this > > taskmanager detail is going to get transferred ? I believe new master > will > > just read the contents from slaves config file. Can anyone give more > > clarity on how this is done ? Or, Is it union of slaves and the > > taskmanager's that are added on the fly ? > > > > -- > > Thanks, > > Deepak >