Hi Deepak!

The "slaves" file is only used by the SSH script to start a standalone
cluster.

As Max said, TaskManagers register dynamically at the JobManager.

Discovery works via:
   - config in non-HA mode
   - ZooKeeper in HA mode



On Wed, Feb 17, 2016 at 10:11 AM, Maximilian Michels <m...@apache.org> wrote:

> Hi Deepak,
>
> The job manager doesn't have to know about task managers. They will
> simply register at the job manager using the provided configuration.
> In HA mode, they will lookup the currently leading job manager first
> and then connect to it. The job manager can then assign work.
>
> Cheers,
> Max
>
> On Tue, Feb 16, 2016 at 10:41 PM, Deepak Jha <dkjhan...@gmail.com> wrote:
> > Hi All,
> > I have a question on scaling-up/scaling-down flink cluster.
> > As per the documentation, in order to scale-up the cluster, I can add a
> new
> > taskmanager on the fly and jobmanager can assign work to it. Assuming, I
> > have Flink HA , so in the event of master JobManager failure, how is this
> > taskmanager detail is going to get transferred ? I believe new master
> will
> > just read the contents from slaves config file. Can anyone give more
> > clarity on how this is done ? Or, Is it union of slaves and the
> > taskmanager's that are added on the fly ?
> >
> > --
> > Thanks,
> > Deepak
>

Reply via email to