It is certainly possible, but when you got dozens of workers, that would
take a very long time, specially if you got a lot of state, as partitions
get reassigned and state moved about.  In fact, it is likely to fail at
some point, as local state that can be stored in a multitude of nodes may
not be able to be stored locally as the number of nodes becomes smaller.

On Wed, Feb 8, 2017 at 12:34 PM, Dmitry Minkovsky <dminkov...@gmail.com>
wrote:

> Can you take them down sequentially? Like, say, with a Kubernetes
> StatefulSet
> <https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-
> set/#ordered-pod-termination>
> .
>
> On Wed, Feb 8, 2017 at 2:15 PM, Elias Levy <fearsome.lucid...@gmail.com>
> wrote:
>
> > What are folks doing to cleanly shutdown a Streams job comprised of
> > multiple workers?
> >
> > Right now I am doing sys.addShutdownHook(streams.close()) but that is
> not
> > working well to shutdown a fleet of workers.  When I signal the fleet to
> > shutdown by sending them all a SIGTERM, some of them will shutdown, but
> > some will persist.  It appears that there is a race condition between the
> > shutdown signal and a rebalancing occurring as a result of other workers
> > shutting down.  If a worker has not started shutting down before the
> > rebalancing starts, the rebalancing will cause the worker to not
> shutdown.
> >
> > Others seeing the same thing?
> >
>

Reply via email to