I think there is something to be said about making this distribution more flexible. A stand alone cluster, still the distribution mechanism for many a installations suffers horribly with the above approach. A healthy cluster requires resources wot be used equitably is possible. I have some pipes that are CPU intense and some not so much. Slots being the primary unit of parallelism does not differentiate that much between a stress inducing pipe or other wise and thus having them spread out becomes essential to avoid skews.
Maybe a simple setting could have forked different approaches to slot distribution. standalone = true for example. I have a n node cluster with 2 nodes being heavily stressed ( LA approaching the slot size ) b'coz of this new setup and that does not seem right. Ot is still a physical node that that these process run on with local drives and so on. On Wed, Jul 18, 2018, 7:54 AM scarmeli <carmeli....@gmail.com> wrote: > Answered in another mailing list > > Hi Shachar, > > with Flink 1.5 we added resource elasticity. This means that Flink is now > able to allocate new containers on a cluster management framework like > Yarn > or Mesos. Due to these changes (which also apply to the standalone mode), > Flink no longer reasons about a fixed set of TaskManagers because if > needed > it will start new containers (does not work in standalone mode). > Therefore, > it is hard for the system to make any decisions about spreading slots > belonging to a single job out across multiple TMs. It gets even harder > when > you consider that some jobs like yours might benefit from such a strategy > whereas others would benefit from co-locating its slots. It gets even more > complicated if you want to do scheduling wrt to multiple jobs which the > system does not have full knowledge about because they are submitted > sequentially. Therefore, Flink currently assumes that slots requests can > be > fulfilled by any TaskManager. > > Cheers, > Till > > > > -- > Sent from: http://apache-flink-user-mailing-list-archive.2336050. > n4.nabble.com/ >