I'm not aware of any plans to change this. It would be possible though, to add a mode where a standalone cluster only accepts a single job.
On Fri, 1 Jul 2016 at 04:50 Jark Wu <wuchong...@alibaba-inc.com> wrote: > As standalone mode has the disadvantage that the TaskManager JVM can’t > isolate different jobs. Does we have any plan to improve this ? > > - Jark Wu > > > 在 2016年6月30日,下午4:19,Kevin Jacobs <kevin.jac...@cern.ch> 写道: > > > > In my opinion the streaming process can be perfectly simulated on a > single node. You can setup a message distribution system like Kafka on a > single node, you can run Spark on a single node and the only thing you need > to change when running it on a cluster is that you need to change the > environment. So there is no need to setup a cluster when testing the > streaming process. > > > > Regards, > > Kevin > > > > On 30-06-16 09:54, Longda Feng wrote: > >> > >> This means Standalone mode is just for prototype. > >> But I think we need a lightweight solution for streaming process, > standalone is the best solution. Some times, we need setup a flink cluster > on a small cluster. setup a yarn cluster isn't convenient. > >> (1) in small company, the number of machine is small (2) When a data > center is small, but we still need do some computing in this data center(3) > some machines are in the while-list, they have been authorited to access > some special data or machine, but the number of these machine is small.(4) > some machine has critical data, they can't be shared with others, but the > number of these machine is small.(5) when a team start to learn flink, he > will setup a small cluster firstly, maybe he wo't want to setup a huge > system, perfer to a small system > >> > >> > >> regardsLongda > >> > >> > ------------------------------------------------------------------From:Aljoscha > Krettek <aljos...@apache.org>Send Time:2016年6月29日(星期三) 21:48To:封仲淹(纪君祥) < > zhongyan.f...@alibaba-inc.com>; dev <dev@flink.apache.org>Subject:Re: > [Discuss] Why different job's tasks can run in the single process. > >> Hi, > >> yes, you are definitely right that allowing to run multiple user code > tasks > >> in the same TaskManager JVM is not good for stability. This mode is > still > >> there from the very early days of Flink where Yarn was not yet > available. > >> In a production environment I would now recommend to always run one > >> Flink-Yarn cluster per job to get good isolation between different jobs. > >> > >> Cheers, > >> Aljoscha > >> > >> On Wed, 29 Jun 2016 at 09:18 Longda Feng <zhongyan.f...@alibaba-inc.com > > > >> wrote: > >> > >>> hi , > >>> Sorry for asking the quest here? Any answer will be apprecated. > >>> Why different job's tasks can run in the single process. (There are > some > >>> different job's tasks in one TaskManager).It seems Flink-on-Yarn can > let > >>> different job run on different process. But for standalone mode, this > >>> problem still exists. > >>> Why design Flink like this?The advantage What I can thought is as > >>> following:(1) All task can share bigger memory pool.(2) The > communication > >>> between the tasks in the same process will be fast. > >>> But this design will impact to the stability. Flink provide > >>> User-Define-Function interface, if one of the User-Define-Function > crash, > >>> It maybe crack the whole JVM, If the TaskManager crash, all other > job's > >>> task in this TaskManager will be impacted. Even if the JVM don't > crash, but > >>> maybe lead to some other unexpected problem, what's more this will > make the > >>> code too sophisticated。Normal framework like Spark/Storm/Samza won't > run > >>> different job's tasks in the same process。As one normal user, > stability has > >>> the highest priority. > >>> > >>> ThanksLongda > >>> > >>> > >>> > >>> > >