The client was not using a config file, it is a stand-alone java app using the flink-client jar file. Thanks for the clarification.
Michael > On Apr 16, 2018, at 2:11 PM, Fabian Hueske <fhue...@gmail.com> wrote: > > The parallelism.default property that is configured in the flink-conf.yaml > file is only considered if the config file belongs to the submitting client. > If you configured the property in the config file of your cluster setup and > used submitted from a client that used a different configuration file, the > property of the other config file is used. > > I tested the behavior on Flink 1.4.2 and setting the parallelism in the > flink-conf.yaml of the client was working correctly in a simple local setup. > > If this doesn't solve your problem, we'd need a bit more information about > the job submission and setup. > > Best, Fabian > > > 2018-04-16 18:37 GMT+02:00 TechnoMage <mla...@technomage.com > <mailto:mla...@technomage.com>>: > 1.4.2. I have since set the parallelism explicitly after creating the env > and that is working. I also made the stream object serializable which may > also be involved in this. I will retest without the explicit parallelism > when I get a chance. > > Michael > > >> On Apr 16, 2018, at 2:05 AM, Fabian Hueske <fhue...@gmail.com >> <mailto:fhue...@gmail.com>> wrote: >> >> (re-adding user mailing list) >> >> A non-serializable function object should cause the job to fail, but not to >> ignore a parallelism setting. >> >> This might be a bug. Most users specify the parallelism directly in the >> application code (via StreamExecutionEnvironment) or when submitting the >> application. >> Which version are you using? >> >> Best, Fabian >> >> 2018-04-14 15:07 GMT+02:00 Michael Latta <mla...@technomage.com >> <mailto:mla...@technomage.com>>: >> Parallelism in config. I think the issue is that some objects used oin the >> stream are not serializable (which I just discovered). I am surprised it >> supports that. >> >> >> Michael >> >> On Apr 14, 2018, at 6:12 AM, Fabian Hueske <fhue...@gmail.com >> <mailto:fhue...@gmail.com>> wrote: >> >>> Hi, >>> >>> The number of Taskmanagers is irrelevant für the parallelism of a job or >>> operator. The scheduler only cares about the number of slots. >>> >>> How did you set the default parallelism? In the config or in the program / >>> StreamExecutionEnvironment? >>> >>> Best, Fabian >>> >>> >>> TechnoMage <mla...@technomage.com <mailto:mla...@technomage.com>> schrieb >>> am Fr., 13. Apr. 2018, 04:30: >>> I am pretty new to flink. I have a flink job that has 10 transforms >>> (mostly CoFlatMap with some simple filters and key extractrs as well. I >>> have the config set for 6 slots and default parallelism of 6, but all my >>> stages show paralellism of 1. Is that because there is only one task >>> manager? Some of what I have read suggested separate slots were needed to >>> use multiple threads on a single box? I have read the section on the docs >>> several times and still not totally sure about the execution model. >>> >>> Michael >> > >