Thats great news!

Are there any plans to expose it in the upcoming Flink release?

On Tue, 15 Jan 2019 at 12:59, Till Rohrmann <trohrm...@apache.org> wrote:

> Hi Alexandru,
>
> at the moment `/jobs/:jobid/rescaling` will always change the parallelism
> for all operators. The maximum is the maximum parallelism which you have
> defined for an operator.
>
> I agree that it should also be possible to rescale an individual operator.
> There internal functionality is already implemented (see
> JobMaster#rescaleOperators) but has not been exposed.
>
> Cheers,
> Till
>
> On Tue, Jan 15, 2019 at 1:03 PM Alexandru Gutan <alex.guta...@gmail.com>
> wrote:
>
>> Thanks Till!
>>
>> To execute the above (using Kubernetes), one would enter the running
>> JobManager service and execute it?
>> The following REST API call does the same */jobs/:jobid/rescaling*?
>>
>> I assume it changes the base parallelism, but what it will do if I had
>> already set the parallelism of my operators?
>> e.g.
>> .source(..)
>> .setParallelism(3)
>> .setUID(..)
>> .map(..)
>> .setParallelism(8)
>> .setUID(..)
>> .sink(..)
>> .setParallelism(3)
>> .setUID(..)
>>
>> I think it would be a good idea to have */jobs/:jobid/rescaling,* 
>> additionally
>> requiring the *operatorUID* as a queryParameter*, *so that the
>> parallelism of specific operators could be changed.
>>
>> Best,
>> Alex.
>>
>> On Tue, 15 Jan 2019 at 10:27, Till Rohrmann <trohrm...@apache.org> wrote:
>>
>>> Hi Alexandru,
>>>
>>> you can use the `modify` command `bin/flink modify <JOB_ID>
>>> --parallelism <PARALLELISM>` to modify the parallelism of a job. At the
>>> moment, it is implemented as first taking a savepoint, stopping the job and
>>> then redeploying the job with the changed parallelism and resuming from the
>>> savepoint.
>>>
>>> Cheers,
>>> Till
>>>
>>> On Mon, Jan 14, 2019 at 4:21 PM Dawid Wysakowicz <dwysakow...@apache.org>
>>> wrote:
>>>
>>>> Hi Alexandru
>>>>
>>>> As for 2, generally speaking the number of required slots depends on
>>>> number of slot sharing groups. By default all operators belong to the
>>>> default slot sharing group, that means a job requires as many slots as
>>>> maximal parallelism in the job. More on the distributed runtime you can
>>>> read here[1]
>>>>
>>>> As for 1 I cc'ed Gary and Till who might better answer your question.
>>>>
>>>> [1]
>>>> https://ci.apache.org/projects/flink/flink-docs-release-1.7/concepts/runtime.html#task-slots-and-resources
>>>>
>>>> Best,
>>>>
>>>> Dawid
>>>> On 14/01/2019 15:26, Alexandru Gutan wrote:
>>>>
>>>> Hi everyone!
>>>>
>>>> 1. Is there a way to increase the parallelism (e.g. through REST) of
>>>> some operators in a job without re-deploying the job? I found this
>>>> <https://stackoverflow.com/questions/50719147/apache-flink-guideliness-for-setting-parallelism>
>>>> answer which mentions scaling at runtime on Yarn/Mesos. Is it possible?
>>>> How? Support for Kubernetes?
>>>> 2. What happens when the number of parallel operator instances exceeds
>>>> the number of task slots? For example: a job with a source (parallelism 3),
>>>> a map (parallelism 8), a sink (parallelism 3), total of *14* operator
>>>> instances and a setup with *8* task slots. Will the operators get
>>>> chained? What if I disable operator chaining?
>>>>
>>>> Thank you!
>>>>
>>>>

Reply via email to