I'm not aware of someone working on this feature right now.

On Tue, Jan 15, 2019 at 3:22 PM Alexandru Gutan <alex.guta...@gmail.com>
wrote:

> Thats great news!
>
> Are there any plans to expose it in the upcoming Flink release?
>
> On Tue, 15 Jan 2019 at 12:59, Till Rohrmann <trohrm...@apache.org> wrote:
>
>> Hi Alexandru,
>>
>> at the moment `/jobs/:jobid/rescaling` will always change the parallelism
>> for all operators. The maximum is the maximum parallelism which you have
>> defined for an operator.
>>
>> I agree that it should also be possible to rescale an individual
>> operator. There internal functionality is already implemented (see
>> JobMaster#rescaleOperators) but has not been exposed.
>>
>> Cheers,
>> Till
>>
>> On Tue, Jan 15, 2019 at 1:03 PM Alexandru Gutan <alex.guta...@gmail.com>
>> wrote:
>>
>>> Thanks Till!
>>>
>>> To execute the above (using Kubernetes), one would enter the running
>>> JobManager service and execute it?
>>> The following REST API call does the same */jobs/:jobid/rescaling*?
>>>
>>> I assume it changes the base parallelism, but what it will do if I had
>>> already set the parallelism of my operators?
>>> e.g.
>>> .source(..)
>>> .setParallelism(3)
>>> .setUID(..)
>>> .map(..)
>>> .setParallelism(8)
>>> .setUID(..)
>>> .sink(..)
>>> .setParallelism(3)
>>> .setUID(..)
>>>
>>> I think it would be a good idea to have */jobs/:jobid/rescaling,* 
>>> additionally
>>> requiring the *operatorUID* as a queryParameter*, *so that the
>>> parallelism of specific operators could be changed.
>>>
>>> Best,
>>> Alex.
>>>
>>> On Tue, 15 Jan 2019 at 10:27, Till Rohrmann <trohrm...@apache.org>
>>> wrote:
>>>
>>>> Hi Alexandru,
>>>>
>>>> you can use the `modify` command `bin/flink modify <JOB_ID>
>>>> --parallelism <PARALLELISM>` to modify the parallelism of a job. At the
>>>> moment, it is implemented as first taking a savepoint, stopping the job and
>>>> then redeploying the job with the changed parallelism and resuming from the
>>>> savepoint.
>>>>
>>>> Cheers,
>>>> Till
>>>>
>>>> On Mon, Jan 14, 2019 at 4:21 PM Dawid Wysakowicz <
>>>> dwysakow...@apache.org> wrote:
>>>>
>>>>> Hi Alexandru
>>>>>
>>>>> As for 2, generally speaking the number of required slots depends on
>>>>> number of slot sharing groups. By default all operators belong to the
>>>>> default slot sharing group, that means a job requires as many slots as
>>>>> maximal parallelism in the job. More on the distributed runtime you can
>>>>> read here[1]
>>>>>
>>>>> As for 1 I cc'ed Gary and Till who might better answer your question.
>>>>>
>>>>> [1]
>>>>> https://ci.apache.org/projects/flink/flink-docs-release-1.7/concepts/runtime.html#task-slots-and-resources
>>>>>
>>>>> Best,
>>>>>
>>>>> Dawid
>>>>> On 14/01/2019 15:26, Alexandru Gutan wrote:
>>>>>
>>>>> Hi everyone!
>>>>>
>>>>> 1. Is there a way to increase the parallelism (e.g. through REST) of
>>>>> some operators in a job without re-deploying the job? I found this
>>>>> <https://stackoverflow.com/questions/50719147/apache-flink-guideliness-for-setting-parallelism>
>>>>> answer which mentions scaling at runtime on Yarn/Mesos. Is it possible?
>>>>> How? Support for Kubernetes?
>>>>> 2. What happens when the number of parallel operator instances exceeds
>>>>> the number of task slots? For example: a job with a source (parallelism 
>>>>> 3),
>>>>> a map (parallelism 8), a sink (parallelism 3), total of *14* operator
>>>>> instances and a setup with *8* task slots. Will the operators get
>>>>> chained? What if I disable operator chaining?
>>>>>
>>>>> Thank you!
>>>>>
>>>>>

Reply via email to