We have observed the same issue in our production cluster. Filed FLINK-7743
for the fix.

~Haohui

On Fri, Sep 29, 2017 at 1:18 AM Till Rohrmann <trohrm...@apache.org> wrote:

> Hi Dan,
>
> I think Aljoscha is right and the 768 MB minimum JM memory is more of a
> legacy artifact which was never properly refactored. If I remember
> correctly, then we had problems when starting Flink in a container with a
> lower memory limit. Therefore this limit was introduced. But I'm actually
> not sure whether this is still valid and should definitely be verified
> again.
>
> Cheers,
> Till
>
> On Thu, Sep 28, 2017 at 10:52 PM, Aljoscha Krettek <aljos...@apache.org>
> wrote:
>
>> I believe this could be from a time when there was not yet the setting 
>> "containerized.heap-cutoff-min"
>> since this part of the code is quite old.
>>
>> I think we could be able to remove that restriction but I'm not sure so
>> I'm cc'ing Till who knows those parts best.
>>
>> @Till, what do you think?
>>
>> On 28. Sep 2017, at 17:47, Dan Circelli <dan.circe...@arcticwolf.com>
>> wrote:
>>
>> In our usage of Flink, our Yarn Job Manager never goes above ~48 MB of
>> heap utilization. In order to maximize the heap available to the Task
>> Managers I thought we could shrink our Job Manager heap setting down from
>> the 1024MB we were using to something tiny like 128MB. However, doing so
>> results in the runtime error:
>>
>> java.lang.IllegalArgumentException: The JobManager memory (64) is below
>> the minimum required memory amount of 768 MB
>> at
>> org.apache.flink.yarn.AbstractYarnClusterDescriptor.setJobManagerMemory(AbstractYarnClusterDescriptor.java:187)
>> …
>>
>> Looking into it: this value isn’t controlled by the settings in
>> yarn-site.xml but is actually hardcoded in Flink code base to 768 MB. (see
>> AbstractYarnDescriptor.java where MIN_JM_MEMORY = 768.)
>>
>>
>> Why is this hardcoded?
>> Why not let value be set via the Yarn Site Configuration xml?
>> Why such a high minimum?
>>
>>
>> Thanks,
>> Dan
>>
>>
>>
>

Reply via email to