In short, no.

G


On Mon, Sep 22, 2025 at 12:25 PM Jean-Marc Paulin <[email protected]>
wrote:

> Hi Gabor,
>
> I hit that error (Record size is too large for CollectSinkFunction.) when
> introspecting a savepoint with the state API. Is that an error a normal
> flink application can hit when resuming from a savepoint or a checkpoint?
> (trying to understand how concerned I should be).
>
> Thanks
>
> JM
>
>
> On Thu, Sep 18, 2025 at 6:20 PM Gabor Somogyi <[email protected]>
> wrote:
>
>> Good to hear :)
>>
>> On Thu, Sep 18, 2025 at 6:35 PM Jean-Marc Paulin <[email protected]>
>> wrote:
>>
>>> Good suggestion,
>>>
>>> Turns out hard coding the config works... Thank you for the suggestion.
>>> Rest to see exactly what typo I made in the yaml.
>>>
>>> Best Regard,
>>> JM
>>>
>>>
>>>
>>> On Thu, Sep 18, 2025 at 4:41 PM Gabor Somogyi <[email protected]>
>>> wrote:
>>>
>>>> Hi Jean-Marc,
>>>>
>>>> Could you please double check that your code is having the mentioned
>>>> fix + give a simple repro steps?
>>>> Please hardcode the batch size value in the code [1] to avoid any yaml
>>>> to config issues.
>>>> I would take a look if you can help a bit.
>>>>
>>>> BR,
>>>> G
>>>>
>>>> [1]
>>>> https://github.com/apache/flink/pull/25764/files#diff-278bac11f68be56ee499b24afe5e1d53a7c61b4d636654fd96b4167e2a45cbacR125
>>>>
>>>>
>>>> On Thu, Sep 18, 2025 at 5:30 PM Jean-Marc Paulin <[email protected]>
>>>> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> Using Flink 1.20.1, we get this error when trying to read a savepoint:
>>>>>
>>>>> Caused by: java.lang.RuntimeException: Record size is too large for
>>>>> CollectSinkFunction. Record size is 5411925 bytes, but max bytes per batch
>>>>> is only 2097152 bytes. Please consider increasing max bytes per batch 
>>>>> value
>>>>> by setting collect-sink.batch-size.max
>>>>>         at
>>>>> org.apache.flink.streaming.api.operators.collect.CollectSinkFunction.invoke(CollectSinkFunction.java:288)
>>>>> ...
>>>>>
>>>>> I tried to set collect-sink.batch-size.max in the flink-conf.yaml, but
>>>>> I still hit the same error. It's like it's not taken into account. I see
>>>>> there is a fix in 1.20.1 (https://github.com/apache/flink/pull/25764)
>>>>> for this, but I still face the same issue,
>>>>>
>>>>> JM
>>>>>
>>>>>
>>>>>

Reply via email to