Hi Gabor,

I hit that error (Record size is too large for CollectSinkFunction.) when
introspecting a savepoint with the state API. Is that an error a normal
flink application can hit when resuming from a savepoint or a checkpoint?
(trying to understand how concerned I should be).

Thanks

JM


On Thu, Sep 18, 2025 at 6:20 PM Gabor Somogyi <[email protected]>
wrote:

> Good to hear :)
>
> On Thu, Sep 18, 2025 at 6:35 PM Jean-Marc Paulin <[email protected]>
> wrote:
>
>> Good suggestion,
>>
>> Turns out hard coding the config works... Thank you for the suggestion.
>> Rest to see exactly what typo I made in the yaml.
>>
>> Best Regard,
>> JM
>>
>>
>>
>> On Thu, Sep 18, 2025 at 4:41 PM Gabor Somogyi <[email protected]>
>> wrote:
>>
>>> Hi Jean-Marc,
>>>
>>> Could you please double check that your code is having the mentioned
>>> fix + give a simple repro steps?
>>> Please hardcode the batch size value in the code [1] to avoid any yaml
>>> to config issues.
>>> I would take a look if you can help a bit.
>>>
>>> BR,
>>> G
>>>
>>> [1]
>>> https://github.com/apache/flink/pull/25764/files#diff-278bac11f68be56ee499b24afe5e1d53a7c61b4d636654fd96b4167e2a45cbacR125
>>>
>>>
>>> On Thu, Sep 18, 2025 at 5:30 PM Jean-Marc Paulin <[email protected]>
>>> wrote:
>>>
>>>> Hi,
>>>>
>>>> Using Flink 1.20.1, we get this error when trying to read a savepoint:
>>>>
>>>> Caused by: java.lang.RuntimeException: Record size is too large for
>>>> CollectSinkFunction. Record size is 5411925 bytes, but max bytes per batch
>>>> is only 2097152 bytes. Please consider increasing max bytes per batch value
>>>> by setting collect-sink.batch-size.max
>>>>         at
>>>> org.apache.flink.streaming.api.operators.collect.CollectSinkFunction.invoke(CollectSinkFunction.java:288)
>>>> ...
>>>>
>>>> I tried to set collect-sink.batch-size.max in the flink-conf.yaml, but
>>>> I still hit the same error. It's like it's not taken into account. I see
>>>> there is a fix in 1.20.1 (https://github.com/apache/flink/pull/25764)
>>>> for this, but I still face the same issue,
>>>>
>>>> JM
>>>>
>>>>
>>>>

Reply via email to