Good to hear :) On Thu, Sep 18, 2025 at 6:35 PM Jean-Marc Paulin <[email protected]> wrote:
> Good suggestion, > > Turns out hard coding the config works... Thank you for the suggestion. > Rest to see exactly what typo I made in the yaml. > > Best Regard, > JM > > > > On Thu, Sep 18, 2025 at 4:41 PM Gabor Somogyi <[email protected]> > wrote: > >> Hi Jean-Marc, >> >> Could you please double check that your code is having the mentioned >> fix + give a simple repro steps? >> Please hardcode the batch size value in the code [1] to avoid any yaml to >> config issues. >> I would take a look if you can help a bit. >> >> BR, >> G >> >> [1] >> https://github.com/apache/flink/pull/25764/files#diff-278bac11f68be56ee499b24afe5e1d53a7c61b4d636654fd96b4167e2a45cbacR125 >> >> >> On Thu, Sep 18, 2025 at 5:30 PM Jean-Marc Paulin <[email protected]> >> wrote: >> >>> Hi, >>> >>> Using Flink 1.20.1, we get this error when trying to read a savepoint: >>> >>> Caused by: java.lang.RuntimeException: Record size is too large for >>> CollectSinkFunction. Record size is 5411925 bytes, but max bytes per batch >>> is only 2097152 bytes. Please consider increasing max bytes per batch value >>> by setting collect-sink.batch-size.max >>> at >>> org.apache.flink.streaming.api.operators.collect.CollectSinkFunction.invoke(CollectSinkFunction.java:288) >>> ... >>> >>> I tried to set collect-sink.batch-size.max in the flink-conf.yaml, but I >>> still hit the same error. It's like it's not taken into account. I see >>> there is a fix in 1.20.1 (https://github.com/apache/flink/pull/25764) >>> for this, but I still face the same issue, >>> >>> JM >>> >>> >>>
