Hi Zhanghao, Thanks for your suggestion. Unfortunately, this does not work, I still get the same error message:
``` Record size is too large for CollectSinkFunction. Record size is 9623137 bytes, but max bytes per batch is only 2097152 bytes. Please consider increasing max bytes per batch value by setting collect-sink.batch-size.max ``` The code looks like this now: ``` Configuration config = new Configuration(); config.setString("collect-sink.batch-size.max", "10mb"); StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment(config); SavepointReader savepoint = SavepointReader.read(env, savepointPath, new HashMapStateBackend()); var matcher = savepoint.readKeyedState("Raw Triggers", new MatcherReader()); var matcherState = matcher.executeAndCollect(1000); ``` I have tried other ways but none has worked (the setting is always ignored in the end). Regards, Salva On Sun, Jul 21, 2024 at 9:10 AM Zhanghao Chen <zhanghao.c...@outlook.com> wrote: > Hi, you could increase it as follows: > > Configuration config = new Configuration(); > config.setString(collect-sink.batch-size.max, "10mb"); > StreamExecutionEnvironment env = > StreamExecutionEnvironment.getExecutionEnvironment(config); > ------------------------------ > *From:* Salva Alcántara <salcantara...@gmail.com> > *Sent:* Saturday, July 20, 2024 15:05 > *To:* user <user@flink.apache.org> > *Subject:* SavepointReader: Record size is too large for > CollectSinkFunction > > Hi all! > > I'm trying to debug a job via inspecting its savepoints but I'm getting > this error message: > > ``` > Caused by: java.lang.RuntimeException: Record size is too large for > CollectSinkFunction. Record size is 9627127 bytes, but max bytes per batch > is only 2097152 bytes. Please consider increasing max bytes per batch value > by setting collect-sink.batch-size.max > ``` > > My code looks like this: > > ``` > private static void run(String savepointPath) throws Exception { > StreamExecutionEnvironment env = > StreamExecutionEnvironment.getExecutionEnvironment(); > org.apache.flink.state.api.SavepointReader savepoint = > org.apache.flink.state.api.SavepointReader.read(env, > savepointPath, new HashMapStateBackend()); > > var operator = savepoint.readKeyedState("uuid", new > MyKeyedOperatorReader()); > var operatorState = matcher.executeAndCollect(1000); > } > ``` > > I haven't found the way to increase the `collect-sink.batch-size.max` as > suggested in the error msg. > > Any help on this will be appreciated! > > Regards, > > Salva >