We using Flink 1.10 running on Mesos. 

Med venlig hilsen / Best regards
Lasse Nedergaard


> Den 30. apr. 2020 kl. 04.53 skrev Yun Tang <myas...@live.com>:
> 
> 
> Hi Lasse
> 
> Which version of Flink did you use? Before Flink-1.10, there might exist 
> memory problem when RocksDB executes savepoint with write batch[1].
> 
> 
> [1] https://issues.apache.org/jira/browse/FLINK-12785
> 
> Best
> Yun Tang
> From: Lasse Nedergaard <lassenedergaardfl...@gmail.com>
> Sent: Wednesday, April 29, 2020 21:17
> To: user <user@flink.apache.org>
> Subject: Savepoint memory overhead
>  
> Hi.
> 
> I would like to know if there are any guidelines/recommendations for the 
> memory overhead we need to calculate for when doing savepoint to s3. We use 
> RockDb state backend.
> 
> We run our job on relative small task managers and we can see we get memory 
> problems if the state size for each task manager get "big" (we haven't found 
> the rule of thumbs yet) and we can remove the problem if we reduce the state 
> size, or increase parallelism and jobs with none or small state don't have 
> any problems.
> So I see a relation between between allocated memory to a task manager and 
> the state it can handle. 
> 
> So do anyone have any recommendations/ base practices for this and can 
> someone explain why savepoint requires memory.
> 
> Thanks
> 
> In advance
> 
> Lasse Nedergaard 
> 

Reply via email to