Hi Martjin, Thanks for the info. We are in the process of moving to 1.15. Is this version actively supported by community?
And coming to my original and follow up questions, I checked the RocksDbStatebackend code from 1.11 and 1.15, it is same. Given K8s configuration with Volume and mounth path, I would like to know the design recommendation for Rocks Db local storage path. Thanks, Vidya On Mon, Nov 14, 2022 at 6:57 AM Martijn Visser <martijnvis...@apache.org> wrote: > Hi Vidya, > > Given that you are still on Flink 1.11 which was released in July 2020 and > no longer supported by the community, I would recommend first upgrading to > a later, supported version like Flink 1.16. > > Best regards, > > Martijn > > On Sat, Nov 12, 2022 at 8:07 PM Vidya Sagar Mula <mulasa...@gmail.com> > wrote: > >> Hi Yanfei, >> >> Thank you for the response. I have follow up answer and questions. >> >> I have two set ups. One is on the local environment and the other one is >> a deployment scenario that is on K8s. >> >> - In K8s set up, I have Volume on the cluster node and mount path is >> specified for the RockDB checkpoints location. So, when the Application TM >> POD is restarted, the older checkpoints are read back from the host path >> again when the TM is UP again. >> In this case, RocksDB local directory is pulled with all the older data >> which is not useful for the JOB ID as the "instanceBasePath" is calculated >> with new random UUID. >> >> Questions: >> - What do you think about the older files that are pulled from the >> hostpath to mount path should be deleted first and then create the new >> instanceBasepath? >> Otherwise, we are going to be ended with the GBs of unwanted data. >> >> What is the general design recommendation is such cases where RocksDB has >> mount path to a Volume on host node? >> Please clarify. >> >> Thanks, >> Vidya Sagar. >> >> >> On Thu, Nov 10, 2022 at 7:52 PM Yanfei Lei <fredia...@gmail.com> wrote: >> >>> Hi Vidya Sagar, >>> >>> Could you please share the reason for TaskManager restart? If the >>> machine or JVM process of TaskManager crashes, the >>> `RocksDBKeyedStateBackend` can't be disposed/closed normally, so the >>> existing rocksdb instance directory would remain. >>> >>> BTW, if you use Application Mode on k8s, if a TaskManager(pod) crashes, >>> the rocksdb directory would be deleted as the pod is released. >>> >>> Best, >>> Yanfei >>> >>> Vidya Sagar Mula <mulasa...@gmail.com> 于2022年11月11日周五 01:39写道: >>> >>>> Hi, >>>> >>>> I am using RocksDB state backend for incremental checkpointing with >>>> Flink 1.11 version. >>>> >>>> Question: >>>> ---------- >>>> For a given Job ID, Intermediate RocksDB checkpoints are stored under >>>> the path defined with "" >>>> >>>> The files are stored with "_jobID+ radom UUID" prefixed to the location. >>>> >>>> Case : 1 >>>> --------- >>>> - When I cancel the job, then all the rocksDB checkpoints are deleted >>>> properly from the location corresponding to that JobId. >>>> (based on "instanceBasePath" variable stored in >>>> RocksDBKeyedStateBackend object). >>>> "NO Issue here. Working as expected". >>>> >>>> Case : 2 >>>> --------- >>>> - When my TaskManger is restarted, the existing rocksDb checkpoints are >>>> not deleted. >>>> New "instanceBasePath" is constructed with the new Random UUID >>>> appended to the directory. >>>> And, old checkpoint directories are still there. >>>> >>>> questions: >>>> - Is this expected behaviour not to delete the existing checkPoint >>>> dirs under the rocksDB local directory? >>>> - I see the "StreamTaskStateInitializerImpl.java", where new >>>> StateBackend objects are created. In this case, new directory is created >>>> for this Job ID appended with new random UUID. >>>> What happens to the old Directories. Are they going to be purged later >>>> on? >>>> If not, the disk is going to filled up with the older checkpoints. >>>> Please clarify this. >>>> >>>> Thanks, >>>> Vidya Sagar. >>>> >>> >>> >>>