We just use the Archive function built into slurm.  That has worked fine for us for the past 6 years.  We keep 6 months of data in the active archive.

If you have 6 years worth of data and you want to prune down to 2 years, I recommend going month by month rather than doing it in one go.  When we initially started archiving data several years back our first pass at archiving (which at that time had 2 years of data in it) took forever and actually caused issues with the archive process.  We worked with SchedMD, improved the archive script built into Slurm but also decided to only archive one month at a time which allowed it to get done in a reasonable amount of time.


The archived data can be pulled into a different slurm database, which is what we do for importing historic data into our XDMod instance.


-Paul Edmon-


On 7/13/2022 4:55 PM, Timony, Mick wrote:
Hi Slurm Users,

Currently we don't archive our SlurmDB and have 6 years' worth of data in our SlurmDB. We are looking to start archiving our database as it starting to get rather large, and we have decided to keep 2 years' worth of data. I'm wondering what approaches or scripts other groups use.

The docs refer to the ArchiveScript setting at:
https://slurm.schedmd.com/slurmdbd.conf.html#OPT_ArchiveScript

I've seen suggestions to import into another database that will require keeping the schema up-to-date which seems like a possible maintenance issue or nightmare if one forgets to update the schema after updating Slurmdb. We also have most of the information in an Elasticsearch <https://slurm.schedmd.com/elasticsearch.html> instance, which will likely suite our needs for long term historical information.


What do you use to archive this information? CSV files, SQL dumps or something else?


Regards
--
Mick Timony
Senior DevOps Engineer
Harvard Medical School
--

Reply via email to