Hi Chad, Here is the most generally useful process I ended up with, implemented in a local custom utility script.
#Update slurm.conf everywhere #Stop slurmctld #Restart all slurmd processes #Start slurmctld per: https://wiki.fysik.dtu.dk/niflheim/SLURM#add-and-remove-nodes I think you only will affect running jobs if you delete a partition, then you have a different procedure: https://slurm.schedmd.com/faq.html#delete_partition However, I think there are a few other parameters in slurm.conf that may be more disruptive but they have warnings in the man page. Regards, Alex On Wed, Jul 24, 2019 at 10:28 AM Julius, Chad <chad.jul...@sdstate.edu> wrote: > All, > > > > As our user base grows, we are getting close to hitting the default > MaxJobCount of 10,000. Is it safe to edit this value, along with some > fairshare settings while the cluster jobs are actively running? As in, > what are the ramifications if I changed the slurm.conf file and then > restarted slurmctld and the slurmd services on all of the nodes. > > > > My assumption is that the jobs will stay running and/or queued but I would > like some reassurance. > > > > Thanks, > > > > Chad > > >