[slurm-users] Re: Wrong MaxRSS Behavior with cgroup v2 in Slurm

2025-06-19 Thread Guillaume COCHARD via slurm-users
ehavior with cgroup v2 in Slurm What kernel are you using? I had a similar issue in an older RHEL 9 kernel which has now been fixed Sean From: Guillaume COCHARD via slurm-users Sent: Thursday, May 22, 2025 7:10:11 PM To: slurm-us...@schedmd.com Subject: [EXT] [slurm-users] Wrong MaxR

[slurm-users] Re: Job information if job is completed

2025-06-18 Thread Guillaume COCHARD via slurm-users
Hello, Have you tried using AccountingStoreFlags=job_script as described in [ https://slurm.schedmd.com/slurm.conf.html#OPT_AccountingStoreFlags | https://slurm.schedmd.com/slurm.conf.html#OPT_AccountingStoreFlags ] ? The user's script should then be accessible through sacct -B (cf. [ https:/

[slurm-users] Re: Wrong MaxRSS Behavior with cgroup v2 in Slurm

2025-05-22 Thread Guillaume COCHARD via slurm-users
2 Mai 2025 11:33:18 Objet: [slurm-users] Re: Wrong MaxRSS Behavior with cgroup v2 in Slurm salut guillaume, nothing else is different between the v1 and v2 setup? (/tmp is tmpfs on v2 setup perhaps?) stijn On 5/22/25 11:10, Guillaume COCHARD via slurm-users wrote: > Hello, > > We&

[slurm-users] Wrong MaxRSS Behavior with cgroup v2 in Slurm

2025-05-22 Thread Guillaume COCHARD via slurm-users
Hello, We've noticed a recent change in how MaxRSS is reported on our cluster. Specifically, the MaxRSS value for many jobs now often matches the allocated memory, which was not the case previously. It appears this change is due to how Slurm accounts for memory when copying large files, likely

[slurm-users] Re: Lots of RPC calls and REQUEST_GETPW calls

2025-05-07 Thread Guillaume COCHARD via slurm-users
Hi, Speaking of RPC rate limiting, we recently encountered an issue with Snakemake making excessive requests to sacct. It seems that the current rate limiting only applies to controller RPCs. Is there a way to also limit the rate of sacct calls? Thanks, Guillaume - Mail original - De:

[slurm-users] Optimizing CPU allocation in Slurm with hyperthreading enabled

2025-01-30 Thread Guillaume COCHARD via slurm-users
Hello, The Slurm documentation states that "The default allocation method within a node is cyclic allocation (allocate available CPUs in a round-robin fashion across the sockets within a node)." In our case, we are using hyperthreading, which means that if two jobs requesting one CPU each arr

[slurm-users] Re: Max TRES per user and node

2024-09-25 Thread Guillaume COCHARD via slurm-users
usrPTphKaUagFzfF9CNo5WYMHOY2GiJMGbhNaCs2qKMrSxpgxSthUZDwCq16zeU03xps1Ar3le5oNqse4SbMDbfEd8anRGRFFhqXTVqHukHok4YQdNAXzQvUG4oI_J5hd11IrO6QMK-jdaUqH3BT1SCR9J4wzbDgUQPsYzZLmhvUjNs7R5Ok5uDUTWbyIUCFgbvDj-/http%3A%2F%2Fwww.dkrz.de%2F> > > Geschäftsführer: Prof. Dr. Thomas Ludwig > Sitz der Gesell

[slurm-users] Re: Max TRES per user and node

2024-09-24 Thread Guillaume COCHARD via slurm-users
as a QOS parameter. You can create a QOS with the restrictions you'd like, and then in the partition definition, you give it that QOS. The QOS will then apply its restrictions to any jobs that use that partition. Rob From: Guillaume COCHARD via slurm-users Sent: Tuesday, September 2

[slurm-users] Re: Max TRES per user and node

2024-09-24 Thread Guillaume COCHARD via slurm-users
partition definition, you give it that QOS. The QOS will then apply its restrictions to any jobs that use that partition. Rob From: Guillaume COCHARD via slurm-users Sent: Tuesday, September 24, 2024 9:30 AM To: slurm-users@lists.schedmd.com Subject: [slurm-users] Max TRES per user and

[slurm-users] Max TRES per user and node

2024-09-24 Thread Guillaume COCHARD via slurm-users
Hello, We are looking for a method to limit the TRES used by each user on a per-node basis. For example, we would like to limit the total memory allocation of jobs from a user to 200G per node. There is MaxTRESperNode (https://slurm.schedmd.com/sacctmgr.html#OPT_MaxTRESPerNode), but unfortuna

[slurm-users] Incorrect hyperthreading with Slurm 23.11

2024-03-27 Thread Guillaume COCHARD via slurm-users
Hello, We have upgraded our cluster to Slurm 23.11.1 then, a few weeks later, to 23.11.4. Since then, Slurm doesn't detect hyperthreaded CPUs. We have downgraded our test cluster, the issue is not present with Slurm 22.05 (we had skipped Slurm 23.02). For example, we are working with this node