ehavior with cgroup v2 in Slurm
What kernel are you using? I had a similar issue in an older RHEL 9 kernel
which has now been fixed
Sean
From: Guillaume COCHARD via slurm-users
Sent: Thursday, May 22, 2025 7:10:11 PM
To: slurm-us...@schedmd.com
Subject: [EXT] [slurm-users] Wrong MaxR
Hello,
Have you tried using AccountingStoreFlags=job_script as described in [
https://slurm.schedmd.com/slurm.conf.html#OPT_AccountingStoreFlags |
https://slurm.schedmd.com/slurm.conf.html#OPT_AccountingStoreFlags ] ? The
user's script should then be accessible through sacct -B (cf. [
https:/
2 Mai 2025 11:33:18
Objet: [slurm-users] Re: Wrong MaxRSS Behavior with cgroup v2 in Slurm
salut guillaume,
nothing else is different between the v1 and v2 setup? (/tmp is tmpfs on
v2 setup perhaps?)
stijn
On 5/22/25 11:10, Guillaume COCHARD via slurm-users wrote:
> Hello,
>
> We&
Hello,
We've noticed a recent change in how MaxRSS is reported on our cluster.
Specifically, the MaxRSS value for many jobs now often matches the allocated
memory, which was not the case previously. It appears this change is due to how
Slurm accounts for memory when copying large files, likely
Hi,
Speaking of RPC rate limiting, we recently encountered an issue with Snakemake
making excessive requests to sacct. It seems that the current rate limiting
only applies to controller RPCs. Is there a way to also limit the rate of sacct
calls?
Thanks,
Guillaume
- Mail original -
De:
Hello,
The Slurm documentation states that "The default allocation method within a
node is cyclic allocation (allocate available CPUs in a round-robin fashion
across the sockets within a node)."
In our case, we are using hyperthreading, which means that if two jobs
requesting one CPU each arr
usrPTphKaUagFzfF9CNo5WYMHOY2GiJMGbhNaCs2qKMrSxpgxSthUZDwCq16zeU03xps1Ar3le5oNqse4SbMDbfEd8anRGRFFhqXTVqHukHok4YQdNAXzQvUG4oI_J5hd11IrO6QMK-jdaUqH3BT1SCR9J4wzbDgUQPsYzZLmhvUjNs7R5Ok5uDUTWbyIUCFgbvDj-/http%3A%2F%2Fwww.dkrz.de%2F>
>
> Geschäftsführer: Prof. Dr. Thomas Ludwig
> Sitz der Gesell
as a QOS parameter.
You can create a QOS with the restrictions you'd like, and then in the
partition definition, you give it that QOS. The QOS will then apply its
restrictions to any jobs that use that partition.
Rob
From: Guillaume COCHARD via slurm-users
Sent: Tuesday, September 2
partition definition, you give it that QOS. The QOS will then apply its
restrictions to any jobs that use that partition.
Rob
From: Guillaume COCHARD via slurm-users
Sent: Tuesday, September 24, 2024 9:30 AM
To: slurm-users@lists.schedmd.com
Subject: [slurm-users] Max TRES per user and
Hello,
We are looking for a method to limit the TRES used by each user on a per-node
basis. For example, we would like to limit the total memory allocation of jobs
from a user to 200G per node.
There is MaxTRESperNode
(https://slurm.schedmd.com/sacctmgr.html#OPT_MaxTRESPerNode), but
unfortuna
Hello,
We have upgraded our cluster to Slurm 23.11.1 then, a few weeks later, to
23.11.4. Since then, Slurm doesn't detect hyperthreaded CPUs. We have
downgraded our test cluster, the issue is not present with Slurm 22.05 (we had
skipped Slurm 23.02).
For example, we are working with this node
11 matches
Mail list logo