Not sure if this is correct but I think you need to leave a bit of RAM for the OS to use so best not to allow slurm to allocate ALL of it. I usually take 8G off to allow for that - negligible when our nodes have at least 768GB of RAM. At least this is my experience when using cgroups.
Emyr James Head of Scientific IT CRG - Centre for Genomic Regulation ________________________________ From: Diego Zuccato via slurm-users <slurm-users@lists.schedmd.com> Sent: 11 July 2024 08:06 To: slurm-users@lists.schedmd.com <slurm-users@lists.schedmd.com> Subject: [slurm-users] Re: Nodes TRES double what is requested Hint: round down a bit the RAM reported by 'slurmd -C'. Or you risk the nodes not coming back up after an upgrade that leaves a bit less free RAM than configured. Diego Il 10/07/2024 17:29, Brian Andrus via slurm-users ha scritto: > Jack, > > To make sure things are set right, run 'slurmd -C' on the node and use > that output in your config. > > It can also give you insight as to what is being seen on the node versus > what you may expect. > > Brian Andrus > > On 7/10/2024 1:25 AM, jack.mellor--- via slurm-users wrote: >> Hi, >> >> We are running slurm 23.02.6. Our nodes have hyperthreading disabled >> and we have slurm.conf set to CPU=32 for each node (each node has 2 >> processes with 16 cores). When we allocated a job, such as salloc -n >> 32, it will allocate a whole node but using sinfo shows double the >> allocation in the TRES=64. It also shows in sinfo that the node has >> 4294967264 idle CPUs. >> >> Not sure if its a known bug, or an issue with our config? I have tried >> various things, like setting the sockets/boards in slurm.conf. >> >> Thanks >> Jack >> > -- Diego Zuccato DIFA - Dip. di Fisica e Astronomia Servizi Informatici Alma Mater Studiorum - Università di Bologna V.le Berti-Pichat 6/2 - 40127 Bologna - Italy tel.: +39 051 20 95786 -- slurm-users mailing list -- slurm-users@lists.schedmd.com To unsubscribe send an email to slurm-users-le...@lists.schedmd.com
-- slurm-users mailing list -- slurm-users@lists.schedmd.com To unsubscribe send an email to slurm-users-le...@lists.schedmd.com