You probably want to look at scontrol show node and scontrol show job
for that node and the jobs on it.
Compare those and you may find someone requested most all the resources,
but are not running them properly. Look at the job itself to see what it
is trying to do.
Brian Andrus
On 7/11/202
Does SACK replace MUNGE? As in - MUNGE is not required when building
Slurm or on compute?
If so, can the Requires and BuildRequires for munge be made optional on
bcond_without_munge in the spec file?
Or is there a reason MUNGE must remain a hard require for Slurm?
Thanks,
--Dani_L.
--
sl
Still learning about SLURM, so please forgive me if I ask a naïve question
I like to use Anders Halager’s gnodes command to visualise the state of our
nodes. I’ve noticed lately that we fairly often see things like this
(apologies for line wrap):
+- core - 46 cores & 186GB
--
On 2024/07/10 16:25, jack.mellor--- via slurm-users wrote:
We are running slurm 23.02.6.
Our nodes have hyperthreading disabled and we have slurm.conf
set to CPU=32 for each node (each node has 2 processes with 16 cores).
When we allocated a job, such as salloc -n 32, it will allocate
a whole no