Thanks Lyn – that was exactly the problem.
Jake
From: slurm-users On Behalf Of Lyn
Gerner
Sent: 03 June 2022 01:51
To: Slurm User Community List
Subject: Re: [slurm-users] New slurm configuration - multiple jobs per host
Jake, my hunch is that your jobs are getting hung up on mem allocation
Jake, my hunch is that your jobs are getting hung up on mem allocation,
such that Slurm is assigning all of memory to each job as it runs; you can
verify w/scontrol show job. If that's what's happening, try setting a
DefMemPerCPU value for your partition(s).
Best of luck,
Lyn
On Thu, May 26, 2022
Hi Ole
I only added the oversubscribe option because without it, it didn’t work - so
in fact, it appears not to have made any difference
I though the RealMemory option just said not to offer any jobs to the node that
didn’t have AT LEAST that amount of RAM
My large node has more than 64GB RAM (
Hi Jake,
Firstly, which Slurm version and which OS do you use?
Next, try simplifying by removing the oversubscribe configuration. Read
the slurm.conf manual page about oversubscribe, it looks a bit tricky.
The RealMemory=1000 is extremely low and might prevent jobs from
starting! Run "slur
Hi
I am just building my first Slurm setup and have got everything running - well,
almost.
I have a two node configuration. All of my setup exists on a single HyperV
server and I have divided up the resources to create my VMs
One node I will use for heavy duty work; this is called compute001
O