Re: [slurm-users] Single Node cluster. How to manage oversubscribing

2023-02-25 Thread Doug Meyer
Hi, You got me, I didn't know that " oversubscribe=FORCE:2" is an option. I'll need to explore that. I missed the question about srun. srun is the preferred I believe. I am not associated with drafting the submit scripts but can ask my peer. You do need to stipulate the number of cores you wa

Re: [slurm-users] Single Node cluster. How to manage oversubscribing

2023-02-25 Thread Analabha Roy
Hi, Thanks for your considered response. Couple of questions linger... On Sat, 25 Feb 2023 at 21:46, Doug Meyer wrote: > Hi, > > Declaring cores=64 will absolutely work but if you start running MPI > you'll want a more detailed config description. The easy way to read it is > "128=2 sockets *

Re: [slurm-users] speed / efficiency of sacct vs. scontrol

2023-02-25 Thread Chris Samuel
On 23/2/23 2:55 am, David Laehnemann wrote: And consequently, would using `scontrol` thus be the better default option (as opposed to `sacct`) for repeated job status checks by a workflow management system? Many others have commented on this, but use of scontrol in this way is really really b

Re: [slurm-users] Single Node cluster. How to manage oversubscribing

2023-02-25 Thread Doug Meyer
Hi, Declaring cores=64 will absolutely work but if you start running MPI you'll want a more detailed config description. The easy way to read it is "128=2 sockets * 32 corespersocket * 2 threads per core". NodeName=hpc[306-308] CPUs=128 Sockets=2 CoresPerSocket=32 ThreadsPerCore=2 RealMemory=512