Re: [slurm-users] [EXT] slurmctld error

2021-04-07 Thread Ioannis Botsis
sacctmgr list cluster Cluster ControlHost ControlPort RPC Share GrpJobs GrpTRES GrpSubmit MaxJobs MaxTRES MaxSubmit MaxWall QOS Def QOS -- --- - - --- - - --- -

Re: [slurm-users] [EXT] slurmctld error

2021-04-07 Thread Ioannis Botsis
Hi Sean I made all the changes you recommended but the problem remains. Attached you will find dbd & ctld log files an slurmd log file from one node wn001. Also slum configuration. scontrol show node wn001 NodeName=wn001 Arch=x86_64 CoresPerSocket=2 CPUAlloc=0 CPUTot=2 CPULoad=

Re: [slurm-users] PartitionName default

2021-04-07 Thread Marcus Boden
Hi everyone, On 08.04.21 02:13, Christopher Samuel wrote: I've not had issues with naming partitions in the past, though I can imagine `default` could cause confusion as there is a `default=yes` setting you can put on the one partition you want as the default choice. more than that. The Parti

Re: [slurm-users] PartitionName default

2021-04-07 Thread Christopher Samuel
On 4/7/21 11:48 am, Administração de Sistemas do Centro de Bioinformática wrote: Unfortunately, I still don't know how to use any other value to PartitionName. We've got about 20 different partitions on our large Cray system, with a variety of names (our submit filter system directs jobs to

[slurm-users] Heterogenous Memory, Partition limits, node preference and backfill.

2021-04-07 Thread Ruffner, Scott (jpr9c)
Hi Everyone, We have a challenge with scheduling jobs in a partition comprised of heterogenous nodes with respect to memory and cores [1]. We further use cores as the unit of measure for charging users. Currently we implement a crude mechanism of using MaxMemPerCore as a proxy for memory use, t

Re: [slurm-users] PartitionName default

2021-04-07 Thread Administração de Sistemas do Centro de Bioinformática
Hi, Christine. I had the same problem: when I configure PartitionName=default, Slurm doesn't work. Then, I tried to change it to PartitionName=bioinfo and I had the same error. So, I had to accept PartitionName=batch (using Ubuntu 20.04 LTS, Slurm 19.05.5). Unfortunately, I still don't know how t

Re: [slurm-users] SLURM reservations with MAGNETIC flag

2021-04-07 Thread Bas van der Vlies
Still have this question. Sometime we have free nodes and users that are allowed to run in the MAGNETIC reservation are first scheduled on the free nodes instead of reservation nodes. Dit I forgot an option or is this the expected behavior? On 25/09/2020 16:47, Bas van der Vlies wrote: Ar

Re: [slurm-users] RawUsage 0??

2021-04-07 Thread Matthias Leopold
The "solution" for my problem was very simple: after reboot of all hosts in this test cluster (login node with slurmctld/slurmdbd + 2 worker nodes) I do get reasonable values in sshare. Maybe I didn't do that after finishing deepops installation procedure (but I didn't know I had to do it and h