sacctmgr list cluster
Cluster ControlHost ControlPort RPC Share GrpJobs GrpTRES
GrpSubmit MaxJobs MaxTRES MaxSubmit MaxWall QOS
Def QOS
-- --- - - --- -
- --- -
Hi Sean
I made all the changes you recommended but the problem remains.
Attached you will find dbd & ctld log files an slurmd log file from one node
wn001. Also slum configuration.
scontrol show node wn001
NodeName=wn001 Arch=x86_64 CoresPerSocket=2
CPUAlloc=0 CPUTot=2 CPULoad=
Hi everyone,
On 08.04.21 02:13, Christopher Samuel wrote:
I've not had issues with naming partitions in the past, though I can
imagine `default` could cause confusion as there is a `default=yes`
setting you can put on the one partition you want as the default choice.
more than that. The Parti
On 4/7/21 11:48 am, Administração de Sistemas do Centro de
Bioinformática wrote:
Unfortunately, I still don't know how to use any other value to
PartitionName.
We've got about 20 different partitions on our large Cray system, with a
variety of names (our submit filter system directs jobs to
Hi Everyone,
We have a challenge with scheduling jobs in a partition comprised of
heterogenous nodes with respect to memory and cores [1]. We further use cores
as the unit of measure for charging users. Currently we implement a crude
mechanism of using MaxMemPerCore as a proxy for memory use, t
Hi, Christine.
I had the same problem: when I configure PartitionName=default, Slurm
doesn't work. Then, I tried to change it to PartitionName=bioinfo and I had
the same error. So, I had to accept PartitionName=batch (using Ubuntu 20.04
LTS, Slurm 19.05.5).
Unfortunately, I still don't know how t
Still have this question. Sometime we have free nodes and users that are
allowed to run in the MAGNETIC reservation are first scheduled on the
free nodes instead of reservation nodes. Dit I forgot an option or is
this the expected behavior?
On 25/09/2020 16:47, Bas van der Vlies wrote:
Ar
The "solution" for my problem was very simple: after reboot of all hosts
in this test cluster (login node with slurmctld/slurmdbd + 2 worker
nodes) I do get reasonable values in sshare. Maybe I didn't do that
after finishing deepops installation procedure (but I didn't know I had
to do it and h