Il 30/09/20 03:49, Brian Andrus ha scritto:
Tks for the answer.
> That means the system has no idea who that user is.
But which system? Being a message generated by slurmctld, I thought it
must be the frontend node. But, as I wrote, that system correctly
identifies the user (he's logged in, 'id'
Il 29/09/20 16:19, Michael Di Domenico ha scritto:
> what leads you to believe that you're getting 2 CPU's instead of 1?
I think I saw that too, once, but thought it was related to hyperthreading.
--
Diego Zuccato
DIFA - Dip. di Fisica e Astronomia
Servizi Informatici
Alma Mater Studiorum - Univ
That means the system has no idea who that user is.
If you are using /etc/passwd, that file is not synched on the slurm
master node(s)
If you are part of a domain or other shared directory (ldap, etc), your
master is likely not configured right.
If you are using SSSD, it is also possible yo
Hi,
I recently upgraded Slurm from 19.05.5 to 20.02.5 and I noticed that I am
no longer able to use the scontrol update command on GRES unless the new
count is 0.
For example, if I have a GRES named "test" defined and in slurm.conf
> NodeName=node1 Gres=test:2
Initially, I have:
$ sinfo -n no
Here are the particulars asked for.
The following is the pertinent information for our cluster and the job run.
Note: server names, IP addresses and user IDs are anonymized.
Slurm.conf
==
TaskPlugin=task/affinity
# SCHEDULING
SchedulerType=s
There are three pieces of information that may provide Useful:
1 - these are VMs and not physical servers
2 - the OS is RedHat 7.8
2 - As far as I can tell, hyperthreading is not enabled, but will check for sure
3 - when we ask for 15Gb memory - we will only get 1 CPU
-Original Message-
F
The following is the pertinent information for our cluster and the job run.
Note: server names, IP addresses and user IDs are anonymized.
Slurm.conf
==
TaskPlugin=task/affinity
# SCHEDULING
SchedulerType=sched/backfill
SelectType=select/cons_
I am working on my first ever SLURM cluster build for use as a resource manager
in a JupyterHub Development environment. I have configured the cluster for
SelectType of 'select/con_res' with DefMemPerCPU and MaxMemPerCPU of 16Gb. The
idea is to essentially provide for jobs that run in a 1 CPU/
I am not familiar with using Slurm with VMs, but do note that Slurm can
behave a bit "unexpectedly" with memory constraints due to the memory
consumed by OS, etc.
E.g., if I had a 16 core machine with 64 GB of RAM and requested 16 cores
with 4 GB/core, it would not fit on this machine because some
what leads you to believe that you're getting 2 CPU's instead of 1?
'scontrol show job ' would be a helpful first start.
On Tue, Sep 29, 2020 at 9:56 AM Luecht, Jeff A wrote:
>
> I am working on my first ever SLURM cluster build for use as a resource
> manager in a JupyterHub Development environ
Hi Loris,
Thank you for clarification
Will figure out if we can apply this approach
Thanks
Alexey
-Original Message-
From: slurm-users On Behalf Of Loris
Bennett
Sent: Tuesday, September 29, 2020 9:37 AM
To: Slurm User Community List
Subject: RE: [EXTERNAL] [slurm-users] Working with
Hello all.
One of the users is unable to submit jobs to our cluster.
The first time he tries, he gets
$ sbatch test.job
sbatch: fatal: Invalid user id: 621049927
then:
$ sbatch test.job
sbatch: error: Batch job submission failed: Invalid account or
account/partition combination specified
In slur
12 matches
Mail list logo