Re: [slurm-users] error: user not found

2020-09-29 Thread Diego Zuccato
Il 30/09/20 03:49, Brian Andrus ha scritto: Tks for the answer. > That means the system has no idea who that user is. But which system? Being a message generated by slurmctld, I thought it must be the frontend node. But, as I wrote, that system correctly identifies the user (he's logged in, 'id'

Re: [slurm-users] Memory per CPU

2020-09-29 Thread Diego Zuccato
Il 29/09/20 16:19, Michael Di Domenico ha scritto: > what leads you to believe that you're getting 2 CPU's instead of 1? I think I saw that too, once, but thought it was related to hyperthreading. -- Diego Zuccato DIFA - Dip. di Fisica e Astronomia Servizi Informatici Alma Mater Studiorum - Univ

Re: [slurm-users] error: user not found

2020-09-29 Thread Brian Andrus
That means the system has no idea who that user is. If you are using /etc/passwd, that file is not synched on the slurm master node(s) If you are part of a domain or other shared directory (ldap, etc), your master is likely not configured right. If you are using SSSD, it is also possible yo

[slurm-users] Scontrol update GRES stops working on Slurm 20.02.5

2020-09-29 Thread Chen Yang
Hi, I recently upgraded Slurm from 19.05.5 to 20.02.5 and I noticed that I am no longer able to use the scontrol update command on GRES unless the new count is 0. For example, if I have a GRES named "test" defined and in slurm.conf > NodeName=node1 Gres=test:2 Initially, I have: $ sinfo -n no

Re: [slurm-users] EXTERNAL: Re: Memory per CPU

2020-09-29 Thread Luecht, Jeff A
Here are the particulars asked for. The following is the pertinent information for our cluster and the job run. Note: server names, IP addresses and user IDs are anonymized. Slurm.conf == TaskPlugin=task/affinity # SCHEDULING SchedulerType=s

Re: [slurm-users] EXTERNAL: Re: Memory per CPU

2020-09-29 Thread Luecht, Jeff A
There are three pieces of information that may provide Useful: 1 - these are VMs and not physical servers 2 - the OS is RedHat 7.8 2 - As far as I can tell, hyperthreading is not enabled, but will check for sure 3 - when we ask for 15Gb memory - we will only get 1 CPU -Original Message- F

Re: [slurm-users] EXTERNAL: Re: Memory per CPU

2020-09-29 Thread Luecht, Jeff A
The following is the pertinent information for our cluster and the job run. Note: server names, IP addresses and user IDs are anonymized. Slurm.conf == TaskPlugin=task/affinity # SCHEDULING SchedulerType=sched/backfill SelectType=select/cons_

[slurm-users] Memory per CPU

2020-09-29 Thread Luecht, Jeff A
I am working on my first ever SLURM cluster build for use as a resource manager in a JupyterHub Development environment. I have configured the cluster for SelectType of 'select/con_res' with DefMemPerCPU and MaxMemPerCPU of 16Gb. The idea is to essentially provide for jobs that run in a 1 CPU/

Re: [slurm-users] EXTERNAL: Re: Memory per CPU

2020-09-29 Thread Thomas M. Payerle
I am not familiar with using Slurm with VMs, but do note that Slurm can behave a bit "unexpectedly" with memory constraints due to the memory consumed by OS, etc. E.g., if I had a 16 core machine with 64 GB of RAM and requested 16 cores with 4 GB/core, it would not fit on this machine because some

Re: [slurm-users] Memory per CPU

2020-09-29 Thread Michael Di Domenico
what leads you to believe that you're getting 2 CPU's instead of 1? 'scontrol show job ' would be a helpful first start. On Tue, Sep 29, 2020 at 9:56 AM Luecht, Jeff A wrote: > > I am working on my first ever SLURM cluster build for use as a resource > manager in a JupyterHub Development environ

Re: [slurm-users] Working with local licenses

2020-09-29 Thread Tager, Alexey
Hi Loris, Thank you for clarification Will figure out if we can apply this approach Thanks Alexey -Original Message- From: slurm-users On Behalf Of Loris Bennett Sent: Tuesday, September 29, 2020 9:37 AM To: Slurm User Community List Subject: RE: [EXTERNAL] [slurm-users] Working with

[slurm-users] error: user not found

2020-09-29 Thread Diego Zuccato
Hello all. One of the users is unable to submit jobs to our cluster. The first time he tries, he gets $ sbatch test.job sbatch: fatal: Invalid user id: 621049927 then: $ sbatch test.job sbatch: error: Batch job submission failed: Invalid account or account/partition combination specified In slur