Re: [slurm-users] swap size

2018-09-21 Thread Raymond Wan
Hi Ashton, On Sat, Sep 22, 2018 at 5:34 AM A wrote: > So I'm wondering if 20% is enough, or whether it should scale by the number > of single jobs I might be running at any one time. E.g. if I'm running 10 > jobs that all use 20 gb of ram, and I suspend, should I need 200 gb of swap? Perhaps

Re: [slurm-users] swap size

2018-09-21 Thread A
Hi John! Thanks for the reply, lots to think about. In terms of suspending/resuming, my situation might be a bit different than other people. As I mentioned this is an install on a single node workstation. This is my daily office machine. I run alot of python processing scripts that have low CPU n

Re: [slurm-users] swap size

2018-09-21 Thread John Hearns
Ashton, on a compute node with 256Gbytes of RAM I would not configure any swap at all. None. I managed an SGI UV1 machine at an F1 team which had 1Tbyte of RAM - and no swap. Also our ICE clusters were diskless - SGI very smartly configured swap over ISCSI - but we disabled this, the reason being

Re: [slurm-users] Job allocating more CPUs than requested

2018-09-21 Thread Ryan Novosielski
-BEGIN PGP SIGNED MESSAGE- Hash: SHA1 On 09/21/2018 11:22 PM, Chris Samuel wrote: > On Saturday, 22 September 2018 2:53:58 AM AEST Nicolas Bock wrote: > >> shows as requesting 1 CPU when in queue, but then allocates all >> CPU cores once running. Why is that? > > Do you mean that Slurm

Re: [slurm-users] Job allocating more CPUs than requested

2018-09-21 Thread Chris Samuel
On Saturday, 22 September 2018 2:53:58 AM AEST Nicolas Bock wrote: > shows as requesting 1 CPU when in queue, but then allocates all > CPU cores once running. Why is that? Do you mean that Slurm expands the cores requested to all the cores on the node or allocates the node in exclusive mode, or

[slurm-users] swap size

2018-09-21 Thread A
I have a single node slurm config on my workstation (18 cores, 256 gb ram, 40 Tb disk space). I recently just extended the array size to its current config and am reconfiguring my LVM logical volumes. I'm curious on people's thoughts on swap sizes for a node. Redhat these days recommends up to 20%

Re: [slurm-users] Job allocating more CPUs than requested

2018-09-21 Thread Dmitri Chebotarov
Hi Nick I'm curious what makes you think job is using all CPU cores once running? Would you be able to share output from 'ps' command while job is running?: ps -p -L -o pid,tid,psr,pcpu Execute it on the compute node where your job is running. is the process id of the job/task. Thank you.

Re: [slurm-users] Job allocating more CPUs than requested

2018-09-21 Thread DENOU, MEHDI
Hello Nick, What is the result with only -n 1 ? Could you provide your slurm.conf ? A lot of parameters are involved in the allocation process. The choice between a few cores or a whole node depends mostly of "SelectType" and "Shared" in the partition definition. Regards, -Original Messag

[slurm-users] Job allocating more CPUs than requested

2018-09-21 Thread Nicolas Bock
Hi, A job run with sbatch --ntasks=1 \ --ntasks-per-node=1 \ --cpus-per-task=1 \ --ntasks-per-core=1 \ --sockets-per-core=1 \ --cores-per-socket=1 \ --threads-per-core=1 shows as requesting 1 CPU when in queue, but then allocates all CPU cores once running. Why is that? Any s

Re: [slurm-users] Use a portion of resources already allocated for a script

2018-09-21 Thread Michael Lamparski
> Today I discovered that sbatch can also create job steps (albeit awkwardly, via --jobid), and it can obviously run sbatch scripts... but as far as I can tell, it cannot run synchronously! Rats. I just discovered that sbatch has a "--wait" option, but it must have been added after 2.2.4, because

Re: [slurm-users] scancel problem

2018-09-21 Thread sysadmin.caos
Try running with "srun", not "mpirun" Hello everybody, i submit a job with sbatch command (sbatch myprog.sh). My prog.sh is = #!/bin/bash #SBATCH --partition=part2 #SBATCH --ntasks=20 #SBATCH --nodelist= #SBATCH --cpus-per-task=1 #SBAT

[slurm-users] scancel problem

2018-09-21 Thread Antonis Potirakis
Hello everybody, i submit a job with sbatch command (sbatch myprog.sh). My prog.sh is = #!/bin/bash #SBATCH --partition=part2 #SBATCH --ntasks=20 #SBATCH --nodelist= #SBATCH --cpus-per-task=1 #SBATCH --mem= # Memory per node specification