Hi Ashton,
On Sat, Sep 22, 2018 at 5:34 AM A wrote:
> So I'm wondering if 20% is enough, or whether it should scale by the number
> of single jobs I might be running at any one time. E.g. if I'm running 10
> jobs that all use 20 gb of ram, and I suspend, should I need 200 gb of swap?
Perhaps
Hi John! Thanks for the reply, lots to think about.
In terms of suspending/resuming, my situation might be a bit different than
other people. As I mentioned this is an install on a single node
workstation. This is my daily office machine. I run alot of python
processing scripts that have low CPU n
Ashton, on a compute node with 256Gbytes of RAM I would not
configure any swap at all. None.
I managed an SGI UV1 machine at an F1 team which had 1Tbyte of RAM -
and no swap.
Also our ICE clusters were diskless - SGI very smartly configured swap
over ISCSI - but we disabled this, the reason being
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1
On 09/21/2018 11:22 PM, Chris Samuel wrote:
> On Saturday, 22 September 2018 2:53:58 AM AEST Nicolas Bock wrote:
>
>> shows as requesting 1 CPU when in queue, but then allocates all
>> CPU cores once running. Why is that?
>
> Do you mean that Slurm
On Saturday, 22 September 2018 2:53:58 AM AEST Nicolas Bock wrote:
> shows as requesting 1 CPU when in queue, but then allocates all
> CPU cores once running. Why is that?
Do you mean that Slurm expands the cores requested to all the cores on the
node or allocates the node in exclusive mode, or
I have a single node slurm config on my workstation (18 cores, 256 gb ram,
40 Tb disk space). I recently just extended the array size to its current
config and am reconfiguring my LVM logical volumes.
I'm curious on people's thoughts on swap sizes for a node. Redhat these
days recommends up to 20%
Hi Nick
I'm curious what makes you think job is using all CPU cores once running?
Would you be able to share output from 'ps' command while job is running?:
ps -p -L -o pid,tid,psr,pcpu
Execute it on the compute node where your job is running.
is the process id of the job/task.
Thank you.
Hello Nick,
What is the result with only -n 1 ?
Could you provide your slurm.conf ?
A lot of parameters are involved in the allocation process. The choice between
a few cores or a whole node depends mostly of "SelectType" and "Shared" in the
partition definition.
Regards,
-Original Messag
Hi,
A job run with
sbatch --ntasks=1 \
--ntasks-per-node=1 \
--cpus-per-task=1 \ --ntasks-per-core=1 \
--sockets-per-core=1 \
--cores-per-socket=1 \
--threads-per-core=1
shows as requesting 1 CPU when in queue, but then allocates all
CPU cores once running. Why is that?
Any s
> Today I discovered that sbatch can also create job steps (albeit
awkwardly, via --jobid), and it can obviously run sbatch scripts... but as
far as I can tell, it cannot run synchronously!
Rats. I just discovered that sbatch has a "--wait" option, but it must
have been added after 2.2.4, because
Try running with "srun", not "mpirun"
Hello everybody,
i submit a job with sbatch command (sbatch myprog.sh). My prog.sh is
=
#!/bin/bash
#SBATCH --partition=part2
#SBATCH --ntasks=20
#SBATCH --nodelist=
#SBATCH --cpus-per-task=1
#SBAT
Hello everybody,
i submit a job with sbatch command (sbatch myprog.sh). My prog.sh is
=
#!/bin/bash
#SBATCH --partition=part2
#SBATCH --ntasks=20
#SBATCH --nodelist=
#SBATCH --cpus-per-task=1
#SBATCH --mem=
# Memory per node specification
12 matches
Mail list logo