Christopher,
This looks like a SLURM issue and Open MPI is (currently) out of the
picture.
What if you
srun --pack-group=0,1 hostname
Do you get a similar error ?
Cheers,
Gilles
On 10/10/2018 3:07 AM, Christopher Benjamin Coffey wrote:
Hi,
I have a user trying to setup a heterogene
Hi,
I have a user trying to setup a heterogeneous job with one MPI_COMM_WORLD with
the following:
==
#!/bin/bash
#SBATCH --job-name=hetero
#SBATCH --output=/scratch/cbc/hetero.txt
#SBATCH --time=2:00
#SBATCH --workdir=/scratch/cbc
#SBATCH --cpu
This should be quite easy..
if job_desc.min_cpus or job_desc.min_cpus < YOUR_NUMBER then
job_desc.partition = "YourPartition"
end
Check slurm.h definition of job_descriptior and (small self advert but
maybe helpful..) you can also check my blog post on job_submit/lua (
https://funinit.wordpress.
My group uses "-p short,batch" in our submit scripts. Jobs will try to run in
the short partition (4 hours or less, no more than 24 nodes per job) but if
there's a wait and enough nodes open up in regular batch partition (4 days /
844 nodes max), the job will run over there. Really helps back
Hello,
We are starting to think about developing a lua job submission script. For
example, we are keen to route jobs requiring no more than 1 compute node
(single core jobs and small parallel jobs) to a slurm shared partition. The
idea being that "small" jobs can share a small set of compute n