Re: [slurm-users] Heterogeneous job one MPI_COMM_WORLD

2018-10-09 Thread Gilles Gouaillardet
Christopher, This looks like a SLURM issue and Open MPI is (currently) out of the picture. What if you srun --pack-group=0,1 hostname Do you get a similar error ? Cheers, Gilles On 10/10/2018 3:07 AM, Christopher Benjamin Coffey wrote: Hi, I have a user trying to setup a heterogene

[slurm-users] Heterogeneous job one MPI_COMM_WORLD

2018-10-09 Thread Christopher Benjamin Coffey
Hi, I have a user trying to setup a heterogeneous job with one MPI_COMM_WORLD with the following: == #!/bin/bash #SBATCH --job-name=hetero #SBATCH --output=/scratch/cbc/hetero.txt #SBATCH --time=2:00 #SBATCH --workdir=/scratch/cbc #SBATCH --cpu

Re: [slurm-users] Help with developing a lua job submit script

2018-10-09 Thread Marcin Stolarek
This should be quite easy.. if job_desc.min_cpus or job_desc.min_cpus < YOUR_NUMBER then job_desc.partition = "YourPartition" end Check slurm.h definition of job_descriptior and (small self advert but maybe helpful..) you can also check my blog post on job_submit/lua ( https://funinit.wordpress.

Re: [slurm-users] [EXTERNAL] Help with developing a lua job submit script

2018-10-09 Thread Amspaugh, Glenn Gregory
My group uses "-p short,batch" in our submit scripts. Jobs will try to run in the short partition (4 hours or less, no more than 24 nodes per job) but if there's a wait and enough nodes open up in regular batch partition (4 days / 844 nodes max), the job will run over there. Really helps back

[slurm-users] Help with developing a lua job submit script

2018-10-09 Thread Baker D . J .
Hello, We are starting to think about developing a lua job submission script. For example, we are keen to route jobs requiring no more than 1 compute node (single core jobs and small parallel jobs) to a slurm shared partition. The idea being that "small" jobs can share a small set of compute n