Re: [OMPI users] OpenMPI2 + slurm

2018-11-30 Thread Lothar Brendel
On Fri, 23 Nov 2018 09:17:00 +0100 Lothar Brendel wrote: [...] > looking into orte/mca/ras/slurm/ras_slurm_module.c, I find that while > orte_ras_slurm_allocate() reads the value of SLURM_CPUS_PER_TASK into its > local variable cpus_per_task, it doesn't use it anywhere. Rather, the number > o

Re: [OMPI users] OpenMPI2 + slurm (Ralph H Castain)

2018-11-25 Thread Lothar Brendel
> Couple of comments. Your original cmd line: > > >> srun -n 2 mpirun MPI-hellow > > tells srun to launch two copies of mpirun, each of which is to run as many > processes as there are slots assigned to the allocation. srun will get an > allocation of two slots, and so you?ll get two concurre

Re: [OMPI users] OpenMPI2 + slurm

2018-11-23 Thread Ralph H Castain
Couple of comments. Your original cmd line: >> srun -n 2 mpirun MPI-hellow tells srun to launch two copies of mpirun, each of which is to run as many processes as there are slots assigned to the allocation. srun will get an allocation of two slots, and so you’ll get two concurrent MPI jobs, e

Re: [OMPI users] OpenMPI2 + slurm

2018-11-23 Thread Gilles Gouaillardet
Lothar, it seems you did not configure Open MPI with --with-pmi= If SLURM was built with PMIx support, then an other option is to use that. First, srun --mpi=list will show you the list of available MPI modules, and then you could srun --mpi=pmix_v2 ... MPI_Hellow If you believe that should be th