Re: [OMPI users] Slot count parameter in hostfile ignored

2017-09-08 Thread r...@open-mpi.org
It isn’t an issue as there is nothing wrong with OMPI. Your method of joining the allocation is a problem. What you have done is to create a job step that has only 1 slot/node. We have no choice but to honor that constraint and run within it. What you should be doing is to use salloc to create

Re: [OMPI users] OMPI users] Slot count parameter in hostfile ignored

2017-09-08 Thread Gilles Gouaillardet
For the time being, you can srun --ntasks-per-node 24 --jobid=... When joining the allocation. This use case looks a bit convoluted to me, so i am not even sure we should consider there is a bug in Open MPI. Ralph, any thoughts ? Cheers, Gilles Gilles Gouaillardet wrote: >Thanks, now i can

Re: [OMPI users] Slot count parameter in hostfile ignored

2017-09-08 Thread Gilles Gouaillardet
Thanks, now i can reproduce the issue Cheers, Gilles On 9/8/2017 5:20 PM, Maksym Planeta wrote: I run start an interactive allocation and I just noticed that the problem happens, when I join this allocation from another shell. Here is how I join: srun --pty --x11 --jobid=$(squeue -u $USE

Re: [OMPI users] Slot count parameter in hostfile ignored

2017-09-08 Thread Maksym Planeta
by any chance, would you be able to test the latest openmpi 2.1.2rc3 ? OpenMPI 2.1.0 is the latest on our cluster. -- Regards, Maksym Planeta smime.p7s Description: S/MIME Cryptographic Signature ___ users mailing list users@lists.open-mpi.org

Re: [OMPI users] Slot count parameter in hostfile ignored

2017-09-08 Thread Maksym Planeta
I run start an interactive allocation and I just noticed that the problem happens, when I join this allocation from another shell. Here is how I join: srun --pty --x11 --jobid=$(squeue -u $USER -o %A | tail -n 1) bash And here is how I create the allocation: srun --pty --nodes 8 --ntasks-per-n

Re: [OMPI users] Slot count parameter in hostfile ignored

2017-09-08 Thread Gilles Gouaillardet
Maxsym, can you please post your sbatch script ? fwiw, i am unable to reproduce the issue with the latest v2.x from github. by any chance, would you be able to test the latest openmpi 2.1.2rc3 ? Cheers, Gilles On 9/8/2017 4:19 PM, Maksym Planeta wrote: Indeed mpirun shows slots=1 per n

Re: [OMPI users] Slot count parameter in hostfile ignored

2017-09-08 Thread Maksym Planeta
Indeed mpirun shows slots=1 per node, but I create allocation with --ntasks-per-node 24, so I do have all cores of the node allocated. When I use srun I can get all the cores. On 09/07/2017 02:12 PM, r...@open-mpi.org wrote: > My best guess is that SLURM has only allocated 2 slots, and we respec

Re: [OMPI users] Slot count parameter in hostfile ignored

2017-09-07 Thread r...@open-mpi.org
My best guess is that SLURM has only allocated 2 slots, and we respect the RM regardless of what you say in the hostfile. You can check this by adding --display-allocation to your cmd line. You probably need to tell slurm to allocate more cpus/node. > On Sep 7, 2017, at 3:33 AM, Maksym Planeta

[OMPI users] Slot count parameter in hostfile ignored

2017-09-07 Thread Maksym Planeta
Hello, I'm trying to tell OpenMPI how many processes per node I want to use, but mpirun seems to ignore the configuration I provide. I create following hostfile: $ cat hostfile.16 taurusi6344 slots=16 taurusi6348 slots=16 And then start the app as follows: $ mpirun --display-map -machinefil