It isn’t an issue as there is nothing wrong with OMPI. Your method of joining
the allocation is a problem. What you have done is to create a job step that
has only 1 slot/node. We have no choice but to honor that constraint and run
within it.
What you should be doing is to use salloc to create
For the time being, you can
srun --ntasks-per-node 24 --jobid=...
When joining the allocation.
This use case looks a bit convoluted to me, so i am not even sure we should
consider there is a bug in Open MPI.
Ralph, any thoughts ?
Cheers,
Gilles
Gilles Gouaillardet wrote:
>Thanks, now i can
Thanks, now i can reproduce the issue
Cheers,
Gilles
On 9/8/2017 5:20 PM, Maksym Planeta wrote:
I run start an interactive allocation and I just noticed that the problem
happens, when I join this allocation from another shell.
Here is how I join:
srun --pty --x11 --jobid=$(squeue -u $USE
by any chance, would you be able to test the latest openmpi 2.1.2rc3 ?
OpenMPI 2.1.0 is the latest on our cluster.
--
Regards,
Maksym Planeta
smime.p7s
Description: S/MIME Cryptographic Signature
___
users mailing list
users@lists.open-mpi.org
I run start an interactive allocation and I just noticed that the problem
happens, when I join this allocation from another shell.
Here is how I join:
srun --pty --x11 --jobid=$(squeue -u $USER -o %A | tail -n 1) bash
And here is how I create the allocation:
srun --pty --nodes 8 --ntasks-per-n
Maxsym,
can you please post your sbatch script ?
fwiw, i am unable to reproduce the issue with the latest v2.x from github.
by any chance, would you be able to test the latest openmpi 2.1.2rc3 ?
Cheers,
Gilles
On 9/8/2017 4:19 PM, Maksym Planeta wrote:
Indeed mpirun shows slots=1 per n
Indeed mpirun shows slots=1 per node, but I create allocation with
--ntasks-per-node 24, so I do have all cores of the node allocated.
When I use srun I can get all the cores.
On 09/07/2017 02:12 PM, r...@open-mpi.org wrote:
> My best guess is that SLURM has only allocated 2 slots, and we respec
My best guess is that SLURM has only allocated 2 slots, and we respect the RM
regardless of what you say in the hostfile. You can check this by adding
--display-allocation to your cmd line. You probably need to tell slurm to
allocate more cpus/node.
> On Sep 7, 2017, at 3:33 AM, Maksym Planeta
Hello,
I'm trying to tell OpenMPI how many processes per node I want to use, but
mpirun seems to ignore the configuration I provide.
I create following hostfile:
$ cat hostfile.16
taurusi6344 slots=16
taurusi6348 slots=16
And then start the app as follows:
$ mpirun --display-map -machinefil