It's supposed to, so it sounds like we have a bug in the connection failover
mechanism. I'll address it
On Jul 23, 2014, at 1:21 AM, Timur Ismagilov wrote:
> Thanks, Ralph!
> When I add --mca oob_tcp_if_include ib0 (where ib0 is infiniband interface
> from ifconfig) to mpirun it starts working
Thanks, Ralph!
When I add --mca oob_tcp_if_include ib0 (where ib0 is infiniband interface from
ifconfig) to mpirun it starts working correct!
Why OpenMPI doesn't do it itself?
Tue, 22 Jul 2014 11:26:16 -0700 от Ralph Castain :
>Okay, the problem is that the connection back to mpirun isn't getti
Okay, the problem is that the connection back to mpirun isn't getting thru. We
are trying on the 10.0.251.53 address - is that blocked, or should we be using
something else? If so, you might want to direct us by adding "-mca
oob_tcp_if_include foo", where foo is the interface you want us to use
With Open MPI 1.9a1r32252 (Jul 16, 2014 (nightly snapshot tarball)) i got this
output (same?):
$ salloc -N2 --exclusive -p test -J ompi
salloc: Granted job allocation 645686
$LD_PRELOAD=/mnt/data/users/dm2/vol3/semenov/_scratch/mxm/mxm-3.0/lib/libmxm.so
mpirun -mca mca_base_env_list 'LD_PRELO
Here it is:
$
LD_PRELOAD=/mnt/data/users/dm2/vol3/semenov/_scratch/mxm/mxm-3.0/lib/libmxm.so
mpirun -x LD_PRELOAD --mca plm_base_verbose 10 --debug-daemons -np 1 hello_c
[access1:29064] mca: base: components_register: registering plm components
[access1:29064] mca: base: components_register:
please add following flags to mpirun "--mca plm_base_verbose 10
--debug-daemons" and attach output.
Thx
On Wed, Jul 16, 2014 at 11:12 AM, Timur Ismagilov
wrote:
> Hello!
> I have Open MPI v1.9a1r32142 and slurm 2.5.6.
>
> I can not use mpirun after salloc:
>
> $salloc -N2 --exclusive -p test -J