I have already sent it On Thu, May 28, 2015 at 10:21 AM.
28 мая 2015 г. 20:07 пользователь Mike Dubman написал:it is fine to recompile OMPI from HPCx to apply site default (choice of job scheduler for example, OMPI from HPCX compiled with ssh support only, etc.).If ssh launcher is working on your
://e.mail.ru/compose/?mailto=mailto%3ami...@dev.mellanox.co.il>>
> wrote:
>
> Alina - could you please take a look?
> Thx
>
>
> -- Forwarded message --
> From: *Timur Ismagilov* <https://e.mail.ru/compose/?mailto=mailto%3atismagi...@mail.ru>>
> D
lib/libmxm.so -x MXM_LOG_LEVEL=data"
>>>
>>>Also, could you please attach the entire output of
>>>"$HPCX_MPI_DIR/bin/ompi_info -a"
>>>
>>>Thank you,
>>>Alina.
>>>
>>>On Tue, May 26, 2015 at 3:39 PM, Mike Dub
look?
>>Thx
>>
>>
>>-- Forwarded message --
>>From: Timur Ismagilov < tismagi...@mail.ru >
>>Date: Tue, May 26, 2015 at 12:40 PM
>>Subject: Re[12]: [OMPI users] MXM problem
>>To: Open MPI Users < us...@open-mpi.org >
>>
It does not work for single node:
1) host: $ $HPCX_MPI_DIR/bin/mpirun -x MXM_IB_PORTS=mlx4_0:1 -x
MXM_SHM_KCOPY_MODE=off -host node5 -mca pml yalla -x MXM_TLS=ud,self,shm
--prefix $HPCX_MPI_DIR -mca plm_base_verbose 5 -mca oob_base_verbose 10 -mca
rml_base_verbose 10 --debug-daemons -np 1 .
1. mxm_perf_test - OK.
2. no_tree_spawn - OK.
3. ompi yalla and "--mca pml cm --mca mtl mxm" still does not work (I use
prebuild ompi-1.8.5 from hpcx-v1.3.330)
3.a) host:$ $HPCX_MPI_DIR/bin/mpirun -x MXM_IB_PORTS=mlx4_0:1 -x
MXM_SHM_KCOPY_MODE=off -host node5,node153 --mca pml cm --mca mtl
scif is a OFA device from Intel.
can you please select export MXM_IB_PORTS=mlx4_0:1 explicitly and retry
On Mon, May 25, 2015 at 8:26 PM, Timur Ismagilov wrote:
> Hi, Mike,
> that is what i have:
>
> $ echo $LD_LIBRARY_PATH | tr ":" "\n"
> /gpfs/NETHOME/oivt1/nicevt/itf/sources/hpcx-v1.3.0-327-i
Hi, Mike,
that is what i have:
$ echo $LD_LIBRARY_PATH | tr ":" "\n"
/gpfs/NETHOME/oivt1/nicevt/itf/sources/hpcx-v1.3.0-327-icc-OFED-1.5.3-redhat6.2/fca/lib
/gpfs/NETHOME/oivt1/nicevt/itf/sources/hpcx-v1.3.0-327-icc-OFED-1.5.3-redhat6.2/hcoll/lib
/gpfs/NETHOME/oivt1/n
Hi Timur,
seems that yalla component was not found in your OMPI tree.
can it be that your mpirun is not from hpcx? Can you please check
LD_LIBRARY_PATH,PATH, LD_PRELOAD and OPAL_PREFIX that it is pointing to the
right mpirun?
Also, could you please check that yalla is present in the ompi_info -l 9
I can password-less ssh to all nodes:
base$ ssh node1
node1$ssh node2
Last login: Mon May 25 18:41:23
node2$ssh node3
Last login: Mon May 25 16:25:01
node3$ssh node4
Last login: Mon May 25 16:27:04
node4$
Is this correct?
In ompi-1.9 i do not have no-tree-spawn problem.
Понедельник, 25 мая 20
I can’t speak to the mxm problem, but the no-tree-spawn issue indicates that
you don’t have password-less ssh authorized between the compute nodes
> On May 25, 2015, at 8:55 AM, Timur Ismagilov wrote:
>
> Hello!
>
> I use ompi-v1.8.4 from hpcx-v1.3.0-327-icc-OFED-1.5.3-redhat6.2;
> OFED-1.5.4
Hello!
I use ompi-v1.8.4 from hpcx-v1.3.0-327-icc-OFED-1.5.3-redhat6.2;
OFED-1.5.4.1;
CentOS release 6.2;
infiniband 4x FDR
I have two problems:
1. I can not use mxm :
1.a) $mpirun --mca pml cm --mca mtl mxm -host node5,node14,node28,node29 -mca
plm_rsh_no_tree_spawn 1 -np 4 ./hello
-
12 matches
Mail list logo