-mca pml_base_verbose 10 you should see:
select: component yalla selected for mxm debug info, please add: -x LD_PRELOAD=$MXM_DIR/lib/libmxm-debug.so -x MXM_LOG_LEVEL=debug On Tue, Apr 28, 2015 at 7:54 AM, Subhra Mazumdar <subhramazumd...@gmail.com> wrote: > Is there any way (probe or trace or other) to sanity check that I am > indeed using #2 ? > > Subhra > > On Fri, Apr 24, 2015 at 12:55 AM, Mike Dubman <mi...@dev.mellanox.co.il> > wrote: > >> yes >> >> #1 - ob1 as pml, openib openib as btl (default: rc) >> #2 - yalla as pml, mxm as IB library (default: ud, use "-x >> MXM_TLS=rc,self,shm" for rc) >> #3 - cm as pml, mxm as mtl and mxm as a transport (default: ud, use >> params from #2 for rc) >> >> On Fri, Apr 24, 2015 at 10:46 AM, Subhra Mazumdar < >> subhramazumd...@gmail.com> wrote: >> >>> I am a little confused now, I ran 3 different ways and got 3 different >>> performance from best to worse in following order: >>> >>> 1) mpirun --allow-run-as-root --mca pml ob1 -n 1 /root/backend >>> localhost : -x LD_PRELOAD=/root/libci.so -n 1 /root/app2 >>> >>> 2) mpirun --allow-run-as-root -n 1 /root/backend localhost : -x >>> LD_PRELOAD=/root/libci.so -n 1 /root/app2 >>> >>> 3) mpirun --allow-run-as-root --mca pml cm --mca mtl mxm -n 1 >>> /root/backend localhost : -x LD_PRELOAD=/root/libci.so -n 1 /root/app2 >>> >>> Are all of the above using infiniband but in different ways? >>> >>> Thanks, >>> Subhra. >>> >>> >>> >>> On Thu, Apr 23, 2015 at 11:57 PM, Mike Dubman <mi...@dev.mellanox.co.il> >>> wrote: >>> >>>> HPCX package uses pml "yalla" by default (part of ompi master branch, >>>> not in v1.8). >>>> So, "-mca mtl mxm" has no effect, unless "-mca pml cm" specified to >>>> disable "pml yalla" and let mtl layer to play. >>>> >>>> >>>> >>>> On Fri, Apr 24, 2015 at 6:36 AM, Subhra Mazumdar < >>>> subhramazumd...@gmail.com> wrote: >>>> >>>>> I changed my downloaded MOFED version to match the one installed on >>>>> the node and now the error goes away and it runs fine. But I still have a >>>>> question, I get the exact same performance on all the below 3 cases: >>>>> >>>>> 1) mpirun --allow-run-as-root --mca mtl mxm -mca mtl_mxm_np 0 -x >>>>> MXM_TLS=self,shm,rc,ud -n 1 /root/backend localhost : -x >>>>> LD_PRELOAD=/root/libci.so -n 1 /root/app2 >>>>> >>>>> 2) mpirun --allow-run-as-root --mca mtl mxm -n 1 /root/backend >>>>> localhost : -x LD_PRELOAD=/root/libci.so -n 1 /root/app2 >>>>> >>>>> 3) mpirun --allow-run-as-root --mca mtl ^mxm -n 1 /root/backend >>>>> localhost : -x LD_PRELOAD=/root/libci.so -n 1 /root/app2 >>>>> >>>>> Seems like it doesn't matter if I use mxm, not use mxm or use it with >>>>> reliable connection (RC). How can I be sure I am indeed using mxm over >>>>> infiniband? >>>>> >>>>> Thanks, >>>>> Subhra. >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> On Thu, Apr 23, 2015 at 1:06 AM, Mike Dubman <mi...@dev.mellanox.co.il >>>>> > wrote: >>>>> >>>>>> /usr/bin/ofed_info >>>>>> >>>>>> So, the OFED on your system is not MellanoxOFED 2.4.x but smth else. >>>>>> >>>>>> try #rpm -qi libibverbs >>>>>> >>>>>> >>>>>> On Thu, Apr 23, 2015 at 7:47 AM, Subhra Mazumdar < >>>>>> subhramazumd...@gmail.com> wrote: >>>>>> >>>>>>> Hi, >>>>>>> >>>>>>> where is the command ofed_info located? I searched from / but didn't >>>>>>> find it. >>>>>>> >>>>>>> Subhra. >>>>>>> >>>>>>> On Tue, Apr 21, 2015 at 10:43 PM, Mike Dubman < >>>>>>> mi...@dev.mellanox.co.il> wrote: >>>>>>> >>>>>>>> cool, progress! >>>>>>>> >>>>>>>> >>1429676565.124664] sys.c:719 MXM WARN Conflicting CPU >>>>>>>> frequencies detected, using: 2601.00 >>>>>>>> >>>>>>>> means that cpu governor on your machine is not on "performance" >>>>>>>> mode >>>>>>>> >>>>>>>> >> MXM ERROR ibv_query_device() returned 38: Function not >>>>>>>> implemented >>>>>>>> >>>>>>>> indicates that ofed installed on your nodes is not indeed >>>>>>>> 2.4.-1.0.0 or there is a mismatch between ofed kernel drivers version >>>>>>>> and >>>>>>>> ofed userspace libraries version. >>>>>>>> or you have multiple ofed libraries installed on your node and use >>>>>>>> incorrect one. >>>>>>>> could you please check that ofed_info -s indeed prints mofed >>>>>>>> 2.4-1.0.0? >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Wed, Apr 22, 2015 at 7:59 AM, Subhra Mazumdar < >>>>>>>> subhramazumd...@gmail.com> wrote: >>>>>>>> >>>>>>>>> Hi, >>>>>>>>> >>>>>>>>> I compiled the openmpi that comes inside the mellanox hpcx package >>>>>>>>> with mxm support instead of separately downloaded openmpi. I also >>>>>>>>> used the >>>>>>>>> environment as in the README so that no LD_PRELOAD (except our own >>>>>>>>> library >>>>>>>>> which is unrelated) is needed. Now it runs fine (no segfault) but we >>>>>>>>> get >>>>>>>>> same errors as before (saying initialization of MXM library failed). >>>>>>>>> Is it >>>>>>>>> using MXM successfully? >>>>>>>>> >>>>>>>>> [root@JARVICE >>>>>>>>> hpcx-v1.2.0-325-gcc-MLNX_OFED_LINUX-2.4-1.0.0-redhat6.5]# mpirun >>>>>>>>> --allow-run-as-root --mca mtl mxm -n 1 /root/backend localhost : -x >>>>>>>>> LD_PRELOAD=/root/libci.so -n 1 /root/app2 >>>>>>>>> >>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>> WARNING: a request was made to bind a process. While the system >>>>>>>>> supports binding the process itself, at least one node does NOT >>>>>>>>> support binding memory to the process location. >>>>>>>>> >>>>>>>>> Node: JARVICE >>>>>>>>> >>>>>>>>> This usually is due to not having the required NUMA support >>>>>>>>> installed >>>>>>>>> on the node. In some Linux distributions, the required support is >>>>>>>>> contained in the libnumactl and libnumactl-devel packages. >>>>>>>>> This is a warning only; your job will continue, though performance >>>>>>>>> may be degraded. >>>>>>>>> >>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>> i am backend >>>>>>>>> [1429676565.121218] sys.c:719 MXM WARN Conflicting CPU >>>>>>>>> frequencies detected, using: 2601.00 >>>>>>>>> [1429676565.122937] [JARVICE:14767:0] ib_dev.c:445 MXM WARN >>>>>>>>> failed call to ibv_exp_use_priv_env(): Function not implemented >>>>>>>>> [1429676565.122950] [JARVICE:14767:0] ib_dev.c:456 MXM >>>>>>>>> ERROR ibv_query_device() returned 38: Function not implemented >>>>>>>>> [1429676565.123535] [JARVICE:14767:0] ib_dev.c:445 MXM WARN >>>>>>>>> failed call to ibv_exp_use_priv_env(): Function not implemented >>>>>>>>> [1429676565.123543] [JARVICE:14767:0] ib_dev.c:456 MXM >>>>>>>>> ERROR ibv_query_device() returned 38: Function not implemented >>>>>>>>> [1429676565.124664] sys.c:719 MXM WARN Conflicting CPU >>>>>>>>> frequencies detected, using: 2601.00 >>>>>>>>> [1429676565.126264] [JARVICE:14768:0] ib_dev.c:445 MXM WARN >>>>>>>>> failed call to ibv_exp_use_priv_env(): Function not implemented >>>>>>>>> [1429676565.126276] [JARVICE:14768:0] ib_dev.c:456 MXM >>>>>>>>> ERROR ibv_query_device() returned 38: Function not implemented >>>>>>>>> [1429676565.126812] [JARVICE:14768:0] ib_dev.c:445 MXM WARN >>>>>>>>> failed call to ibv_exp_use_priv_env(): Function not implemented >>>>>>>>> [1429676565.126821] [JARVICE:14768:0] ib_dev.c:456 MXM >>>>>>>>> ERROR ibv_query_device() returned 38: Function not implemented >>>>>>>>> >>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>> Initialization of MXM library failed. >>>>>>>>> >>>>>>>>> Error: Input/output error >>>>>>>>> >>>>>>>>> >>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>> >>>>>>>>> <application runs fine> >>>>>>>>> >>>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> Subhra. >>>>>>>>> >>>>>>>>> >>>>>>>>> On Sat, Apr 18, 2015 at 12:28 AM, Mike Dubman < >>>>>>>>> mi...@dev.mellanox.co.il> wrote: >>>>>>>>> >>>>>>>>>> could you please check that ofed_info -s indeed prints mofed >>>>>>>>>> 2.4-1.0.0? >>>>>>>>>> why LD_PRELOAD needed in your command line? Can you try >>>>>>>>>> >>>>>>>>>> module load hpcx >>>>>>>>>> mpirun -np $np test.exe >>>>>>>>>> ? >>>>>>>>>> >>>>>>>>>> On Sat, Apr 18, 2015 at 8:39 AM, Subhra Mazumdar < >>>>>>>>>> subhramazumd...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> I followed the instructions as in the README, now getting a >>>>>>>>>>> different error: >>>>>>>>>>> >>>>>>>>>>> [root@JARVICE >>>>>>>>>>> hpcx-v1.2.0-325-gcc-MLNX_OFED_LINUX-2.4-1.0.0-redhat6.5]# >>>>>>>>>>> ../openmpi-1.8.4/openmpinstall/bin/mpirun --allow-run-as-root --mca >>>>>>>>>>> mtl mxm >>>>>>>>>>> -x LD_PRELOAD="../openmpi-1.8.4/openmpinstall/lib/libmpi.so.1 >>>>>>>>>>> ./mxm/lib/libmxm.so.2" -n 1 ../backend localhost : -x >>>>>>>>>>> LD_PRELOAD="../openmpi-1.8.4/openmpinstall/lib/libmpi.so.1 >>>>>>>>>>> ./mxm/lib/libmxm.so.2 ../libci.so" -n 1 ../app2 >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>>>> >>>>>>>>>>> WARNING: a request was made to bind a process. While the system >>>>>>>>>>> >>>>>>>>>>> supports binding the process itself, at least one node does NOT >>>>>>>>>>> >>>>>>>>>>> support binding memory to the process location. >>>>>>>>>>> >>>>>>>>>>> Node: JARVICE >>>>>>>>>>> >>>>>>>>>>> This usually is due to not having the required NUMA support >>>>>>>>>>> installed >>>>>>>>>>> >>>>>>>>>>> on the node. In some Linux distributions, the required support is >>>>>>>>>>> >>>>>>>>>>> contained in the libnumactl and libnumactl-devel packages. >>>>>>>>>>> >>>>>>>>>>> This is a warning only; your job will continue, though >>>>>>>>>>> performance may be degraded. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>>>> >>>>>>>>>>> i am backend >>>>>>>>>>> >>>>>>>>>>> [1429334876.139452] [JARVICE:449 :0] ib_dev.c:445 MXM WARN >>>>>>>>>>> failed call to ibv_exp_use_priv_env(): Function not implemented >>>>>>>>>>> >>>>>>>>>>> [1429334876.139464] [JARVICE:449 :0] ib_dev.c:456 MXM >>>>>>>>>>> ERROR ibv_query_device() returned 38: Function not implemented >>>>>>>>>>> >>>>>>>>>>> [1429334876.139982] [JARVICE:449 :0] ib_dev.c:445 MXM WARN >>>>>>>>>>> failed call to ibv_exp_use_priv_env(): Function not implemented >>>>>>>>>>> >>>>>>>>>>> [1429334876.139990] [JARVICE:449 :0] ib_dev.c:456 MXM >>>>>>>>>>> ERROR ibv_query_device() returned 38: Function not implemented >>>>>>>>>>> >>>>>>>>>>> [1429334876.142649] [JARVICE:450 :0] ib_dev.c:445 MXM WARN >>>>>>>>>>> failed call to ibv_exp_use_priv_env(): Function not implemented >>>>>>>>>>> >>>>>>>>>>> [1429334876.142666] [JARVICE:450 :0] ib_dev.c:456 MXM >>>>>>>>>>> ERROR ibv_query_device() returned 38: Function not implemented >>>>>>>>>>> >>>>>>>>>>> [1429334876.143235] [JARVICE:450 :0] ib_dev.c:445 MXM WARN >>>>>>>>>>> failed call to ibv_exp_use_priv_env(): Function not implemented >>>>>>>>>>> >>>>>>>>>>> [1429334876.143243] [JARVICE:450 :0] ib_dev.c:456 MXM >>>>>>>>>>> ERROR ibv_query_device() returned 38: Function not implemented >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>>>> >>>>>>>>>>> Initialization of MXM library failed. >>>>>>>>>>> >>>>>>>>>>> Error: Input/output error >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>>>> >>>>>>>>>>> [JARVICE:449 :0] Caught signal 11 (Segmentation fault) >>>>>>>>>>> >>>>>>>>>>> [JARVICE:450 :0] Caught signal 11 (Segmentation fault) >>>>>>>>>>> >>>>>>>>>>> ==== backtrace ==== >>>>>>>>>>> >>>>>>>>>>> 2 0x000000000005640c mxm_handle_error() >>>>>>>>>>> >>>>>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u5-x86-64-MOFED-CHECKER/hpcx_root/src/hpcx-v1.2.0-325-gcc-MLNX_OFED_LINUX-2.4-1.0.0-redhat6.5/mxm-v3.2/src/mxm/util/debug/debug.c:641 >>>>>>>>>>> >>>>>>>>>>> 3 0x000000000005657c mxm_error_signal_handler() >>>>>>>>>>> >>>>>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u5-x86-64-MOFED-CHECKER/hpcx_root/src/hpcx-v1.2.0-325-gcc-MLNX_OFED_LINUX-2.4-1.0.0-redhat6.5/mxm-v3.2/src/mxm/util/debug/debug.c:616 >>>>>>>>>>> >>>>>>>>>>> 4 0x00000000000329a0 killpg() ??:0 >>>>>>>>>>> >>>>>>>>>>> 5 0x000000000004812c _IO_vfprintf() ??:0 >>>>>>>>>>> >>>>>>>>>>> 6 0x000000000006f6da vasprintf() ??:0 >>>>>>>>>>> >>>>>>>>>>> 7 0x0000000000059b3b opal_show_help_vstring() ??:0 >>>>>>>>>>> >>>>>>>>>>> 8 0x0000000000026630 orte_show_help() ??:0 >>>>>>>>>>> >>>>>>>>>>> 9 0x0000000000001a3f mca_bml_r2_add_procs() >>>>>>>>>>> >>>>>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u5-x86-64-MOFED-CHECKER/hpcx_root/src/hpcx-v1.2.0-325-gcc-MLNX_OFED_LINUX-2.4-1.0.0-redhat6.5/ompi-mellanox-v1.8/ompi/mca/bml/r2/bml_r2.c:409 >>>>>>>>>>> >>>>>>>>>>> 10 0x0000000000004475 mca_pml_ob1_add_procs() >>>>>>>>>>> >>>>>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u5-x86-64-MOFED-CHECKER/hpcx_root/src/hpcx-v1.2.0-325-gcc-MLNX_OFED_LINUX-2.4-1.0.0-redhat6.5/ompi-mellanox-v1.8/ompi/mca/pml/ob1/pml_ob1.c:332 >>>>>>>>>>> >>>>>>>>>>> 11 0x00000000000442f3 ompi_mpi_init() ??:0 >>>>>>>>>>> >>>>>>>>>>> 12 0x0000000000067cb0 PMPI_Init_thread() ??:0 >>>>>>>>>>> >>>>>>>>>>> 13 0x000000000000d0ca l_getLocalFromConfig() >>>>>>>>>>> /root/rain_ib/interposer/libciutils.c:83 >>>>>>>>>>> >>>>>>>>>>> 14 0x000000000000c7b4 __cudaRegisterFatBinary() >>>>>>>>>>> /root/rain_ib/interposer/libci.c:4055 >>>>>>>>>>> >>>>>>>>>>> 15 0x0000000000402b59 >>>>>>>>>>> _ZL70__sti____cudaRegisterAll_39_tmpxft_00000703_00000000_6_app2_cpp1_ii_hwv() >>>>>>>>>>> tmpxft_00000703_00000000-3_app2.cudafe1.cpp:0 >>>>>>>>>>> >>>>>>>>>>> 16 0x0000000000402dd6 __do_global_ctors_aux() crtstuff.c:0 >>>>>>>>>>> >>>>>>>>>>> =================== >>>>>>>>>>> >>>>>>>>>>> ==== backtrace ==== >>>>>>>>>>> >>>>>>>>>>> 2 0x000000000005640c mxm_handle_error() >>>>>>>>>>> >>>>>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u5-x86-64-MOFED-CHECKER/hpcx_root/src/hpcx-v1.2.0-325-gcc-MLNX_OFED_LINUX-2.4-1.0.0-redhat6.5/mxm-v3.2/src/mxm/util/debug/debug.c:641 >>>>>>>>>>> >>>>>>>>>>> 3 0x000000000005657c mxm_error_signal_handler() >>>>>>>>>>> >>>>>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u5-x86-64-MOFED-CHECKER/hpcx_root/src/hpcx-v1.2.0-325-gcc-MLNX_OFED_LINUX-2.4-1.0.0-redhat6.5/mxm-v3.2/src/mxm/util/debug/debug.c:616 >>>>>>>>>>> >>>>>>>>>>> 4 0x00000000000329a0 killpg() ??:0 >>>>>>>>>>> >>>>>>>>>>> 5 0x000000000004812c _IO_vfprintf() ??:0 >>>>>>>>>>> >>>>>>>>>>> 6 0x000000000006f6da vasprintf() ??:0 >>>>>>>>>>> >>>>>>>>>>> 7 0x0000000000059b3b opal_show_help_vstring() ??:0 >>>>>>>>>>> >>>>>>>>>>> 8 0x0000000000026630 orte_show_help() ??:0 >>>>>>>>>>> >>>>>>>>>>> 9 0x0000000000001a3f mca_bml_r2_add_procs() >>>>>>>>>>> >>>>>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u5-x86-64-MOFED-CHECKER/hpcx_root/src/hpcx-v1.2.0-325-gcc-MLNX_OFED_LINUX-2.4-1.0.0-redhat6.5/ompi-mellanox-v1.8/ompi/mca/bml/r2/bml_r2.c:409 >>>>>>>>>>> >>>>>>>>>>> 10 0x0000000000004475 mca_pml_ob1_add_procs() >>>>>>>>>>> >>>>>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u5-x86-64-MOFED-CHECKER/hpcx_root/src/hpcx-v1.2.0-325-gcc-MLNX_OFED_LINUX-2.4-1.0.0-redhat6.5/ompi-mellanox-v1.8/ompi/mca/pml/ob1/pml_ob1.c:332 >>>>>>>>>>> >>>>>>>>>>> 11 0x00000000000442f3 ompi_mpi_init() ??:0 >>>>>>>>>>> >>>>>>>>>>> 12 0x0000000000067cb0 PMPI_Init_thread() ??:0 >>>>>>>>>>> >>>>>>>>>>> 13 0x0000000000404fdf main() >>>>>>>>>>> /root/rain_ib/backend/backend.c:1237 >>>>>>>>>>> >>>>>>>>>>> 14 0x000000000001ed1d __libc_start_main() ??:0 >>>>>>>>>>> >>>>>>>>>>> 15 0x0000000000402db9 _start() ??:0 >>>>>>>>>>> >>>>>>>>>>> =================== >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>>>> >>>>>>>>>>> mpirun noticed that process rank 1 with PID 450 on node JARVICE >>>>>>>>>>> exited on signal 11 (Segmentation fault). >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>>>> >>>>>>>>>>> [JARVICE:00447] 1 more process has sent help message >>>>>>>>>>> help-mtl-mxm.txt / mxm init >>>>>>>>>>> >>>>>>>>>>> [JARVICE:00447] Set MCA parameter "orte_base_help_aggregate" to >>>>>>>>>>> 0 to see all help / error messages >>>>>>>>>>> >>>>>>>>>>> [root@JARVICE >>>>>>>>>>> hpcx-v1.2.0-325-gcc-MLNX_OFED_LINUX-2.4-1.0.0-redhat6.5]# >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Subhra. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Mon, Apr 13, 2015 at 10:58 PM, Mike Dubman < >>>>>>>>>>> mi...@dev.mellanox.co.il> wrote: >>>>>>>>>>> >>>>>>>>>>>> Have you followed installation steps from README (Also here for >>>>>>>>>>>> reference http://bgate.mellanox.com/products/hpcx/README.txt) >>>>>>>>>>>> >>>>>>>>>>>> ... >>>>>>>>>>>> >>>>>>>>>>>> * Load OpenMPI/OpenSHMEM v1.8 based package: >>>>>>>>>>>> >>>>>>>>>>>> % source $HPCX_HOME/hpcx-init.sh >>>>>>>>>>>> % hpcx_load >>>>>>>>>>>> % env | grep HPCX >>>>>>>>>>>> % mpirun -np 2 $HPCX_MPI_TESTS_DIR/examples/hello_usempi >>>>>>>>>>>> % oshrun -np 2 $HPCX_MPI_TESTS_DIR/examples/hello_oshmem >>>>>>>>>>>> % hpcx_unload >>>>>>>>>>>> >>>>>>>>>>>> 3. Load HPCX environment from modules >>>>>>>>>>>> >>>>>>>>>>>> * Load OpenMPI/OpenSHMEM based package: >>>>>>>>>>>> >>>>>>>>>>>> % module use $HPCX_HOME/modulefiles >>>>>>>>>>>> % module load hpcx >>>>>>>>>>>> % mpirun -np 2 $HPCX_MPI_TESTS_DIR/examples/hello_c >>>>>>>>>>>> % oshrun -np 2 $HPCX_MPI_TESTS_DIR/examples/hello_oshmem >>>>>>>>>>>> % module unload hpcx >>>>>>>>>>>> >>>>>>>>>>>> ... >>>>>>>>>>>> >>>>>>>>>>>> On Tue, Apr 14, 2015 at 5:42 AM, Subhra Mazumdar < >>>>>>>>>>>> subhramazumd...@gmail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> I am using 2.4-1.0.0 mellanox ofed. >>>>>>>>>>>>> >>>>>>>>>>>>> I downloaded mofed tarball >>>>>>>>>>>>> hpcx-v1.2.0-325-gcc-MLNX_OFED_LINUX-2.4-1.0.0-redhat6.5.tar and >>>>>>>>>>>>> extracted >>>>>>>>>>>>> it. It has mxm directory. >>>>>>>>>>>>> >>>>>>>>>>>>> hpcx-v1.2.0-325-[root@JARVICE ~]# ls >>>>>>>>>>>>> hpcx-v1.2.0-325-gcc-MLNX_OFED_LINUX-2.4-1.0.0-redhat6.5 >>>>>>>>>>>>> archive fca hpcx-init-ompi-mellanox-v1.8.sh ibprof >>>>>>>>>>>>> modulefiles ompi-mellanox-v1.8 sources VERSION >>>>>>>>>>>>> bupc-master hcoll hpcx-init.sh knem >>>>>>>>>>>>> mxm README.txt utils >>>>>>>>>>>>> >>>>>>>>>>>>> I tried using LD_PRELOAD for libmxm, but getting a different >>>>>>>>>>>>> error stack now as following >>>>>>>>>>>>> >>>>>>>>>>>>> [root@JARVICE ~]# ./openmpi-1.8.4/openmpinstall/bin/mpirun >>>>>>>>>>>>> --allow-run-as-root --mca mtl mxm -x >>>>>>>>>>>>> LD_PRELOAD="./openmpi-1.8.4/openmpinstall/lib/libmpi.so.1 >>>>>>>>>>>>> ./hpcx-v1.2.0-325-gcc-MLNX_OFED_LINUX-2.4-1.0.0-redhat6.5/mxm/lib/libmxm.so.2" >>>>>>>>>>>>> -n 1 ./backend localhost : -x >>>>>>>>>>>>> LD_PRELOAD="./openmpi-1.8.4/openmpinstall/lib/libmpi.so.1 >>>>>>>>>>>>> ./hpcx-v1.2.0-325-gcc-MLNX_OFED_LINUX-2.4-1.0.0-redhat6.5/mxm/lib/libmxm.so.2 >>>>>>>>>>>>> ./libci.so" -n 1 ./app2 >>>>>>>>>>>>> i am backend >>>>>>>>>>>>> [JARVICE:00564] mca: base: components_open: component pml / cm >>>>>>>>>>>>> open function failed >>>>>>>>>>>>> [JARVICE:564 :0] Caught signal 11 (Segmentation fault) >>>>>>>>>>>>> [JARVICE:00565] mca: base: components_open: component pml / cm >>>>>>>>>>>>> open function failed >>>>>>>>>>>>> [JARVICE:565 :0] Caught signal 11 (Segmentation fault) >>>>>>>>>>>>> ==== backtrace ==== >>>>>>>>>>>>> 2 0x000000000005640c mxm_handle_error() >>>>>>>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u5-x86-64-MOFED-CHECKER/hpcx_root/src/hpcx-v1.2.0-325-gcc-MLNX_OFED_LINUX-2.4-1.0.0-redhat6.5/mxm-v3.2/src/mxm/util/debug/debug.c:641 >>>>>>>>>>>>> 3 0x000000000005657c mxm_error_signal_handler() >>>>>>>>>>>>> /scrap/jenkins/workspace/hpc-power-pack/label/r-vmb-rhel6-u5-x86-64-MOFED-CHECKER/hpcx_root/src/hpcx-v1.2.0-325-gcc-MLNX_OFED_LINUX-2.4-1.0.0-redhat6.5/mxm-v3.2/src/mxm/util/debug/debug.c:616 >>>>>>>>>>>>> 4 0x00000000000329a0 killpg() ??:0 >>>>>>>>>>>>> 5 0x0000000000045491 mca_base_components_close() ??:0 >>>>>>>>>>>>> 6 0x000000000004e99a mca_base_framework_close() ??:0 >>>>>>>>>>>>> 7 0x0000000000045431 mca_base_component_close() ??:0 >>>>>>>>>>>>> 8 0x000000000004515c mca_base_framework_components_open() >>>>>>>>>>>>> ??:0 >>>>>>>>>>>>> 9 0x00000000000a0de9 mca_pml_base_open() pml_base_frame.c:0 >>>>>>>>>>>>> 10 0x000000000004eb1c mca_base_framework_open() ??:0 >>>>>>>>>>>>> 11 0x0000000000043eb3 ompi_mpi_init() ??:0 >>>>>>>>>>>>> 12 0x0000000000067cb0 PMPI_Init_thread() ??:0 >>>>>>>>>>>>> 13 0x0000000000404fdf main() >>>>>>>>>>>>> /root/rain_ib/backend/backend.c:1237 >>>>>>>>>>>>> 14 0x000000000001ed1d __libc_start_main() ??:0 >>>>>>>>>>>>> 15 0x0000000000402db9 _start() ??:0 >>>>>>>>>>>>> =================== >>>>>>>>>>>>> >>>>>>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>>>>>> A requested component was not found, or was unable to be >>>>>>>>>>>>> opened. This >>>>>>>>>>>>> means that this component is either not installed or is unable >>>>>>>>>>>>> to be >>>>>>>>>>>>> used on your system (e.g., sometimes this means that shared >>>>>>>>>>>>> libraries >>>>>>>>>>>>> that the component requires are unable to be found/loaded). >>>>>>>>>>>>> Note that >>>>>>>>>>>>> Open MPI stopped checking at the first component that it did >>>>>>>>>>>>> not find. >>>>>>>>>>>>> >>>>>>>>>>>>> Host: JARVICE >>>>>>>>>>>>> Framework: mtl >>>>>>>>>>>>> Component: mxm >>>>>>>>>>>>> >>>>>>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>>>>>> >>>>>>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>>>>>> mpirun noticed that process rank 0 with PID 564 on node >>>>>>>>>>>>> JARVICE exited on signal 11 (Segmentation fault). >>>>>>>>>>>>> >>>>>>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>>>>>> [JARVICE:00562] 1 more process has sent help message >>>>>>>>>>>>> help-mca-base.txt / find-available:not-valid >>>>>>>>>>>>> [JARVICE:00562] Set MCA parameter "orte_base_help_aggregate" >>>>>>>>>>>>> to 0 to see all help / error messages >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Subhra >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Sun, Apr 12, 2015 at 10:48 PM, Mike Dubman < >>>>>>>>>>>>> mi...@dev.mellanox.co.il> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> seems like mxm was not found in your ld_library_path. >>>>>>>>>>>>>> >>>>>>>>>>>>>> what mofed version do you use? >>>>>>>>>>>>>> does it have /opt/mellanox/mxm in it? >>>>>>>>>>>>>> You could just run mpirun from HPCX package which looks for >>>>>>>>>>>>>> mxm internally and recompile ompi as mentioned in README. >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Mon, Apr 13, 2015 at 3:24 AM, Subhra Mazumdar < >>>>>>>>>>>>>> subhramazumd...@gmail.com> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I used mxm mtl as follows but getting segfault. It says mxm >>>>>>>>>>>>>>> component not found but I have compiled openmpi with mxm. Any >>>>>>>>>>>>>>> idea what I >>>>>>>>>>>>>>> might be missing? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> [root@JARVICE ~]# ./openmpi-1.8.4/openmpinstall/bin/mpirun >>>>>>>>>>>>>>> --allow-run-as-root --mca pml cm --mca mtl mxm -n 1 -x >>>>>>>>>>>>>>> LD_PRELOAD=./openmpi-1.8.4/openmpinstall/lib/libmpi.so.1 >>>>>>>>>>>>>>> ./backend >>>>>>>>>>>>>>> localhosst : -n 1 -x LD_PRELOAD="./libci.so >>>>>>>>>>>>>>> ./openmpi-1.8.4/openmpinstall/lib/libmpi.so.1" ./app2 >>>>>>>>>>>>>>> i am backend >>>>>>>>>>>>>>> [JARVICE:08398] *** Process received signal *** >>>>>>>>>>>>>>> [JARVICE:08398] Signal: Segmentation fault (11) >>>>>>>>>>>>>>> [JARVICE:08398] Signal code: Address not mapped (1) >>>>>>>>>>>>>>> [JARVICE:08398] Failing at address: 0x10 >>>>>>>>>>>>>>> [JARVICE:08398] [ 0] >>>>>>>>>>>>>>> /lib64/libpthread.so.0(+0xf710)[0x7ff8d0ddb710] >>>>>>>>>>>>>>> [JARVICE:08398] [ 1] >>>>>>>>>>>>>>> /root/openmpi-1.8.4/openmpinstall/lib/libopen-pal.so.6(mca_base_components_close+0x21)[0x7ff8cf9ae491] >>>>>>>>>>>>>>> [JARVICE:08398] [ 2] >>>>>>>>>>>>>>> /root/openmpi-1.8.4/openmpinstall/lib/libopen-pal.so.6(mca_base_framework_close+0x6a)[0x7ff8cf9b799a] >>>>>>>>>>>>>>> [JARVICE:08398] [ 3] >>>>>>>>>>>>>>> /root/openmpi-1.8.4/openmpinstall/lib/libopen-pal.so.6(mca_base_component_close+0x21)[0x7ff8cf9ae431] >>>>>>>>>>>>>>> [JARVICE:08398] [ 4] >>>>>>>>>>>>>>> /root/openmpi-1.8.4/openmpinstall/lib/libopen-pal.so.6(mca_base_framework_components_open+0x11c)[0x7ff8cf9ae15c] >>>>>>>>>>>>>>> [JARVICE:08398] [ 5] >>>>>>>>>>>>>>> ./openmpi-1.8.4/openmpinstall/lib/libmpi.so.1(+0xa0de9)[0x7ff8d1089de9] >>>>>>>>>>>>>>> [JARVICE:08398] [ 6] >>>>>>>>>>>>>>> /root/openmpi-1.8.4/openmpinstall/lib/libopen-pal.so.6(mca_base_framework_open+0x7c)[0x7ff8cf9b7b1c] >>>>>>>>>>>>>>> [JARVICE:08398] [ 7] [JARVICE:08398] mca: base: >>>>>>>>>>>>>>> components_open: component pml / cm open function failed >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> ./openmpi-1.8.4/openmpinstall/lib/libmpi.so.1(ompi_mpi_init+0x4b3)[0x7ff8d102ceb3] >>>>>>>>>>>>>>> [JARVICE:08398] [ 8] >>>>>>>>>>>>>>> ./openmpi-1.8.4/openmpinstall/lib/libmpi.so.1(PMPI_Init_thread+0x100)[0x7ff8d1050cb0] >>>>>>>>>>>>>>> [JARVICE:08398] [ 9] ./backend[0x404fdf] >>>>>>>>>>>>>>> [JARVICE:08398] [10] >>>>>>>>>>>>>>> /lib64/libc.so.6(__libc_start_main+0xfd)[0x7ff8cfeded1d] >>>>>>>>>>>>>>> [JARVICE:08398] [11] ./backend[0x402db9] >>>>>>>>>>>>>>> [JARVICE:08398] *** End of error message *** >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>>>>>>>> A requested component was not found, or was unable to be >>>>>>>>>>>>>>> opened. This >>>>>>>>>>>>>>> means that this component is either not installed or is >>>>>>>>>>>>>>> unable to be >>>>>>>>>>>>>>> used on your system (e.g., sometimes this means that shared >>>>>>>>>>>>>>> libraries >>>>>>>>>>>>>>> that the component requires are unable to be found/loaded). >>>>>>>>>>>>>>> Note that >>>>>>>>>>>>>>> Open MPI stopped checking at the first component that it did >>>>>>>>>>>>>>> not find. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Host: JARVICE >>>>>>>>>>>>>>> Framework: mtl >>>>>>>>>>>>>>> Component: mxm >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>>>>>>>> mpirun noticed that process rank 0 with PID 8398 on node >>>>>>>>>>>>>>> JARVICE exited on signal 11 (Segmentation fault). >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Subhra. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Fri, Apr 10, 2015 at 12:12 AM, Mike Dubman < >>>>>>>>>>>>>>> mi...@dev.mellanox.co.il> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> no need IPoIB, mxm uses native IB. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Please see HPCX (pre-compiled ompi, integrated with MXM and >>>>>>>>>>>>>>>> FCA) README file for details how to compile/select. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> The default transport is UD for internode communication and >>>>>>>>>>>>>>>> shared-memory for intra-node. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> http://bgate,mellanox.com/products/hpcx/ >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Also, mxm included in the Mellanox OFED. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On Fri, Apr 10, 2015 at 5:26 AM, Subhra Mazumdar < >>>>>>>>>>>>>>>> subhramazumd...@gmail.com> wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Does ipoib need to be configured on the ib cards for mxm >>>>>>>>>>>>>>>>> (I have a separate ethernet connection too)? Also are there >>>>>>>>>>>>>>>>> special flags >>>>>>>>>>>>>>>>> in mpirun to select from UD/RC/DC? What is the default? >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>> Subhra. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On Tue, Mar 31, 2015 at 9:46 AM, Mike Dubman < >>>>>>>>>>>>>>>>> mi...@dev.mellanox.co.il> wrote: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>>>> mxm uses IB rdma/roce technologies. Once can select >>>>>>>>>>>>>>>>>> UD/RC/DC transports to be used in mxm. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> By selecting mxm, all MPI p2p routines will be mapped to >>>>>>>>>>>>>>>>>> appropriate mxm functions. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> M >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> On Mon, Mar 30, 2015 at 7:32 PM, Subhra Mazumdar < >>>>>>>>>>>>>>>>>> subhramazumd...@gmail.com> wrote: >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Hi MIke, >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Does the mxm mtl use infiniband rdma? Also from >>>>>>>>>>>>>>>>>>> programming perspective, do I need to use anything else >>>>>>>>>>>>>>>>>>> other than >>>>>>>>>>>>>>>>>>> MPI_Send/MPI_Recv? >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>> Subhra. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> On Sun, Mar 29, 2015 at 11:14 PM, Mike Dubman < >>>>>>>>>>>>>>>>>>> mi...@dev.mellanox.co.il> wrote: >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>>>>>> openib btl does not support this thread model. >>>>>>>>>>>>>>>>>>>> You can use OMPI w/ mxm (-mca mtl mxm) and multiple >>>>>>>>>>>>>>>>>>>> thread mode lin 1.8 x series or (-mca pml yalla) in the >>>>>>>>>>>>>>>>>>>> master branch. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> M >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> On Mon, Mar 30, 2015 at 9:09 AM, Subhra Mazumdar < >>>>>>>>>>>>>>>>>>>> subhramazumd...@gmail.com> wrote: >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Can MPI_THREAD_MULTIPLE and openib btl work together >>>>>>>>>>>>>>>>>>>>> in open mpi 1.8.4? If so are there any command line >>>>>>>>>>>>>>>>>>>>> options needed during >>>>>>>>>>>>>>>>>>>>> run time? >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>>>> Subhra. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>>>>>> users mailing list >>>>>>>>>>>>>>>>>>>>> us...@open-mpi.org >>>>>>>>>>>>>>>>>>>>> Subscription: >>>>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>>>>>>>>>>> Link to this post: >>>>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/03/26574.php >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> Kind Regards, >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> M. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>>>>> users mailing list >>>>>>>>>>>>>>>>>>>> us...@open-mpi.org >>>>>>>>>>>>>>>>>>>> Subscription: >>>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>>>>>>>>>> Link to this post: >>>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/03/26575.php >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>>>> users mailing list >>>>>>>>>>>>>>>>>>> us...@open-mpi.org >>>>>>>>>>>>>>>>>>> Subscription: >>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>>>>>>>>> Link to this post: >>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/03/26580.php >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Kind Regards, >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> M. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>>> users mailing list >>>>>>>>>>>>>>>>>> us...@open-mpi.org >>>>>>>>>>>>>>>>>> Subscription: >>>>>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>>>>>>>> Link to this post: >>>>>>>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/03/26584.php >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>> users mailing list >>>>>>>>>>>>>>>>> us...@open-mpi.org >>>>>>>>>>>>>>>>> Subscription: >>>>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>>>>>>> Link to this post: >>>>>>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/04/26663.php >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Kind Regards, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> M. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>> users mailing list >>>>>>>>>>>>>>>> us...@open-mpi.org >>>>>>>>>>>>>>>> Subscription: >>>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>>>>>> Link to this post: >>>>>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/04/26665.php >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>> users mailing list >>>>>>>>>>>>>>> us...@open-mpi.org >>>>>>>>>>>>>>> Subscription: >>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>>>>> Link to this post: >>>>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/04/26686.php >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> >>>>>>>>>>>>>> Kind Regards, >>>>>>>>>>>>>> >>>>>>>>>>>>>> M. >>>>>>>>>>>>>> >>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>> users mailing list >>>>>>>>>>>>>> us...@open-mpi.org >>>>>>>>>>>>>> Subscription: >>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>>>> Link to this post: >>>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/04/26688.php >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>> users mailing list >>>>>>>>>>>>> us...@open-mpi.org >>>>>>>>>>>>> Subscription: >>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>>> Link to this post: >>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/04/26711.php >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> >>>>>>>>>>>> Kind Regards, >>>>>>>>>>>> >>>>>>>>>>>> M. >>>>>>>>>>>> >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> users mailing list >>>>>>>>>>>> us...@open-mpi.org >>>>>>>>>>>> Subscription: >>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>> Link to this post: >>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/04/26712.php >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> users mailing list >>>>>>>>>>> us...@open-mpi.org >>>>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>> Link to this post: >>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/04/26752.php >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> >>>>>>>>>> Kind Regards, >>>>>>>>>> >>>>>>>>>> M. >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> users mailing list >>>>>>>>>> us...@open-mpi.org >>>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>> Link to this post: >>>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/04/26754.php >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> users mailing list >>>>>>>>> us...@open-mpi.org >>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>> Link to this post: >>>>>>>>> http://www.open-mpi.org/community/lists/users/2015/04/26761.php >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> >>>>>>>> Kind Regards, >>>>>>>> >>>>>>>> M. >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> users mailing list >>>>>>>> us...@open-mpi.org >>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>> Link to this post: >>>>>>>> http://www.open-mpi.org/community/lists/users/2015/04/26762.php >>>>>>>> >>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> users mailing list >>>>>>> us...@open-mpi.org >>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>> Link to this post: >>>>>>> http://www.open-mpi.org/community/lists/users/2015/04/26766.php >>>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> >>>>>> Kind Regards, >>>>>> >>>>>> M. >>>>>> >>>>>> _______________________________________________ >>>>>> users mailing list >>>>>> us...@open-mpi.org >>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>> Link to this post: >>>>>> http://www.open-mpi.org/community/lists/users/2015/04/26768.php >>>>>> >>>>> >>>>> >>>>> _______________________________________________ >>>>> users mailing list >>>>> us...@open-mpi.org >>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>> Link to this post: >>>>> http://www.open-mpi.org/community/lists/users/2015/04/26777.php >>>>> >>>> >>>> >>>> >>>> -- >>>> >>>> Kind Regards, >>>> >>>> M. >>>> >>>> _______________________________________________ >>>> users mailing list >>>> us...@open-mpi.org >>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>> Link to this post: >>>> http://www.open-mpi.org/community/lists/users/2015/04/26779.php >>>> >>> >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>> Link to this post: >>> http://www.open-mpi.org/community/lists/users/2015/04/26781.php >>> >> >> >> >> -- >> >> Kind Regards, >> >> M. >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >> Link to this post: >> http://www.open-mpi.org/community/lists/users/2015/04/26782.php >> > > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > http://www.open-mpi.org/community/lists/users/2015/04/26799.php > -- Kind Regards, M.