Please send Jeff and I the opal/mca/pmix/pmix4x/pmix/config.log again - we’ll need to see why it isn’t building. The patch definitely is not in the v4.0 branch, but it should have been in master.
> On Oct 5, 2018, at 2:04 AM, Siegmar Gross > <siegmar.gr...@informatik.hs-fulda.de> wrote: > > Hi Ralph, hi Jeff, > > > On 10/3/18 8:14 PM, Ralph H Castain wrote: >> Jeff and I talked and believe the patch in >> https://github.com/open-mpi/ompi/pull/5836 >> <https://github.com/open-mpi/ompi/pull/5836> should fix the problem. > > > Today I've installed openmpi-master-201810050304-5f1c940 and > openmpi-v4.0.x-201810050241-c079666. Unfortunately, I still get the > same error for all seven versions that I was able to build. > > loki hello_1 114 mpicc --showme > gcc -I/usr/local/openmpi-master_64_gcc/include -fexceptions -pthread -std=c11 > -m64 -Wl,-rpath -Wl,/usr/local/openmpi-master_64_gcc/lib64 > -Wl,--enable-new-dtags -L/usr/local/openmpi-master_64_gcc/lib64 -lmpi > > loki hello_1 115 ompi_info | grep "Open MPI repo revision" > Open MPI repo revision: v2.x-dev-6262-g5f1c940 > > loki hello_1 116 mpicc hello_1_mpi.c > > loki hello_1 117 mpiexec -np 2 a.out > [loki:25575] [[64603,0],0] ORTE_ERROR_LOG: Not found in file > ../../../../../openmpi-master-201810050304-5f1c940/orte/mca/ess/hnp/ess_hnp_module.c > at line 320 > -------------------------------------------------------------------------- > It looks like orte_init failed for some reason; your parallel process is > likely to abort. There are many reasons that a parallel process can > fail during orte_init; some of which are due to configuration or > environment problems. This failure appears to be an internal failure; > here's some additional information (which may only be relevant to an > Open MPI developer): > > opal_pmix_base_select failed > --> Returned value Not found (-13) instead of ORTE_SUCCESS > -------------------------------------------------------------------------- > loki hello_1 118 > > > I don't know, if you have already applied your suggested patch or if the > error message is still from a version without that patch. Do you need > anything else? > > > Best regards > > Siegmar > > >>> On Oct 2, 2018, at 2:50 PM, Jeff Squyres (jsquyres) via users >>> <users@lists.open-mpi.org> wrote: >>> >>> (Ralph sent me Siegmar's pmix config.log, which Siegmar sent to him >>> off-list) >>> >>> It looks like Siegmar passed --with-hwloc=internal. >>> >>> Open MPI's configure understood this and did the appropriate things. >>> PMIX's configure didn't. >>> >>> I think we need to add an adjustment into the PMIx configure.m4 in OMPI... >>> >>> >>>> On Oct 2, 2018, at 5:25 PM, Ralph H Castain <r...@open-mpi.org> wrote: >>>> >>>> Hi Siegmar >>>> >>>> I honestly have no idea - for some reason, the PMIx component isn’t seeing >>>> the internal hwloc code in your environment. >>>> >>>> Jeff, Brice - any ideas? >>>> >>>> >>>>> On Oct 2, 2018, at 1:18 PM, Siegmar Gross >>>>> <siegmar.gr...@informatik.hs-fulda.de> wrote: >>>>> >>>>> Hi Ralph, >>>>> >>>>> how can I confirm that HWLOC built? Some hwloc files are available >>>>> in the built directory. >>>>> >>>>> loki openmpi-master-201809290304-73075b8-Linux.x86_64.64_gcc 111 find . >>>>> -name '*hwloc*' >>>>> ./opal/mca/btl/usnic/.deps/btl_usnic_hwloc.Plo >>>>> ./opal/mca/hwloc >>>>> ./opal/mca/hwloc/external/.deps/hwloc_external_component.Plo >>>>> ./opal/mca/hwloc/base/hwloc_base_frame.lo >>>>> ./opal/mca/hwloc/base/.deps/hwloc_base_dt.Plo >>>>> ./opal/mca/hwloc/base/.deps/hwloc_base_maffinity.Plo >>>>> ./opal/mca/hwloc/base/.deps/hwloc_base_frame.Plo >>>>> ./opal/mca/hwloc/base/.deps/hwloc_base_util.Plo >>>>> ./opal/mca/hwloc/base/hwloc_base_dt.lo >>>>> ./opal/mca/hwloc/base/hwloc_base_util.lo >>>>> ./opal/mca/hwloc/base/hwloc_base_maffinity.lo >>>>> ./opal/mca/hwloc/base/.libs/hwloc_base_util.o >>>>> ./opal/mca/hwloc/base/.libs/hwloc_base_dt.o >>>>> ./opal/mca/hwloc/base/.libs/hwloc_base_maffinity.o >>>>> ./opal/mca/hwloc/base/.libs/hwloc_base_frame.o >>>>> ./opal/mca/hwloc/.libs/libmca_hwloc.la >>>>> ./opal/mca/hwloc/.libs/libmca_hwloc.a >>>>> ./opal/mca/hwloc/libmca_hwloc.la >>>>> ./opal/mca/hwloc/hwloc201 >>>>> ./opal/mca/hwloc/hwloc201/.deps/hwloc201_component.Plo >>>>> ./opal/mca/hwloc/hwloc201/hwloc201_component.lo >>>>> ./opal/mca/hwloc/hwloc201/hwloc >>>>> ./opal/mca/hwloc/hwloc201/hwloc/include/hwloc >>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc >>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/libhwloc_embedded.la >>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_pci_la-topology-pci.Plo >>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_gl_la-topology-gl.Plo >>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_cuda_la-topology-cuda.Plo >>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_xml_libxml_la-topology-xml-libxml.Plo >>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_opencl_la-topology-opencl.Plo >>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_nvml_la-topology-nvml.Plo >>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.libs/libhwloc_embedded.la >>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.libs/libhwloc_embedded.a >>>>> ./opal/mca/hwloc/hwloc201/.libs/hwloc201_component.o >>>>> ./opal/mca/hwloc/hwloc201/.libs/libmca_hwloc_hwloc201.la >>>>> ./opal/mca/hwloc/hwloc201/.libs/libmca_hwloc_hwloc201.a >>>>> ./opal/mca/hwloc/hwloc201/libmca_hwloc_hwloc201.la >>>>> ./orte/mca/rtc/hwloc >>>>> ./orte/mca/rtc/hwloc/rtc_hwloc.lo >>>>> ./orte/mca/rtc/hwloc/.deps/rtc_hwloc.Plo >>>>> ./orte/mca/rtc/hwloc/.deps/rtc_hwloc_component.Plo >>>>> ./orte/mca/rtc/hwloc/mca_rtc_hwloc.la >>>>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.so >>>>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.la >>>>> ./orte/mca/rtc/hwloc/.libs/rtc_hwloc.o >>>>> ./orte/mca/rtc/hwloc/.libs/rtc_hwloc_component.o >>>>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.soT >>>>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.lai >>>>> ./orte/mca/rtc/hwloc/rtc_hwloc_component.lo >>>>> loki openmpi-master-201809290304-73075b8-Linux.x86_64.64_gcc 112 >>>>> >>>>> And some files are available in the install directory. >>>>> >>>>> loki openmpi-master_64_gcc 116 find . -name '*hwloc*' >>>>> ./share/openmpi/help-orte-rtc-hwloc.txt >>>>> ./share/openmpi/help-opal-hwloc-base.txt >>>>> ./lib64/openmpi/mca_rtc_hwloc.so >>>>> ./lib64/openmpi/mca_rtc_hwloc.la >>>>> loki openmpi-master_64_gcc 117 >>>>> >>>>> I don't see any unavailable libraries so that the only available >>>>> hwloc library should work. >>>>> >>>>> loki openmpi 126 ldd -v mca_rtc_hwloc.so >>>>> linux-vdso.so.1 (0x00007ffd2df5b000) >>>>> libopen-rte.so.0 => >>>>> /usr/local/openmpi-master_64_gcc/lib64/libopen-rte.so.0 >>>>> (0x00007f082b7fb000) >>>>> libopen-pal.so.0 => >>>>> /usr/local/openmpi-master_64_gcc/lib64/libopen-pal.so.0 >>>>> (0x00007f082b493000) >>>>> libdl.so.2 => /lib64/libdl.so.2 (0x00007f082b28f000) >>>>> libudev.so.1 => /usr/lib64/libudev.so.1 (0x00007f082b06e000) >>>>> libpciaccess.so.0 => /usr/lib64/libpciaccess.so.0 >>>>> (0x00007f082ae64000) >>>>> librt.so.1 => /lib64/librt.so.1 (0x00007f082ac5c000) >>>>> libm.so.6 => /lib64/libm.so.6 (0x00007f082a95f000) >>>>> libutil.so.1 => /lib64/libutil.so.1 (0x00007f082a75c000) >>>>> libz.so.1 => /lib64/libz.so.1 (0x00007f082a546000) >>>>> libpthread.so.0 => /lib64/libpthread.so.0 (0x00007f082a329000) >>>>> libc.so.6 => /lib64/libc.so.6 (0x00007f0829f84000) >>>>> libgcc_s.so.1 => /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1 >>>>> (0x00007f0829d6c000) >>>>> /lib64/ld-linux-x86-64.so.2 (0x00007f082bd24000) >>>>> libselinux.so.1 => /lib64/libselinux.so.1 (0x00007f0829b46000) >>>>> libcap.so.2 => /lib64/libcap.so.2 (0x00007f0829941000) >>>>> libresolv.so.2 => /lib64/libresolv.so.2 (0x00007f082972a000) >>>>> libpcre.so.1 => /usr/lib64/libpcre.so.1 (0x00007f08294bb000) >>>>> >>>>> Version information: >>>>> ./mca_rtc_hwloc.so: >>>>> libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0 >>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>> /usr/local/openmpi-master_64_gcc/lib64/libopen-rte.so.0: >>>>> libz.so.1 (ZLIB_1.2.0) => /lib64/libz.so.1 >>>>> libpthread.so.0 (GLIBC_2.3.2) => /lib64/libpthread.so.0 >>>>> libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0 >>>>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>> /usr/local/openmpi-master_64_gcc/lib64/libopen-pal.so.0: >>>>> librt.so.1 (GLIBC_2.2.5) => /lib64/librt.so.1 >>>>> libgcc_s.so.1 (GCC_3.0) => >>>>> /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1 >>>>> libgcc_s.so.1 (GCC_3.3.1) => >>>>> /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1 >>>>> libdl.so.2 (GLIBC_2.2.5) => /lib64/libdl.so.2 >>>>> libutil.so.1 (GLIBC_2.2.5) => /lib64/libutil.so.1 >>>>> libudev.so.1 (LIBUDEV_183) => /usr/lib64/libudev.so.1 >>>>> libm.so.6 (GLIBC_2.2.5) => /lib64/libm.so.6 >>>>> libpthread.so.0 (GLIBC_2.3.4) => /lib64/libpthread.so.0 >>>>> libpthread.so.0 (GLIBC_2.3.2) => /lib64/libpthread.so.0 >>>>> libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0 >>>>> libc.so.6 (GLIBC_2.6) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.7) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.3.2) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>> /lib64/libdl.so.2: >>>>> ld-linux-x86-64.so.2 (GLIBC_PRIVATE) => >>>>> /lib64/ld-linux-x86-64.so.2 >>>>> libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>> /usr/lib64/libudev.so.1: >>>>> librt.so.1 (GLIBC_2.2.5) => /lib64/librt.so.1 >>>>> ld-linux-x86-64.so.2 (GLIBC_2.3) => >>>>> /lib64/ld-linux-x86-64.so.2 >>>>> libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0 >>>>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.9) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.16) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.8) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>> /usr/lib64/libpciaccess.so.0: >>>>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.7) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>>>> /lib64/librt.so.1: >>>>> libpthread.so.0 (GLIBC_2.3.2) => /lib64/libpthread.so.0 >>>>> libpthread.so.0 (GLIBC_PRIVATE) => /lib64/libpthread.so.0 >>>>> libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0 >>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.3.2) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6 >>>>> /lib64/libm.so.6: >>>>> libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>> /lib64/libutil.so.1: >>>>> libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>> /lib64/libz.so.1: >>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>>>> /lib64/libpthread.so.0: >>>>> ld-linux-x86-64.so.2 (GLIBC_2.2.5) => >>>>> /lib64/ld-linux-x86-64.so.2 >>>>> ld-linux-x86-64.so.2 (GLIBC_PRIVATE) => >>>>> /lib64/ld-linux-x86-64.so.2 >>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.3.2) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6 >>>>> /lib64/libc.so.6: >>>>> ld-linux-x86-64.so.2 (GLIBC_2.3) => >>>>> /lib64/ld-linux-x86-64.so.2 >>>>> ld-linux-x86-64.so.2 (GLIBC_PRIVATE) => >>>>> /lib64/ld-linux-x86-64.so.2 >>>>> /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1: >>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>> /lib64/libselinux.so.1: >>>>> libdl.so.2 (GLIBC_2.2.5) => /lib64/libdl.so.2 >>>>> ld-linux-x86-64.so.2 (GLIBC_2.3) => >>>>> /lib64/ld-linux-x86-64.so.2 >>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.8) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.7) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>>>> /lib64/libcap.so.2: >>>>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.8) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>> /lib64/libresolv.so.2: >>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>>>> /usr/lib64/libpcre.so.1: >>>>> libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0 >>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>>>> loki openmpi 127 >>>>> >>>>> Hopefully that helps to find the problem. I will answer your emails >>>>> tommorrow if you need anything else. >>>>> >>>>> >>>>> Best regards >>>>> >>>>> Siegmar >>>>> >>>>> >>>>> Am 02.10.2018 um 19:48 schrieb Ralph H Castain: >>>>>> So the problem is here when configuring the internal PMIx code: >>>>>> configure:3383: === HWLOC >>>>>> configure:36189: checking for hwloc in >>>>>> configure:36201: result: Could not find internal/lib or internal/lib64 >>>>>> configure:36203: error: Can not continue >>>>>> Can you confirm that HWLOC built? I believe we require it, but perhaps >>>>>> something is different about this environment. >>>>>>> On Oct 2, 2018, at 6:36 AM, Ralph H Castain <r...@open-mpi.org> wrote: >>>>>>> >>>>>>> Looks like PMIx failed to build - can you send the config.log? >>>>>>> >>>>>>>> On Oct 2, 2018, at 12:00 AM, Siegmar Gross >>>>>>>> <siegmar.gr...@informatik.hs-fulda.de> wrote: >>>>>>>> >>>>>>>> Hi, >>>>>>>> >>>>>>>> yesterday I've installed openmpi-v4.0.x-201809290241-a7e275c and >>>>>>>> openmpi-master-201805080348-b39bbfb on my "SUSE Linux Enterprise Server >>>>>>>> 12.3 (x86_64)" with Sun C 5.15, gcc 6.4.0, Intel icc 18.0.3, and >>>>>>>> Portland >>>>>>>> Group pgcc 18.4-0. Unfortunately, I get the following error for all >>>>>>>> seven >>>>>>>> installed versions (Sun C couldn't built master as I mentioned in >>>>>>>> another >>>>>>>> email). >>>>>>>> >>>>>>>> >>>>>>>> loki hello_1 118 mpiexec -np 4 --host loki:2,nfs2:2 hello_1_mpi >>>>>>>> [loki:11423] [[45859,0],0] ORTE_ERROR_LOG: Not found in file >>>>>>>> ../../../../../openmpi-v4.0.x-201809290241-a7e275c/orte/mca/ess/hnp/ess_hnp_module.c >>>>>>>> at line 321 >>>>>>>> -------------------------------------------------------------------------- >>>>>>>> It looks like orte_init failed for some reason; your parallel process >>>>>>>> is >>>>>>>> likely to abort. There are many reasons that a parallel process can >>>>>>>> fail during orte_init; some of which are due to configuration or >>>>>>>> environment problems. This failure appears to be an internal failure; >>>>>>>> here's some additional information (which may only be relevant to an >>>>>>>> Open MPI developer): >>>>>>>> >>>>>>>> opal_pmix_base_select failed >>>>>>>> --> Returned value Not found (-13) instead of ORTE_SUCCESS >>>>>>>> -------------------------------------------------------------------------- >>>>>>>> loki hello_1 119 >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> I would be grateful, if somebody can fix the problem. Do you need >>>>>>>> anything >>>>>>>> else? Thank you very much for any help in advance. >>>>>>>> >>>>>>>> >>>>>>>> Kind regards >>>>>>>> >>>>>>>> Siegmar >>>>>>>> _______________________________________________ >>>>>>>> users mailing list >>>>>>>> users@lists.open-mpi.org >>>>>>>> https://lists.open-mpi.org/mailman/listinfo/users >>>>>>> >>>>>>> _______________________________________________ >>>>>>> users mailing list >>>>>>> users@lists.open-mpi.org >>>>>>> https://lists.open-mpi.org/mailman/listinfo/users >>>>>> _______________________________________________ >>>>>> users mailing list >>>>>> users@lists.open-mpi.org >>>>>> https://lists.open-mpi.org/mailman/listinfo/users >>>> >>>> _______________________________________________ >>>> users mailing list >>>> users@lists.open-mpi.org >>>> https://lists.open-mpi.org/mailman/listinfo/users >>> >>> >>> -- >>> Jeff Squyres >>> jsquy...@cisco.com >>> >>> _______________________________________________ >>> users mailing list >>> users@lists.open-mpi.org >>> https://lists.open-mpi.org/mailman/listinfo/users >> _______________________________________________ >> users mailing list >> users@lists.open-mpi.org >> https://lists.open-mpi.org/mailman/listinfo/users > _______________________________________________ > users mailing list > users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> > https://lists.open-mpi.org/mailman/listinfo/users > <https://lists.open-mpi.org/mailman/listinfo/users>
_______________________________________________ users mailing list users@lists.open-mpi.org https://lists.open-mpi.org/mailman/listinfo/users