Please send Jeff and I the opal/mca/pmix/pmix4x/pmix/config.log again - we’ll 
need to see why it isn’t building. The patch definitely is not in the v4.0 
branch, but it should have been in master.


> On Oct 5, 2018, at 2:04 AM, Siegmar Gross 
> <siegmar.gr...@informatik.hs-fulda.de> wrote:
> 
> Hi Ralph, hi Jeff,
> 
> 
> On 10/3/18 8:14 PM, Ralph H Castain wrote:
>> Jeff and I talked and believe the patch in 
>> https://github.com/open-mpi/ompi/pull/5836 
>> <https://github.com/open-mpi/ompi/pull/5836> should fix the problem.
> 
> 
> Today I've installed openmpi-master-201810050304-5f1c940 and
> openmpi-v4.0.x-201810050241-c079666. Unfortunately, I still get the
> same error for all seven versions that I was able to build.
> 
> loki hello_1 114 mpicc --showme
> gcc -I/usr/local/openmpi-master_64_gcc/include -fexceptions -pthread -std=c11 
> -m64 -Wl,-rpath -Wl,/usr/local/openmpi-master_64_gcc/lib64 
> -Wl,--enable-new-dtags -L/usr/local/openmpi-master_64_gcc/lib64 -lmpi
> 
> loki hello_1 115 ompi_info | grep "Open MPI repo revision"
>  Open MPI repo revision: v2.x-dev-6262-g5f1c940
> 
> loki hello_1 116 mpicc hello_1_mpi.c
> 
> loki hello_1 117 mpiexec -np 2 a.out
> [loki:25575] [[64603,0],0] ORTE_ERROR_LOG: Not found in file 
> ../../../../../openmpi-master-201810050304-5f1c940/orte/mca/ess/hnp/ess_hnp_module.c
>  at line 320
> --------------------------------------------------------------------------
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>  opal_pmix_base_select failed
>  --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --------------------------------------------------------------------------
> loki hello_1 118
> 
> 
> I don't know, if you have already applied your suggested patch or if the
> error message is still from a version without that patch. Do you need
> anything else?
> 
> 
> Best regards
> 
> Siegmar
> 
> 
>>> On Oct 2, 2018, at 2:50 PM, Jeff Squyres (jsquyres) via users 
>>> <users@lists.open-mpi.org> wrote:
>>> 
>>> (Ralph sent me Siegmar's pmix config.log, which Siegmar sent to him 
>>> off-list)
>>> 
>>> It looks like Siegmar passed --with-hwloc=internal.
>>> 
>>> Open MPI's configure understood this and did the appropriate things.
>>> PMIX's configure didn't.
>>> 
>>> I think we need to add an adjustment into the PMIx configure.m4 in OMPI...
>>> 
>>> 
>>>> On Oct 2, 2018, at 5:25 PM, Ralph H Castain <r...@open-mpi.org> wrote:
>>>> 
>>>> Hi Siegmar
>>>> 
>>>> I honestly have no idea - for some reason, the PMIx component isn’t seeing 
>>>> the internal hwloc code in your environment.
>>>> 
>>>> Jeff, Brice - any ideas?
>>>> 
>>>> 
>>>>> On Oct 2, 2018, at 1:18 PM, Siegmar Gross 
>>>>> <siegmar.gr...@informatik.hs-fulda.de> wrote:
>>>>> 
>>>>> Hi Ralph,
>>>>> 
>>>>> how can I confirm that HWLOC built? Some hwloc files are available
>>>>> in the built directory.
>>>>> 
>>>>> loki openmpi-master-201809290304-73075b8-Linux.x86_64.64_gcc 111 find . 
>>>>> -name '*hwloc*'
>>>>> ./opal/mca/btl/usnic/.deps/btl_usnic_hwloc.Plo
>>>>> ./opal/mca/hwloc
>>>>> ./opal/mca/hwloc/external/.deps/hwloc_external_component.Plo
>>>>> ./opal/mca/hwloc/base/hwloc_base_frame.lo
>>>>> ./opal/mca/hwloc/base/.deps/hwloc_base_dt.Plo
>>>>> ./opal/mca/hwloc/base/.deps/hwloc_base_maffinity.Plo
>>>>> ./opal/mca/hwloc/base/.deps/hwloc_base_frame.Plo
>>>>> ./opal/mca/hwloc/base/.deps/hwloc_base_util.Plo
>>>>> ./opal/mca/hwloc/base/hwloc_base_dt.lo
>>>>> ./opal/mca/hwloc/base/hwloc_base_util.lo
>>>>> ./opal/mca/hwloc/base/hwloc_base_maffinity.lo
>>>>> ./opal/mca/hwloc/base/.libs/hwloc_base_util.o
>>>>> ./opal/mca/hwloc/base/.libs/hwloc_base_dt.o
>>>>> ./opal/mca/hwloc/base/.libs/hwloc_base_maffinity.o
>>>>> ./opal/mca/hwloc/base/.libs/hwloc_base_frame.o
>>>>> ./opal/mca/hwloc/.libs/libmca_hwloc.la
>>>>> ./opal/mca/hwloc/.libs/libmca_hwloc.a
>>>>> ./opal/mca/hwloc/libmca_hwloc.la
>>>>> ./opal/mca/hwloc/hwloc201
>>>>> ./opal/mca/hwloc/hwloc201/.deps/hwloc201_component.Plo
>>>>> ./opal/mca/hwloc/hwloc201/hwloc201_component.lo
>>>>> ./opal/mca/hwloc/hwloc201/hwloc
>>>>> ./opal/mca/hwloc/hwloc201/hwloc/include/hwloc
>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc
>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/libhwloc_embedded.la
>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_pci_la-topology-pci.Plo
>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_gl_la-topology-gl.Plo
>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_cuda_la-topology-cuda.Plo
>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_xml_libxml_la-topology-xml-libxml.Plo
>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_opencl_la-topology-opencl.Plo
>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_nvml_la-topology-nvml.Plo
>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.libs/libhwloc_embedded.la
>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.libs/libhwloc_embedded.a
>>>>> ./opal/mca/hwloc/hwloc201/.libs/hwloc201_component.o
>>>>> ./opal/mca/hwloc/hwloc201/.libs/libmca_hwloc_hwloc201.la
>>>>> ./opal/mca/hwloc/hwloc201/.libs/libmca_hwloc_hwloc201.a
>>>>> ./opal/mca/hwloc/hwloc201/libmca_hwloc_hwloc201.la
>>>>> ./orte/mca/rtc/hwloc
>>>>> ./orte/mca/rtc/hwloc/rtc_hwloc.lo
>>>>> ./orte/mca/rtc/hwloc/.deps/rtc_hwloc.Plo
>>>>> ./orte/mca/rtc/hwloc/.deps/rtc_hwloc_component.Plo
>>>>> ./orte/mca/rtc/hwloc/mca_rtc_hwloc.la
>>>>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.so
>>>>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.la
>>>>> ./orte/mca/rtc/hwloc/.libs/rtc_hwloc.o
>>>>> ./orte/mca/rtc/hwloc/.libs/rtc_hwloc_component.o
>>>>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.soT
>>>>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.lai
>>>>> ./orte/mca/rtc/hwloc/rtc_hwloc_component.lo
>>>>> loki openmpi-master-201809290304-73075b8-Linux.x86_64.64_gcc 112
>>>>> 
>>>>> And some files are available in the install directory.
>>>>> 
>>>>> loki openmpi-master_64_gcc 116 find . -name '*hwloc*'
>>>>> ./share/openmpi/help-orte-rtc-hwloc.txt
>>>>> ./share/openmpi/help-opal-hwloc-base.txt
>>>>> ./lib64/openmpi/mca_rtc_hwloc.so
>>>>> ./lib64/openmpi/mca_rtc_hwloc.la
>>>>> loki openmpi-master_64_gcc 117
>>>>> 
>>>>> I don't see any unavailable libraries so that the only available
>>>>> hwloc library should work.
>>>>> 
>>>>> loki openmpi 126 ldd -v mca_rtc_hwloc.so
>>>>>      linux-vdso.so.1 (0x00007ffd2df5b000)
>>>>>      libopen-rte.so.0 => 
>>>>> /usr/local/openmpi-master_64_gcc/lib64/libopen-rte.so.0 
>>>>> (0x00007f082b7fb000)
>>>>>      libopen-pal.so.0 => 
>>>>> /usr/local/openmpi-master_64_gcc/lib64/libopen-pal.so.0 
>>>>> (0x00007f082b493000)
>>>>>      libdl.so.2 => /lib64/libdl.so.2 (0x00007f082b28f000)
>>>>>      libudev.so.1 => /usr/lib64/libudev.so.1 (0x00007f082b06e000)
>>>>>      libpciaccess.so.0 => /usr/lib64/libpciaccess.so.0 
>>>>> (0x00007f082ae64000)
>>>>>      librt.so.1 => /lib64/librt.so.1 (0x00007f082ac5c000)
>>>>>      libm.so.6 => /lib64/libm.so.6 (0x00007f082a95f000)
>>>>>      libutil.so.1 => /lib64/libutil.so.1 (0x00007f082a75c000)
>>>>>      libz.so.1 => /lib64/libz.so.1 (0x00007f082a546000)
>>>>>      libpthread.so.0 => /lib64/libpthread.so.0 (0x00007f082a329000)
>>>>>      libc.so.6 => /lib64/libc.so.6 (0x00007f0829f84000)
>>>>>      libgcc_s.so.1 => /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1 
>>>>> (0x00007f0829d6c000)
>>>>>      /lib64/ld-linux-x86-64.so.2 (0x00007f082bd24000)
>>>>>      libselinux.so.1 => /lib64/libselinux.so.1 (0x00007f0829b46000)
>>>>>      libcap.so.2 => /lib64/libcap.so.2 (0x00007f0829941000)
>>>>>      libresolv.so.2 => /lib64/libresolv.so.2 (0x00007f082972a000)
>>>>>      libpcre.so.1 => /usr/lib64/libpcre.so.1 (0x00007f08294bb000)
>>>>> 
>>>>>      Version information:
>>>>>      ./mca_rtc_hwloc.so:
>>>>>              libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0
>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>      /usr/local/openmpi-master_64_gcc/lib64/libopen-rte.so.0:
>>>>>              libz.so.1 (ZLIB_1.2.0) => /lib64/libz.so.1
>>>>>              libpthread.so.0 (GLIBC_2.3.2) => /lib64/libpthread.so.0
>>>>>              libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0
>>>>>              libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>      /usr/local/openmpi-master_64_gcc/lib64/libopen-pal.so.0:
>>>>>              librt.so.1 (GLIBC_2.2.5) => /lib64/librt.so.1
>>>>>              libgcc_s.so.1 (GCC_3.0) => 
>>>>> /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1
>>>>>              libgcc_s.so.1 (GCC_3.3.1) => 
>>>>> /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1
>>>>>              libdl.so.2 (GLIBC_2.2.5) => /lib64/libdl.so.2
>>>>>              libutil.so.1 (GLIBC_2.2.5) => /lib64/libutil.so.1
>>>>>              libudev.so.1 (LIBUDEV_183) => /usr/lib64/libudev.so.1
>>>>>              libm.so.6 (GLIBC_2.2.5) => /lib64/libm.so.6
>>>>>              libpthread.so.0 (GLIBC_2.3.4) => /lib64/libpthread.so.0
>>>>>              libpthread.so.0 (GLIBC_2.3.2) => /lib64/libpthread.so.0
>>>>>              libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0
>>>>>              libc.so.6 (GLIBC_2.6) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.7) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.3.2) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>      /lib64/libdl.so.2:
>>>>>              ld-linux-x86-64.so.2 (GLIBC_PRIVATE) => 
>>>>> /lib64/ld-linux-x86-64.so.2
>>>>>              libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>      /usr/lib64/libudev.so.1:
>>>>>              librt.so.1 (GLIBC_2.2.5) => /lib64/librt.so.1
>>>>>              ld-linux-x86-64.so.2 (GLIBC_2.3) => 
>>>>> /lib64/ld-linux-x86-64.so.2
>>>>>              libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0
>>>>>              libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.9) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.16) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.8) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>      /usr/lib64/libpciaccess.so.0:
>>>>>              libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.7) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6
>>>>>      /lib64/librt.so.1:
>>>>>              libpthread.so.0 (GLIBC_2.3.2) => /lib64/libpthread.so.0
>>>>>              libpthread.so.0 (GLIBC_PRIVATE) => /lib64/libpthread.so.0
>>>>>              libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0
>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.3.2) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6
>>>>>      /lib64/libm.so.6:
>>>>>              libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>      /lib64/libutil.so.1:
>>>>>              libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>      /lib64/libz.so.1:
>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6
>>>>>      /lib64/libpthread.so.0:
>>>>>              ld-linux-x86-64.so.2 (GLIBC_2.2.5) => 
>>>>> /lib64/ld-linux-x86-64.so.2
>>>>>              ld-linux-x86-64.so.2 (GLIBC_PRIVATE) => 
>>>>> /lib64/ld-linux-x86-64.so.2
>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.3.2) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6
>>>>>      /lib64/libc.so.6:
>>>>>              ld-linux-x86-64.so.2 (GLIBC_2.3) => 
>>>>> /lib64/ld-linux-x86-64.so.2
>>>>>              ld-linux-x86-64.so.2 (GLIBC_PRIVATE) => 
>>>>> /lib64/ld-linux-x86-64.so.2
>>>>>      /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1:
>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>      /lib64/libselinux.so.1:
>>>>>              libdl.so.2 (GLIBC_2.2.5) => /lib64/libdl.so.2
>>>>>              ld-linux-x86-64.so.2 (GLIBC_2.3) => 
>>>>> /lib64/ld-linux-x86-64.so.2
>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.8) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.7) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6
>>>>>      /lib64/libcap.so.2:
>>>>>              libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.8) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>      /lib64/libresolv.so.2:
>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6
>>>>>      /usr/lib64/libpcre.so.1:
>>>>>              libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0
>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>              libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6
>>>>> loki openmpi 127
>>>>> 
>>>>> Hopefully that helps to find the problem. I will answer your emails
>>>>> tommorrow if you need anything else.
>>>>> 
>>>>> 
>>>>> Best regards
>>>>> 
>>>>> Siegmar
>>>>> 
>>>>> 
>>>>> Am 02.10.2018 um 19:48 schrieb Ralph H Castain:
>>>>>> So the problem is here when configuring the internal PMIx code:
>>>>>> configure:3383: === HWLOC
>>>>>> configure:36189: checking for hwloc in
>>>>>> configure:36201: result: Could not find internal/lib or internal/lib64
>>>>>> configure:36203: error: Can not continue
>>>>>> Can you confirm that HWLOC built? I believe we require it, but perhaps 
>>>>>> something is different about this environment.
>>>>>>> On Oct 2, 2018, at 6:36 AM, Ralph H Castain <r...@open-mpi.org> wrote:
>>>>>>> 
>>>>>>> Looks like PMIx failed to build - can you send the config.log?
>>>>>>> 
>>>>>>>> On Oct 2, 2018, at 12:00 AM, Siegmar Gross 
>>>>>>>> <siegmar.gr...@informatik.hs-fulda.de> wrote:
>>>>>>>> 
>>>>>>>> Hi,
>>>>>>>> 
>>>>>>>> yesterday I've installed openmpi-v4.0.x-201809290241-a7e275c and
>>>>>>>> openmpi-master-201805080348-b39bbfb on my "SUSE Linux Enterprise Server
>>>>>>>> 12.3 (x86_64)" with Sun C 5.15, gcc 6.4.0, Intel icc 18.0.3, and 
>>>>>>>> Portland
>>>>>>>> Group pgcc 18.4-0. Unfortunately, I get the following error for all 
>>>>>>>> seven
>>>>>>>> installed versions (Sun C couldn't built master as I mentioned in 
>>>>>>>> another
>>>>>>>> email).
>>>>>>>> 
>>>>>>>> 
>>>>>>>> loki hello_1 118 mpiexec -np 4 --host loki:2,nfs2:2 hello_1_mpi
>>>>>>>> [loki:11423] [[45859,0],0] ORTE_ERROR_LOG: Not found in file 
>>>>>>>> ../../../../../openmpi-v4.0.x-201809290241-a7e275c/orte/mca/ess/hnp/ess_hnp_module.c
>>>>>>>>  at line 321
>>>>>>>> --------------------------------------------------------------------------
>>>>>>>> It looks like orte_init failed for some reason; your parallel process 
>>>>>>>> is
>>>>>>>> likely to abort.  There are many reasons that a parallel process can
>>>>>>>> fail during orte_init; some of which are due to configuration or
>>>>>>>> environment problems.  This failure appears to be an internal failure;
>>>>>>>> here's some additional information (which may only be relevant to an
>>>>>>>> Open MPI developer):
>>>>>>>> 
>>>>>>>> opal_pmix_base_select failed
>>>>>>>> --> Returned value Not found (-13) instead of ORTE_SUCCESS
>>>>>>>> --------------------------------------------------------------------------
>>>>>>>> loki hello_1 119
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> I would be grateful, if somebody can fix the problem. Do you need 
>>>>>>>> anything
>>>>>>>> else? Thank you very much for any help in advance.
>>>>>>>> 
>>>>>>>> 
>>>>>>>> Kind regards
>>>>>>>> 
>>>>>>>> Siegmar
>>>>>>>> _______________________________________________
>>>>>>>> users mailing list
>>>>>>>> users@lists.open-mpi.org
>>>>>>>> https://lists.open-mpi.org/mailman/listinfo/users
>>>>>>> 
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> users@lists.open-mpi.org
>>>>>>> https://lists.open-mpi.org/mailman/listinfo/users
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> users@lists.open-mpi.org
>>>>>> https://lists.open-mpi.org/mailman/listinfo/users
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> users@lists.open-mpi.org
>>>> https://lists.open-mpi.org/mailman/listinfo/users
>>> 
>>> 
>>> -- 
>>> Jeff Squyres
>>> jsquy...@cisco.com
>>> 
>>> _______________________________________________
>>> users mailing list
>>> users@lists.open-mpi.org
>>> https://lists.open-mpi.org/mailman/listinfo/users
>> _______________________________________________
>> users mailing list
>> users@lists.open-mpi.org
>> https://lists.open-mpi.org/mailman/listinfo/users
> _______________________________________________
> users mailing list
> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
> https://lists.open-mpi.org/mailman/listinfo/users 
> <https://lists.open-mpi.org/mailman/listinfo/users>
_______________________________________________
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

Reply via email to