Oops!  We had a typo in yesterday's fix -- fixed:

    https://github.com/open-mpi/ompi/pull/5847

Ralph also put double extra super protection to make triple sure that this 
error can't happen again in:

    https://github.com/open-mpi/ompi/pull/5846

Both of these should be in tonight's nightly snapshot.

Thank you!


> On Oct 5, 2018, at 5:45 AM, Ralph H Castain <r...@open-mpi.org> wrote:
> 
> Please send Jeff and I the opal/mca/pmix/pmix4x/pmix/config.log again - we’ll 
> need to see why it isn’t building. The patch definitely is not in the v4.0 
> branch, but it should have been in master.
> 
> 
>> On Oct 5, 2018, at 2:04 AM, Siegmar Gross 
>> <siegmar.gr...@informatik.hs-fulda.de> wrote:
>> 
>> Hi Ralph, hi Jeff,
>> 
>> 
>> On 10/3/18 8:14 PM, Ralph H Castain wrote:
>>> Jeff and I talked and believe the patch in 
>>> https://github.com/open-mpi/ompi/pull/5836 should fix the problem.
>> 
>> 
>> Today I've installed openmpi-master-201810050304-5f1c940 and
>> openmpi-v4.0.x-201810050241-c079666. Unfortunately, I still get the
>> same error for all seven versions that I was able to build.
>> 
>> loki hello_1 114 mpicc --showme
>> gcc -I/usr/local/openmpi-master_64_gcc/include -fexceptions -pthread 
>> -std=c11 -m64 -Wl,-rpath -Wl,/usr/local/openmpi-master_64_gcc/lib64 
>> -Wl,--enable-new-dtags -L/usr/local/openmpi-master_64_gcc/lib64 -lmpi
>> 
>> loki hello_1 115 ompi_info | grep "Open MPI repo revision"
>>  Open MPI repo revision: v2.x-dev-6262-g5f1c940
>> 
>> loki hello_1 116 mpicc hello_1_mpi.c
>> 
>> loki hello_1 117 mpiexec -np 2 a.out
>> [loki:25575] [[64603,0],0] ORTE_ERROR_LOG: Not found in file 
>> ../../../../../openmpi-master-201810050304-5f1c940/orte/mca/ess/hnp/ess_hnp_module.c
>>  at line 320
>> --------------------------------------------------------------------------
>> It looks like orte_init failed for some reason; your parallel process is
>> likely to abort.  There are many reasons that a parallel process can
>> fail during orte_init; some of which are due to configuration or
>> environment problems.  This failure appears to be an internal failure;
>> here's some additional information (which may only be relevant to an
>> Open MPI developer):
>> 
>>  opal_pmix_base_select failed
>>  --> Returned value Not found (-13) instead of ORTE_SUCCESS
>> --------------------------------------------------------------------------
>> loki hello_1 118
>> 
>> 
>> I don't know, if you have already applied your suggested patch or if the
>> error message is still from a version without that patch. Do you need
>> anything else?
>> 
>> 
>> Best regards
>> 
>> Siegmar
>> 
>> 
>>>> On Oct 2, 2018, at 2:50 PM, Jeff Squyres (jsquyres) via users 
>>>> <users@lists.open-mpi.org> wrote:
>>>> 
>>>> (Ralph sent me Siegmar's pmix config.log, which Siegmar sent to him 
>>>> off-list)
>>>> 
>>>> It looks like Siegmar passed --with-hwloc=internal.
>>>> 
>>>> Open MPI's configure understood this and did the appropriate things.
>>>> PMIX's configure didn't.
>>>> 
>>>> I think we need to add an adjustment into the PMIx configure.m4 in OMPI...
>>>> 
>>>> 
>>>>> On Oct 2, 2018, at 5:25 PM, Ralph H Castain <r...@open-mpi.org> wrote:
>>>>> 
>>>>> Hi Siegmar
>>>>> 
>>>>> I honestly have no idea - for some reason, the PMIx component isn’t 
>>>>> seeing the internal hwloc code in your environment.
>>>>> 
>>>>> Jeff, Brice - any ideas?
>>>>> 
>>>>> 
>>>>>> On Oct 2, 2018, at 1:18 PM, Siegmar Gross 
>>>>>> <siegmar.gr...@informatik.hs-fulda.de> wrote:
>>>>>> 
>>>>>> Hi Ralph,
>>>>>> 
>>>>>> how can I confirm that HWLOC built? Some hwloc files are available
>>>>>> in the built directory.
>>>>>> 
>>>>>> loki openmpi-master-201809290304-73075b8-Linux.x86_64.64_gcc 111 find . 
>>>>>> -name '*hwloc*'
>>>>>> ./opal/mca/btl/usnic/.deps/btl_usnic_hwloc.Plo
>>>>>> ./opal/mca/hwloc
>>>>>> ./opal/mca/hwloc/external/.deps/hwloc_external_component.Plo
>>>>>> ./opal/mca/hwloc/base/hwloc_base_frame.lo
>>>>>> ./opal/mca/hwloc/base/.deps/hwloc_base_dt.Plo
>>>>>> ./opal/mca/hwloc/base/.deps/hwloc_base_maffinity.Plo
>>>>>> ./opal/mca/hwloc/base/.deps/hwloc_base_frame.Plo
>>>>>> ./opal/mca/hwloc/base/.deps/hwloc_base_util.Plo
>>>>>> ./opal/mca/hwloc/base/hwloc_base_dt.lo
>>>>>> ./opal/mca/hwloc/base/hwloc_base_util.lo
>>>>>> ./opal/mca/hwloc/base/hwloc_base_maffinity.lo
>>>>>> ./opal/mca/hwloc/base/.libs/hwloc_base_util.o
>>>>>> ./opal/mca/hwloc/base/.libs/hwloc_base_dt.o
>>>>>> ./opal/mca/hwloc/base/.libs/hwloc_base_maffinity.o
>>>>>> ./opal/mca/hwloc/base/.libs/hwloc_base_frame.o
>>>>>> ./opal/mca/hwloc/.libs/libmca_hwloc.la
>>>>>> ./opal/mca/hwloc/.libs/libmca_hwloc.a
>>>>>> ./opal/mca/hwloc/libmca_hwloc.la
>>>>>> ./opal/mca/hwloc/hwloc201
>>>>>> ./opal/mca/hwloc/hwloc201/.deps/hwloc201_component.Plo
>>>>>> ./opal/mca/hwloc/hwloc201/hwloc201_component.lo
>>>>>> ./opal/mca/hwloc/hwloc201/hwloc
>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/include/hwloc
>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc
>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/libhwloc_embedded.la
>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_pci_la-topology-pci.Plo
>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_gl_la-topology-gl.Plo
>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_cuda_la-topology-cuda.Plo
>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_xml_libxml_la-topology-xml-libxml.Plo
>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_opencl_la-topology-opencl.Plo
>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_nvml_la-topology-nvml.Plo
>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.libs/libhwloc_embedded.la
>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.libs/libhwloc_embedded.a
>>>>>> ./opal/mca/hwloc/hwloc201/.libs/hwloc201_component.o
>>>>>> ./opal/mca/hwloc/hwloc201/.libs/libmca_hwloc_hwloc201.la
>>>>>> ./opal/mca/hwloc/hwloc201/.libs/libmca_hwloc_hwloc201.a
>>>>>> ./opal/mca/hwloc/hwloc201/libmca_hwloc_hwloc201.la
>>>>>> ./orte/mca/rtc/hwloc
>>>>>> ./orte/mca/rtc/hwloc/rtc_hwloc.lo
>>>>>> ./orte/mca/rtc/hwloc/.deps/rtc_hwloc.Plo
>>>>>> ./orte/mca/rtc/hwloc/.deps/rtc_hwloc_component.Plo
>>>>>> ./orte/mca/rtc/hwloc/mca_rtc_hwloc.la
>>>>>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.so
>>>>>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.la
>>>>>> ./orte/mca/rtc/hwloc/.libs/rtc_hwloc.o
>>>>>> ./orte/mca/rtc/hwloc/.libs/rtc_hwloc_component.o
>>>>>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.soT
>>>>>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.lai
>>>>>> ./orte/mca/rtc/hwloc/rtc_hwloc_component.lo
>>>>>> loki openmpi-master-201809290304-73075b8-Linux.x86_64.64_gcc 112
>>>>>> 
>>>>>> And some files are available in the install directory.
>>>>>> 
>>>>>> loki openmpi-master_64_gcc 116 find . -name '*hwloc*'
>>>>>> ./share/openmpi/help-orte-rtc-hwloc.txt
>>>>>> ./share/openmpi/help-opal-hwloc-base.txt
>>>>>> ./lib64/openmpi/mca_rtc_hwloc.so
>>>>>> ./lib64/openmpi/mca_rtc_hwloc.la
>>>>>> loki openmpi-master_64_gcc 117
>>>>>> 
>>>>>> I don't see any unavailable libraries so that the only available
>>>>>> hwloc library should work.
>>>>>> 
>>>>>> loki openmpi 126 ldd -v mca_rtc_hwloc.so
>>>>>>      linux-vdso.so.1 (0x00007ffd2df5b000)
>>>>>>      libopen-rte.so.0 => 
>>>>>> /usr/local/openmpi-master_64_gcc/lib64/libopen-rte.so.0 
>>>>>> (0x00007f082b7fb000)
>>>>>>      libopen-pal.so.0 => 
>>>>>> /usr/local/openmpi-master_64_gcc/lib64/libopen-pal.so.0 
>>>>>> (0x00007f082b493000)
>>>>>>      libdl.so.2 => /lib64/libdl.so.2 (0x00007f082b28f000)
>>>>>>      libudev.so.1 => /usr/lib64/libudev.so.1 (0x00007f082b06e000)
>>>>>>      libpciaccess.so.0 => /usr/lib64/libpciaccess.so.0 
>>>>>> (0x00007f082ae64000)
>>>>>>      librt.so.1 => /lib64/librt.so.1 (0x00007f082ac5c000)
>>>>>>      libm.so.6 => /lib64/libm.so.6 (0x00007f082a95f000)
>>>>>>      libutil.so.1 => /lib64/libutil.so.1 (0x00007f082a75c000)
>>>>>>      libz.so.1 => /lib64/libz.so.1 (0x00007f082a546000)
>>>>>>      libpthread.so.0 => /lib64/libpthread.so.0 (0x00007f082a329000)
>>>>>>      libc.so.6 => /lib64/libc.so.6 (0x00007f0829f84000)
>>>>>>      libgcc_s.so.1 => /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1 
>>>>>> (0x00007f0829d6c000)
>>>>>>      /lib64/ld-linux-x86-64.so.2 (0x00007f082bd24000)
>>>>>>      libselinux.so.1 => /lib64/libselinux.so.1 (0x00007f0829b46000)
>>>>>>      libcap.so.2 => /lib64/libcap.so.2 (0x00007f0829941000)
>>>>>>      libresolv.so.2 => /lib64/libresolv.so.2 (0x00007f082972a000)
>>>>>>      libpcre.so.1 => /usr/lib64/libpcre.so.1 (0x00007f08294bb000)
>>>>>> 
>>>>>>      Version information:
>>>>>>      ./mca_rtc_hwloc.so:
>>>>>>              libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0
>>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>>      /usr/local/openmpi-master_64_gcc/lib64/libopen-rte.so.0:
>>>>>>              libz.so.1 (ZLIB_1.2.0) => /lib64/libz.so.1
>>>>>>              libpthread.so.0 (GLIBC_2.3.2) => /lib64/libpthread.so.0
>>>>>>              libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0
>>>>>>              libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>>      /usr/local/openmpi-master_64_gcc/lib64/libopen-pal.so.0:
>>>>>>              librt.so.1 (GLIBC_2.2.5) => /lib64/librt.so.1
>>>>>>              libgcc_s.so.1 (GCC_3.0) => 
>>>>>> /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1
>>>>>>              libgcc_s.so.1 (GCC_3.3.1) => 
>>>>>> /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1
>>>>>>              libdl.so.2 (GLIBC_2.2.5) => /lib64/libdl.so.2
>>>>>>              libutil.so.1 (GLIBC_2.2.5) => /lib64/libutil.so.1
>>>>>>              libudev.so.1 (LIBUDEV_183) => /usr/lib64/libudev.so.1
>>>>>>              libm.so.6 (GLIBC_2.2.5) => /lib64/libm.so.6
>>>>>>              libpthread.so.0 (GLIBC_2.3.4) => /lib64/libpthread.so.0
>>>>>>              libpthread.so.0 (GLIBC_2.3.2) => /lib64/libpthread.so.0
>>>>>>              libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0
>>>>>>              libc.so.6 (GLIBC_2.6) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.7) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.3.2) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>>      /lib64/libdl.so.2:
>>>>>>              ld-linux-x86-64.so.2 (GLIBC_PRIVATE) => 
>>>>>> /lib64/ld-linux-x86-64.so.2
>>>>>>              libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>>      /usr/lib64/libudev.so.1:
>>>>>>              librt.so.1 (GLIBC_2.2.5) => /lib64/librt.so.1
>>>>>>              ld-linux-x86-64.so.2 (GLIBC_2.3) => 
>>>>>> /lib64/ld-linux-x86-64.so.2
>>>>>>              libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0
>>>>>>              libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.9) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.16) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.8) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>>      /usr/lib64/libpciaccess.so.0:
>>>>>>              libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.7) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6
>>>>>>      /lib64/librt.so.1:
>>>>>>              libpthread.so.0 (GLIBC_2.3.2) => /lib64/libpthread.so.0
>>>>>>              libpthread.so.0 (GLIBC_PRIVATE) => /lib64/libpthread.so.0
>>>>>>              libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0
>>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.3.2) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6
>>>>>>      /lib64/libm.so.6:
>>>>>>              libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>>      /lib64/libutil.so.1:
>>>>>>              libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>>      /lib64/libz.so.1:
>>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6
>>>>>>      /lib64/libpthread.so.0:
>>>>>>              ld-linux-x86-64.so.2 (GLIBC_2.2.5) => 
>>>>>> /lib64/ld-linux-x86-64.so.2
>>>>>>              ld-linux-x86-64.so.2 (GLIBC_PRIVATE) => 
>>>>>> /lib64/ld-linux-x86-64.so.2
>>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.3.2) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6
>>>>>>      /lib64/libc.so.6:
>>>>>>              ld-linux-x86-64.so.2 (GLIBC_2.3) => 
>>>>>> /lib64/ld-linux-x86-64.so.2
>>>>>>              ld-linux-x86-64.so.2 (GLIBC_PRIVATE) => 
>>>>>> /lib64/ld-linux-x86-64.so.2
>>>>>>      /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1:
>>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>>      /lib64/libselinux.so.1:
>>>>>>              libdl.so.2 (GLIBC_2.2.5) => /lib64/libdl.so.2
>>>>>>              ld-linux-x86-64.so.2 (GLIBC_2.3) => 
>>>>>> /lib64/ld-linux-x86-64.so.2
>>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.8) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.7) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6
>>>>>>      /lib64/libcap.so.2:
>>>>>>              libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.8) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>>      /lib64/libresolv.so.2:
>>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6
>>>>>>      /usr/lib64/libpcre.so.1:
>>>>>>              libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0
>>>>>>              libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>>>>>>              libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6
>>>>>> loki openmpi 127
>>>>>> 
>>>>>> Hopefully that helps to find the problem. I will answer your emails
>>>>>> tommorrow if you need anything else.
>>>>>> 
>>>>>> 
>>>>>> Best regards
>>>>>> 
>>>>>> Siegmar
>>>>>> 
>>>>>> 
>>>>>> Am 02.10.2018 um 19:48 schrieb Ralph H Castain:
>>>>>>> So the problem is here when configuring the internal PMIx code:
>>>>>>> configure:3383: === HWLOC
>>>>>>> configure:36189: checking for hwloc in
>>>>>>> configure:36201: result: Could not find internal/lib or internal/lib64
>>>>>>> configure:36203: error: Can not continue
>>>>>>> Can you confirm that HWLOC built? I believe we require it, but perhaps 
>>>>>>> something is different about this environment.
>>>>>>>> On Oct 2, 2018, at 6:36 AM, Ralph H Castain <r...@open-mpi.org> wrote:
>>>>>>>> 
>>>>>>>> Looks like PMIx failed to build - can you send the config.log?
>>>>>>>> 
>>>>>>>>> On Oct 2, 2018, at 12:00 AM, Siegmar Gross 
>>>>>>>>> <siegmar.gr...@informatik.hs-fulda.de> wrote:
>>>>>>>>> 
>>>>>>>>> Hi,
>>>>>>>>> 
>>>>>>>>> yesterday I've installed openmpi-v4.0.x-201809290241-a7e275c and
>>>>>>>>> openmpi-master-201805080348-b39bbfb on my "SUSE Linux Enterprise 
>>>>>>>>> Server
>>>>>>>>> 12.3 (x86_64)" with Sun C 5.15, gcc 6.4.0, Intel icc 18.0.3, and 
>>>>>>>>> Portland
>>>>>>>>> Group pgcc 18.4-0. Unfortunately, I get the following error for all 
>>>>>>>>> seven
>>>>>>>>> installed versions (Sun C couldn't built master as I mentioned in 
>>>>>>>>> another
>>>>>>>>> email).
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> loki hello_1 118 mpiexec -np 4 --host loki:2,nfs2:2 hello_1_mpi
>>>>>>>>> [loki:11423] [[45859,0],0] ORTE_ERROR_LOG: Not found in file 
>>>>>>>>> ../../../../../openmpi-v4.0.x-201809290241-a7e275c/orte/mca/ess/hnp/ess_hnp_module.c
>>>>>>>>>  at line 321
>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>> It looks like orte_init failed for some reason; your parallel process 
>>>>>>>>> is
>>>>>>>>> likely to abort.  There are many reasons that a parallel process can
>>>>>>>>> fail during orte_init; some of which are due to configuration or
>>>>>>>>> environment problems.  This failure appears to be an internal failure;
>>>>>>>>> here's some additional information (which may only be relevant to an
>>>>>>>>> Open MPI developer):
>>>>>>>>> 
>>>>>>>>> opal_pmix_base_select failed
>>>>>>>>> --> Returned value Not found (-13) instead of ORTE_SUCCESS
>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>> loki hello_1 119
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> I would be grateful, if somebody can fix the problem. Do you need 
>>>>>>>>> anything
>>>>>>>>> else? Thank you very much for any help in advance.
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> Kind regards
>>>>>>>>> 
>>>>>>>>> Siegmar
>>>>>>>>> _______________________________________________
>>>>>>>>> users mailing list
>>>>>>>>> users@lists.open-mpi.org
>>>>>>>>> https://lists.open-mpi.org/mailman/listinfo/users
>>>>>>>> 
>>>>>>>> _______________________________________________
>>>>>>>> users mailing list
>>>>>>>> users@lists.open-mpi.org
>>>>>>>> https://lists.open-mpi.org/mailman/listinfo/users
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> users@lists.open-mpi.org
>>>>>>> https://lists.open-mpi.org/mailman/listinfo/users
>>>>> 
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> users@lists.open-mpi.org
>>>>> https://lists.open-mpi.org/mailman/listinfo/users
>>>> 
>>>> 
>>>> -- 
>>>> Jeff Squyres
>>>> jsquy...@cisco.com
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> users@lists.open-mpi.org
>>>> https://lists.open-mpi.org/mailman/listinfo/users
>>> _______________________________________________
>>> users mailing list
>>> users@lists.open-mpi.org
>>> https://lists.open-mpi.org/mailman/listinfo/users
>> _______________________________________________
>> users mailing list
>> users@lists.open-mpi.org
>> https://lists.open-mpi.org/mailman/listinfo/users
> 
> _______________________________________________
> users mailing list
> users@lists.open-mpi.org
> https://lists.open-mpi.org/mailman/listinfo/users


-- 
Jeff Squyres
jsquy...@cisco.com

_______________________________________________
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

Reply via email to