Hi Siegmar The patch was merged into the v4.0.0 branch on Oct 10th, so should be available in the nightly tarball from that date onward.
> On Oct 6, 2018, at 2:12 AM, Siegmar Gross > <siegmar.gr...@informatik.hs-fulda.de> wrote: > > Hi Jeff, hi Ralph, > > Great, it works again! Thank you very much for your help. I'm really happy, > if the undefined references for Sun C are resolved and there are no new > problems for that compiler :-)). Do you know when the pmix patch will be > integrated into version 4.0.0? > > > Best regards > > Siegmar > > > On 10/5/18 4:33 PM, Jeff Squyres (jsquyres) via users wrote: >> Oops! We had a typo in yesterday's fix -- fixed: >> https://github.com/open-mpi/ompi/pull/5847 >> Ralph also put double extra super protection to make triple sure that this >> error can't happen again in: >> https://github.com/open-mpi/ompi/pull/5846 >> Both of these should be in tonight's nightly snapshot. >> Thank you! >>> On Oct 5, 2018, at 5:45 AM, Ralph H Castain <r...@open-mpi.org> wrote: >>> >>> Please send Jeff and I the opal/mca/pmix/pmix4x/pmix/config.log again - >>> we’ll need to see why it isn’t building. The patch definitely is not in the >>> v4.0 branch, but it should have been in master. >>> >>> >>>> On Oct 5, 2018, at 2:04 AM, Siegmar Gross >>>> <siegmar.gr...@informatik.hs-fulda.de> wrote: >>>> >>>> Hi Ralph, hi Jeff, >>>> >>>> >>>> On 10/3/18 8:14 PM, Ralph H Castain wrote: >>>>> Jeff and I talked and believe the patch in >>>>> https://github.com/open-mpi/ompi/pull/5836 should fix the problem. >>>> >>>> >>>> Today I've installed openmpi-master-201810050304-5f1c940 and >>>> openmpi-v4.0.x-201810050241-c079666. Unfortunately, I still get the >>>> same error for all seven versions that I was able to build. >>>> >>>> loki hello_1 114 mpicc --showme >>>> gcc -I/usr/local/openmpi-master_64_gcc/include -fexceptions -pthread >>>> -std=c11 -m64 -Wl,-rpath -Wl,/usr/local/openmpi-master_64_gcc/lib64 >>>> -Wl,--enable-new-dtags -L/usr/local/openmpi-master_64_gcc/lib64 -lmpi >>>> >>>> loki hello_1 115 ompi_info | grep "Open MPI repo revision" >>>> Open MPI repo revision: v2.x-dev-6262-g5f1c940 >>>> >>>> loki hello_1 116 mpicc hello_1_mpi.c >>>> >>>> loki hello_1 117 mpiexec -np 2 a.out >>>> [loki:25575] [[64603,0],0] ORTE_ERROR_LOG: Not found in file >>>> ../../../../../openmpi-master-201810050304-5f1c940/orte/mca/ess/hnp/ess_hnp_module.c >>>> at line 320 >>>> -------------------------------------------------------------------------- >>>> It looks like orte_init failed for some reason; your parallel process is >>>> likely to abort. There are many reasons that a parallel process can >>>> fail during orte_init; some of which are due to configuration or >>>> environment problems. This failure appears to be an internal failure; >>>> here's some additional information (which may only be relevant to an >>>> Open MPI developer): >>>> >>>> opal_pmix_base_select failed >>>> --> Returned value Not found (-13) instead of ORTE_SUCCESS >>>> -------------------------------------------------------------------------- >>>> loki hello_1 118 >>>> >>>> >>>> I don't know, if you have already applied your suggested patch or if the >>>> error message is still from a version without that patch. Do you need >>>> anything else? >>>> >>>> >>>> Best regards >>>> >>>> Siegmar >>>> >>>> >>>>>> On Oct 2, 2018, at 2:50 PM, Jeff Squyres (jsquyres) via users >>>>>> <users@lists.open-mpi.org> wrote: >>>>>> >>>>>> (Ralph sent me Siegmar's pmix config.log, which Siegmar sent to him >>>>>> off-list) >>>>>> >>>>>> It looks like Siegmar passed --with-hwloc=internal. >>>>>> >>>>>> Open MPI's configure understood this and did the appropriate things. >>>>>> PMIX's configure didn't. >>>>>> >>>>>> I think we need to add an adjustment into the PMIx configure.m4 in >>>>>> OMPI... >>>>>> >>>>>> >>>>>>> On Oct 2, 2018, at 5:25 PM, Ralph H Castain <r...@open-mpi.org> wrote: >>>>>>> >>>>>>> Hi Siegmar >>>>>>> >>>>>>> I honestly have no idea - for some reason, the PMIx component isn’t >>>>>>> seeing the internal hwloc code in your environment. >>>>>>> >>>>>>> Jeff, Brice - any ideas? >>>>>>> >>>>>>> >>>>>>>> On Oct 2, 2018, at 1:18 PM, Siegmar Gross >>>>>>>> <siegmar.gr...@informatik.hs-fulda.de> wrote: >>>>>>>> >>>>>>>> Hi Ralph, >>>>>>>> >>>>>>>> how can I confirm that HWLOC built? Some hwloc files are available >>>>>>>> in the built directory. >>>>>>>> >>>>>>>> loki openmpi-master-201809290304-73075b8-Linux.x86_64.64_gcc 111 find >>>>>>>> . -name '*hwloc*' >>>>>>>> ./opal/mca/btl/usnic/.deps/btl_usnic_hwloc.Plo >>>>>>>> ./opal/mca/hwloc >>>>>>>> ./opal/mca/hwloc/external/.deps/hwloc_external_component.Plo >>>>>>>> ./opal/mca/hwloc/base/hwloc_base_frame.lo >>>>>>>> ./opal/mca/hwloc/base/.deps/hwloc_base_dt.Plo >>>>>>>> ./opal/mca/hwloc/base/.deps/hwloc_base_maffinity.Plo >>>>>>>> ./opal/mca/hwloc/base/.deps/hwloc_base_frame.Plo >>>>>>>> ./opal/mca/hwloc/base/.deps/hwloc_base_util.Plo >>>>>>>> ./opal/mca/hwloc/base/hwloc_base_dt.lo >>>>>>>> ./opal/mca/hwloc/base/hwloc_base_util.lo >>>>>>>> ./opal/mca/hwloc/base/hwloc_base_maffinity.lo >>>>>>>> ./opal/mca/hwloc/base/.libs/hwloc_base_util.o >>>>>>>> ./opal/mca/hwloc/base/.libs/hwloc_base_dt.o >>>>>>>> ./opal/mca/hwloc/base/.libs/hwloc_base_maffinity.o >>>>>>>> ./opal/mca/hwloc/base/.libs/hwloc_base_frame.o >>>>>>>> ./opal/mca/hwloc/.libs/libmca_hwloc.la >>>>>>>> ./opal/mca/hwloc/.libs/libmca_hwloc.a >>>>>>>> ./opal/mca/hwloc/libmca_hwloc.la >>>>>>>> ./opal/mca/hwloc/hwloc201 >>>>>>>> ./opal/mca/hwloc/hwloc201/.deps/hwloc201_component.Plo >>>>>>>> ./opal/mca/hwloc/hwloc201/hwloc201_component.lo >>>>>>>> ./opal/mca/hwloc/hwloc201/hwloc >>>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/include/hwloc >>>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc >>>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/libhwloc_embedded.la >>>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_pci_la-topology-pci.Plo >>>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_gl_la-topology-gl.Plo >>>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_cuda_la-topology-cuda.Plo >>>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_xml_libxml_la-topology-xml-libxml.Plo >>>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_opencl_la-topology-opencl.Plo >>>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_nvml_la-topology-nvml.Plo >>>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.libs/libhwloc_embedded.la >>>>>>>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.libs/libhwloc_embedded.a >>>>>>>> ./opal/mca/hwloc/hwloc201/.libs/hwloc201_component.o >>>>>>>> ./opal/mca/hwloc/hwloc201/.libs/libmca_hwloc_hwloc201.la >>>>>>>> ./opal/mca/hwloc/hwloc201/.libs/libmca_hwloc_hwloc201.a >>>>>>>> ./opal/mca/hwloc/hwloc201/libmca_hwloc_hwloc201.la >>>>>>>> ./orte/mca/rtc/hwloc >>>>>>>> ./orte/mca/rtc/hwloc/rtc_hwloc.lo >>>>>>>> ./orte/mca/rtc/hwloc/.deps/rtc_hwloc.Plo >>>>>>>> ./orte/mca/rtc/hwloc/.deps/rtc_hwloc_component.Plo >>>>>>>> ./orte/mca/rtc/hwloc/mca_rtc_hwloc.la >>>>>>>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.so >>>>>>>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.la >>>>>>>> ./orte/mca/rtc/hwloc/.libs/rtc_hwloc.o >>>>>>>> ./orte/mca/rtc/hwloc/.libs/rtc_hwloc_component.o >>>>>>>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.soT >>>>>>>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.lai >>>>>>>> ./orte/mca/rtc/hwloc/rtc_hwloc_component.lo >>>>>>>> loki openmpi-master-201809290304-73075b8-Linux.x86_64.64_gcc 112 >>>>>>>> >>>>>>>> And some files are available in the install directory. >>>>>>>> >>>>>>>> loki openmpi-master_64_gcc 116 find . -name '*hwloc*' >>>>>>>> ./share/openmpi/help-orte-rtc-hwloc.txt >>>>>>>> ./share/openmpi/help-opal-hwloc-base.txt >>>>>>>> ./lib64/openmpi/mca_rtc_hwloc.so >>>>>>>> ./lib64/openmpi/mca_rtc_hwloc.la >>>>>>>> loki openmpi-master_64_gcc 117 >>>>>>>> >>>>>>>> I don't see any unavailable libraries so that the only available >>>>>>>> hwloc library should work. >>>>>>>> >>>>>>>> loki openmpi 126 ldd -v mca_rtc_hwloc.so >>>>>>>> linux-vdso.so.1 (0x00007ffd2df5b000) >>>>>>>> libopen-rte.so.0 => >>>>>>>> /usr/local/openmpi-master_64_gcc/lib64/libopen-rte.so.0 >>>>>>>> (0x00007f082b7fb000) >>>>>>>> libopen-pal.so.0 => >>>>>>>> /usr/local/openmpi-master_64_gcc/lib64/libopen-pal.so.0 >>>>>>>> (0x00007f082b493000) >>>>>>>> libdl.so.2 => /lib64/libdl.so.2 (0x00007f082b28f000) >>>>>>>> libudev.so.1 => /usr/lib64/libudev.so.1 (0x00007f082b06e000) >>>>>>>> libpciaccess.so.0 => /usr/lib64/libpciaccess.so.0 >>>>>>>> (0x00007f082ae64000) >>>>>>>> librt.so.1 => /lib64/librt.so.1 (0x00007f082ac5c000) >>>>>>>> libm.so.6 => /lib64/libm.so.6 (0x00007f082a95f000) >>>>>>>> libutil.so.1 => /lib64/libutil.so.1 (0x00007f082a75c000) >>>>>>>> libz.so.1 => /lib64/libz.so.1 (0x00007f082a546000) >>>>>>>> libpthread.so.0 => /lib64/libpthread.so.0 (0x00007f082a329000) >>>>>>>> libc.so.6 => /lib64/libc.so.6 (0x00007f0829f84000) >>>>>>>> libgcc_s.so.1 => /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1 >>>>>>>> (0x00007f0829d6c000) >>>>>>>> /lib64/ld-linux-x86-64.so.2 (0x00007f082bd24000) >>>>>>>> libselinux.so.1 => /lib64/libselinux.so.1 (0x00007f0829b46000) >>>>>>>> libcap.so.2 => /lib64/libcap.so.2 (0x00007f0829941000) >>>>>>>> libresolv.so.2 => /lib64/libresolv.so.2 (0x00007f082972a000) >>>>>>>> libpcre.so.1 => /usr/lib64/libpcre.so.1 (0x00007f08294bb000) >>>>>>>> >>>>>>>> Version information: >>>>>>>> ./mca_rtc_hwloc.so: >>>>>>>> libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0 >>>>>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>>>>> /usr/local/openmpi-master_64_gcc/lib64/libopen-rte.so.0: >>>>>>>> libz.so.1 (ZLIB_1.2.0) => /lib64/libz.so.1 >>>>>>>> libpthread.so.0 (GLIBC_2.3.2) => /lib64/libpthread.so.0 >>>>>>>> libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0 >>>>>>>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>>>>> /usr/local/openmpi-master_64_gcc/lib64/libopen-pal.so.0: >>>>>>>> librt.so.1 (GLIBC_2.2.5) => /lib64/librt.so.1 >>>>>>>> libgcc_s.so.1 (GCC_3.0) => >>>>>>>> /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1 >>>>>>>> libgcc_s.so.1 (GCC_3.3.1) => >>>>>>>> /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1 >>>>>>>> libdl.so.2 (GLIBC_2.2.5) => /lib64/libdl.so.2 >>>>>>>> libutil.so.1 (GLIBC_2.2.5) => /lib64/libutil.so.1 >>>>>>>> libudev.so.1 (LIBUDEV_183) => /usr/lib64/libudev.so.1 >>>>>>>> libm.so.6 (GLIBC_2.2.5) => /lib64/libm.so.6 >>>>>>>> libpthread.so.0 (GLIBC_2.3.4) => /lib64/libpthread.so.0 >>>>>>>> libpthread.so.0 (GLIBC_2.3.2) => /lib64/libpthread.so.0 >>>>>>>> libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0 >>>>>>>> libc.so.6 (GLIBC_2.6) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.7) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.3.2) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>>>>> /lib64/libdl.so.2: >>>>>>>> ld-linux-x86-64.so.2 (GLIBC_PRIVATE) => >>>>>>>> /lib64/ld-linux-x86-64.so.2 >>>>>>>> libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>>>>> /usr/lib64/libudev.so.1: >>>>>>>> librt.so.1 (GLIBC_2.2.5) => /lib64/librt.so.1 >>>>>>>> ld-linux-x86-64.so.2 (GLIBC_2.3) => >>>>>>>> /lib64/ld-linux-x86-64.so.2 >>>>>>>> libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0 >>>>>>>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.9) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.16) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.8) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>>>>> /usr/lib64/libpciaccess.so.0: >>>>>>>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.7) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>>>>>>> /lib64/librt.so.1: >>>>>>>> libpthread.so.0 (GLIBC_2.3.2) => /lib64/libpthread.so.0 >>>>>>>> libpthread.so.0 (GLIBC_PRIVATE) => /lib64/libpthread.so.0 >>>>>>>> libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0 >>>>>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.3.2) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6 >>>>>>>> /lib64/libm.so.6: >>>>>>>> libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>>>>> /lib64/libutil.so.1: >>>>>>>> libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>>>>> /lib64/libz.so.1: >>>>>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>>>>>>> /lib64/libpthread.so.0: >>>>>>>> ld-linux-x86-64.so.2 (GLIBC_2.2.5) => >>>>>>>> /lib64/ld-linux-x86-64.so.2 >>>>>>>> ld-linux-x86-64.so.2 (GLIBC_PRIVATE) => >>>>>>>> /lib64/ld-linux-x86-64.so.2 >>>>>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.3.2) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6 >>>>>>>> /lib64/libc.so.6: >>>>>>>> ld-linux-x86-64.so.2 (GLIBC_2.3) => >>>>>>>> /lib64/ld-linux-x86-64.so.2 >>>>>>>> ld-linux-x86-64.so.2 (GLIBC_PRIVATE) => >>>>>>>> /lib64/ld-linux-x86-64.so.2 >>>>>>>> /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1: >>>>>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>>>>> /lib64/libselinux.so.1: >>>>>>>> libdl.so.2 (GLIBC_2.2.5) => /lib64/libdl.so.2 >>>>>>>> ld-linux-x86-64.so.2 (GLIBC_2.3) => >>>>>>>> /lib64/ld-linux-x86-64.so.2 >>>>>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.8) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.7) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>>>>>>> /lib64/libcap.so.2: >>>>>>>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.8) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>>>>> /lib64/libresolv.so.2: >>>>>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_PRIVATE) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>>>>>>> /usr/lib64/libpcre.so.1: >>>>>>>> libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0 >>>>>>>> libc.so.6 (GLIBC_2.14) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.3.4) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.4) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6 >>>>>>>> libc.so.6 (GLIBC_2.3) => /lib64/libc.so.6 >>>>>>>> loki openmpi 127 >>>>>>>> >>>>>>>> Hopefully that helps to find the problem. I will answer your emails >>>>>>>> tommorrow if you need anything else. >>>>>>>> >>>>>>>> >>>>>>>> Best regards >>>>>>>> >>>>>>>> Siegmar >>>>>>>> >>>>>>>> >>>>>>>> Am 02.10.2018 um 19:48 schrieb Ralph H Castain: >>>>>>>>> So the problem is here when configuring the internal PMIx code: >>>>>>>>> configure:3383: === HWLOC >>>>>>>>> configure:36189: checking for hwloc in >>>>>>>>> configure:36201: result: Could not find internal/lib or internal/lib64 >>>>>>>>> configure:36203: error: Can not continue >>>>>>>>> Can you confirm that HWLOC built? I believe we require it, but >>>>>>>>> perhaps something is different about this environment. >>>>>>>>>> On Oct 2, 2018, at 6:36 AM, Ralph H Castain <r...@open-mpi.org> >>>>>>>>>> wrote: >>>>>>>>>> >>>>>>>>>> Looks like PMIx failed to build - can you send the config.log? >>>>>>>>>> >>>>>>>>>>> On Oct 2, 2018, at 12:00 AM, Siegmar Gross >>>>>>>>>>> <siegmar.gr...@informatik.hs-fulda.de> wrote: >>>>>>>>>>> >>>>>>>>>>> Hi, >>>>>>>>>>> >>>>>>>>>>> yesterday I've installed openmpi-v4.0.x-201809290241-a7e275c and >>>>>>>>>>> openmpi-master-201805080348-b39bbfb on my "SUSE Linux Enterprise >>>>>>>>>>> Server >>>>>>>>>>> 12.3 (x86_64)" with Sun C 5.15, gcc 6.4.0, Intel icc 18.0.3, and >>>>>>>>>>> Portland >>>>>>>>>>> Group pgcc 18.4-0. Unfortunately, I get the following error for all >>>>>>>>>>> seven >>>>>>>>>>> installed versions (Sun C couldn't built master as I mentioned in >>>>>>>>>>> another >>>>>>>>>>> email). >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> loki hello_1 118 mpiexec -np 4 --host loki:2,nfs2:2 hello_1_mpi >>>>>>>>>>> [loki:11423] [[45859,0],0] ORTE_ERROR_LOG: Not found in file >>>>>>>>>>> ../../../../../openmpi-v4.0.x-201809290241-a7e275c/orte/mca/ess/hnp/ess_hnp_module.c >>>>>>>>>>> at line 321 >>>>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>>>> It looks like orte_init failed for some reason; your parallel >>>>>>>>>>> process is >>>>>>>>>>> likely to abort. There are many reasons that a parallel process can >>>>>>>>>>> fail during orte_init; some of which are due to configuration or >>>>>>>>>>> environment problems. This failure appears to be an internal >>>>>>>>>>> failure; >>>>>>>>>>> here's some additional information (which may only be relevant to an >>>>>>>>>>> Open MPI developer): >>>>>>>>>>> >>>>>>>>>>> opal_pmix_base_select failed >>>>>>>>>>> --> Returned value Not found (-13) instead of ORTE_SUCCESS >>>>>>>>>>> -------------------------------------------------------------------------- >>>>>>>>>>> loki hello_1 119 >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> I would be grateful, if somebody can fix the problem. Do you need >>>>>>>>>>> anything >>>>>>>>>>> else? Thank you very much for any help in advance. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Kind regards >>>>>>>>>>> >>>>>>>>>>> Siegmar >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> users mailing list >>>>>>>>>>> users@lists.open-mpi.org >>>>>>>>>>> https://lists.open-mpi.org/mailman/listinfo/users >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> users mailing list >>>>>>>>>> users@lists.open-mpi.org >>>>>>>>>> https://lists.open-mpi.org/mailman/listinfo/users >>>>>>>>> _______________________________________________ >>>>>>>>> users mailing list >>>>>>>>> users@lists.open-mpi.org >>>>>>>>> https://lists.open-mpi.org/mailman/listinfo/users >>>>>>> >>>>>>> _______________________________________________ >>>>>>> users mailing list >>>>>>> users@lists.open-mpi.org >>>>>>> https://lists.open-mpi.org/mailman/listinfo/users >>>>>> >>>>>> >>>>>> -- >>>>>> Jeff Squyres >>>>>> jsquy...@cisco.com >>>>>> >>>>>> _______________________________________________ >>>>>> users mailing list >>>>>> users@lists.open-mpi.org >>>>>> https://lists.open-mpi.org/mailman/listinfo/users >>>>> _______________________________________________ >>>>> users mailing list >>>>> users@lists.open-mpi.org >>>>> https://lists.open-mpi.org/mailman/listinfo/users >>>> _______________________________________________ >>>> users mailing list >>>> users@lists.open-mpi.org >>>> https://lists.open-mpi.org/mailman/listinfo/users >>> >>> _______________________________________________ >>> users mailing list >>> users@lists.open-mpi.org >>> https://lists.open-mpi.org/mailman/listinfo/users > > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://lists.open-mpi.org/mailman/listinfo/users _______________________________________________ users mailing list users@lists.open-mpi.org https://lists.open-mpi.org/mailman/listinfo/users