Ouch - thanks for finding this, Gilles! I’ll take care of it on Friday.

> On Apr 28, 2016, at 6:38 PM, Gilles Gouaillardet 
> <gilles.gouaillar...@gmail.com> wrote:
> 
> Siegmar,
> 
> in pmix_bfrop_pack_app,
> app->argc
> must be replaced with
> app[i].argc
> 
> I will PR to pmix, ompi and ompi-release when I am back at work on Monday
> 
> Cheers,
> 
> Gilles
> 
> On Thursday, April 28, 2016, Gilles Gouaillardet <gil...@rist.or.jp 
> <mailto:gil...@rist.or.jp>> wrote:
> Siegmar,
> 
> 
> 
> can you please also post the source of spawn_slave ?
> 
> 
> 
> Cheers,
> 
> Gilles
> 
> 
> On 4/28/2016 1:17 AM, Siegmar Gross wrote:
>> Hi Gilles, 
>> 
>> it is not necessary to have a heterogeneous environment to reproduce 
>> the error as you can see below. All machines are 64 bit. 
>> 
>> tyr spawn 119 ompi_info | grep -e "OPAL repo revision" -e "C compiler 
>> absolute" 
>>       OPAL repo revision: v2.x-dev-1290-gbd0e4e1 
>>      C compiler absolute: /usr/local/gcc-5.1.0/bin/gcc 
>> tyr spawn 120 uname -a 
>> SunOS tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/> 5.10 
>> Generic_150400-11 sun4u sparc SUNW,A70 Solaris 
>> tyr spawn 121 mpiexec -np 1 --host tyr,tyr,tyr,tyr spawn_multiple_master 
>> 
>> Parent process 0 running on tyr.informatik.hs-fulda.de 
>> <http://tyr.informatik.hs-fulda.de/> 
>>   I create 3 slave processes. 
>> 
>> [tyr.informatik.hs-fulda.de:27286 
>> <http://tyr.informatik.hs-fulda.de:27286/>] PMIX ERROR: UNPACK-PAST-END in 
>> file 
>> ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c
>>  at line 829 
>> [tyr.informatik.hs-fulda.de:27286 
>> <http://tyr.informatik.hs-fulda.de:27286/>] PMIX ERROR: UNPACK-PAST-END in 
>> file 
>> ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c
>>  at line 2176 
>> [tyr:27288] *** An error occurred in MPI_Comm_spawn_multiple 
>> [tyr:27288] *** reported by process [3434086401,0] 
>> [tyr:27288] *** on communicator MPI_COMM_WORLD 
>> [tyr:27288] *** MPI_ERR_SPAWN: could not spawn processes 
>> [tyr:27288] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will 
>> now abort, 
>> [tyr:27288] ***    and potentially your MPI job) 
>> tyr spawn 122 
>> 
>> 
>> 
>> 
>> 
>> 
>> sunpc1 fd1026 105 ompi_info | grep -e "OPAL repo revision" -e "C compiler 
>> absolute" 
>>       OPAL repo revision: v2.x-dev-1290-gbd0e4e1 
>>      C compiler absolute: /usr/local/gcc-5.1.0/bin/gcc 
>> sunpc1 fd1026 106 uname -a 
>> SunOS sunpc1 5.10 Generic_147441-21 i86pc i386 i86pc Solaris 
>> sunpc1 fd1026 107 mpiexec -np 1 --host sunpc1,sunpc1,sunpc1,sunpc1 
>> spawn_multiple_master 
>> 
>> Parent process 0 running on sunpc1 
>>   I create 3 slave processes. 
>> 
>> [sunpc1:00368] PMIX ERROR: UNPACK-PAST-END in file 
>> ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c
>>  at line 829 
>> [sunpc1:00368] PMIX ERROR: UNPACK-PAST-END in file 
>> ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c
>>  at line 2176 
>> [sunpc1:370] *** An error occurred in MPI_Comm_spawn_multiple 
>> [sunpc1:370] *** reported by process [43909121,0] 
>> [sunpc1:370] *** on communicator MPI_COMM_WORLD 
>> [sunpc1:370] *** MPI_ERR_SPAWN: could not spawn processes 
>> [sunpc1:370] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will 
>> now abort, 
>> [sunpc1:370] ***    and potentially your MPI job) 
>> sunpc1 fd1026 108 
>> 
>> 
>> 
>> 
>> 
>> linpc1 fd1026 105 ompi_info | grep -e "OPAL repo revision" -e "C compiler 
>> absolute" 
>>       OPAL repo revision: v2.x-dev-1290-gbd0e4e1 
>>      C compiler absolute: /usr/local/gcc-5.1.0/bin/gcc 
>> linpc1 fd1026 106 uname -a 
>> Linux linpc1 3.1.10-1.29-desktop #1 SMP PREEMPT Fri May 31 20:10:04 UTC 2013 
>> (2529847) x86_64 x86_64 x86_64 GNU/Linux 
>> linpc1 fd1026 107 mpiexec -np 1 --host linpc1,linpc1,linpc1,linpc1 
>> spawn_multiple_master 
>> 
>> Parent process 0 running on linpc1 
>>   I create 3 slave processes. 
>> 
>> [linpc1:21502] PMIX ERROR: UNPACK-PAST-END in file 
>> ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c
>>  at line 829 
>> [linpc1:21502] PMIX ERROR: UNPACK-PAST-END in file 
>> ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c
>>  at line 2176 
>> [linpc1:21507] *** An error occurred in MPI_Comm_spawn_multiple 
>> [linpc1:21507] *** reported by process [1005518849,0] 
>> [linpc1:21507] *** on communicator MPI_COMM_WORLD 
>> [linpc1:21507] *** MPI_ERR_SPAWN: could not spawn processes 
>> [linpc1:21507] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will 
>> now abort, 
>> [linpc1:21507] ***    and potentially your MPI job) 
>> linpc1 fd1026 108 
>> 
>> 
>> I used the following configure command. 
>> 
>> ../openmpi-v2.x-dev-1290-gbd0e4e1/configure \ 
>>   --prefix=/usr/local/openmpi-2.0.0_64_gcc \ 
>>   --libdir=/usr/local/openmpi-2.0.0_64_gcc/lib64 \ 
>>   --with-jdk-bindir=/usr/local/jdk1.8.0/bin \ 
>>   --with-jdk-headers=/usr/local/jdk1.8.0/include \ 
>>   JAVA_HOME=/usr/local/jdk1.8.0 \ 
>>   LDFLAGS="-m64" CC="gcc" CXX="g++" FC="gfortran" \ 
>>   CFLAGS="-m64" CXXFLAGS="-m64" FCFLAGS="-m64" \ 
>>   CPP="cpp" CXXCPP="cpp" \ 
>>   --enable-mpi-cxx \ 
>>   --enable-cxx-exceptions \ 
>>   --enable-mpi-java \ 
>>   --enable-heterogeneous \ 
>>   --enable-mpi-thread-multiple \ 
>>   --with-hwloc=internal \ 
>>   --without-verbs \ 
>>   --with-wrapper-cflags="-std=c11 -m64" \ 
>>   --with-wrapper-cxxflags="-m64" \ 
>>   --with-wrapper-fcflags="-m64" \ 
>>   --enable-debug \ 
>>   |& tee log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_gcc 
>> 
>> 
>> Kind regards 
>> 
>> Siegmar 
>> 
>> 
>> 
>> Am 27.04.2016 um 13:21 schrieb Gilles Gouaillardet: 
>>> Siegmar, 
>>> 
>>> please add this to your CFLAGS for the time being. 
>>> 
>>> configure tries to detect which flags must be added for C99 support, and it 
>>> seems 
>>> the test is not working for Solaris 10 and Oracle compilers. 
>>> this is no more a widely used environment, and I am not sure I can find the 
>>> time to fix this 
>>> in a near future. 
>>> 
>>> 
>>> regarding the runtime issue, can you please describe your 4 hosts (os, 
>>> endianness and bitness) 
>>> 
>>> Cheers, 
>>> 
>>> Gilles 
>>> 
>>> On Wednesday, April 27, 2016, Siegmar Gross 
>>> <siegmar.gr...@informatik.hs-fulda.de <> 
>>> <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de');> <>> 
>>> wrote: 
>>> 
>>>     Hi Gilles, 
>>> 
>>>     adding "-std=c99" to CFLAGS solves the problem with the missing 
>>> library. 
>>>     Shall I add it permanently to my configure command or will you add it, 
>>>     so that I will not run into problems if you need the C11 standard 
>>> later? 
>>> 
>>>     "spawn_multiple_master" breaks with the same error that I reported 
>>>     yesterday for my gcc-version of Open MPI. Hopefully you can solve the 
>>>     problem as well. 
>>> 
>>> 
>>>     Kind regards and thank you very much for your help 
>>> 
>>>     Siegmar 
>>> 
>>> 
>>>     Am 27.04.2016 um 08:05 schrieb Gilles Gouaillardet: 
>>> 
>>>         Siegmar, 
>>> 
>>> 
>>>         here is the error : 
>>> 
>>>         configure:17969: cc -o conftest -m64 -D_REENTRANT -g  -g 
>>>         -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1 
>>>         
>>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc
>>>         
>>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/include 
>>>         
>>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/include
>>>         -D_REENTRANT 
>>>         
>>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/hwloc/hwloc1112/hwloc/include
>>>         
>>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/hwloc/hwloc1112/hwloc/include
>>>         
>>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/event/libevent2022/libevent
>>>         
>>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/event/libevent2022/libevent/include
>>>         
>>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/event/libevent2022/libevent/include
>>>         -m64 conftest.c  >&5 
>>>         "/usr/include/stdbool.h", line 42: #error: "Use of <stdbool.h> is 
>>>         valid only 
>>>         in a c99 compilation environment." 
>>> 
>>> 
>>>         i cannot reproduce this on solaris 11 with oracle studio 5.3 
>>> compiler, 
>>>         and i 
>>>         do not have solaris 10 yet. 
>>> 
>>>         could you please re-configure with '-std=c99' appended to your 
>>> CFLAGS 
>>>         and see 
>>>         if it helps ? 
>>> 
>>> 
>>>         Cheers, 
>>> 
>>> 
>>>         Gilles 
>>> 
>>> 
>>>         On 4/26/2016 7:57 PM, Siegmar Gross wrote: 
>>> 
>>>             Hi Gilles and Ralph, 
>>> 
>>>             I was able to sort out my mess. In my last email I compared the 
>>>             files from "SunOS_sparc/openmpi-2.0.0_64_gcc/lib64/openmpi" 
>>> from 
>>>             the attachment of my email to Ralph with the files from 
>>>             "SunOS_sparc/openmpi-2.0.0_64_cc/lib64/openmpi" from my current 
>>>             file system. That's the reason while I have had different 
>>>             timestamps. The other problem was that Ralph didn't recognize 
>>>             that "mca_pmix_pmix112.so" wasn't built on Solaris with the 
>>>             Sun C compiler. I've removed most of the files from the 
>>> attachment 
>>>             of my email so that it is easier to see the relevant files. 
>>> Below 
>>>             I try to give you more information that may be relevant to 
>>> track 
>>>             down the problem. I still get an error running one of my small 
>>>             test programs, when I use my gcc-version of Open MPI. 
>>>             "mca_pmix_pmix112.so" is a 64 bits library. 
>>> 
>>>             Linux_x86_64/openmpi-2.0.0_64_cc/lib64/openmpi: 
>>>             ... 
>>>             -rwxr-xr-x 1 root root  261327 Apr 19 16:46 mca_plm_slurm.so 
>>>             -rwxr-xr-x 1 root root    1002 Apr 19 16:45 mca_pmix_pmix112.la 
>>> <http://mca_pmix_pmix112.la/> 
>>>             <http://mca_pmix_pmix112.la> <http://mca_pmix_pmix112.la/> 
>>>             -rwxr-xr-x 1 root root 3906526 Apr 19 16:45 mca_pmix_pmix112.so 
>>>             -rwxr-xr-x 1 root root     966 Apr 19 16:51 mca_pml_cm.la 
>>> <http://mca_pml_cm.la/> 
>>>             <http://mca_pml_cm.la> <http://mca_pml_cm.la/> 
>>>             -rwxr-xr-x 1 root root 1574265 Apr 19 16:51 mca_pml_cm.so 
>>>             ... 
>>> 
>>>             Linux_x86_64/openmpi-2.0.0_64_gcc/lib64/openmpi: 
>>>             ... 
>>>             -rwxr-xr-x 1 root root   70371 Apr 19 16:43 mca_plm_slurm.so 
>>>             -rwxr-xr-x 1 root root    1008 Apr 19 16:42 mca_pmix_pmix112.la 
>>> <http://mca_pmix_pmix112.la/> 
>>>             <http://mca_pmix_pmix112.la> <http://mca_pmix_pmix112.la/> 
>>>             -rwxr-xr-x 1 root root 1029005 Apr 19 16:42 mca_pmix_pmix112.so 
>>>             -rwxr-xr-x 1 root root     972 Apr 19 16:46 mca_pml_cm.la 
>>> <http://mca_pml_cm.la/> 
>>>             <http://mca_pml_cm.la> <http://mca_pml_cm.la/> 
>>>             -rwxr-xr-x 1 root root  284858 Apr 19 16:46 mca_pml_cm.so 
>>>             ... 
>>> 
>>>             SunOS_sparc/openmpi-2.0.0_64_cc/lib64/openmpi: 
>>>             ... 
>>>             -rwxr-xr-x 1 root root  319816 Apr 19 19:58 mca_plm_rsh.so 
>>>             -rwxr-xr-x 1 root root     970 Apr 19 20:00 mca_pml_cm.la 
>>> <http://mca_pml_cm.la/> 
>>>             <http://mca_pml_cm.la> <http://mca_pml_cm.la/> 
>>>             -rwxr-xr-x 1 root root 1507440 Apr 19 20:00 mca_pml_cm.so 
>>>             ... 
>>> 
>>>             SunOS_sparc/openmpi-2.0.0_64_gcc/lib64/openmpi: 
>>>             ... 
>>>             -rwxr-xr-x 1 root root  153280 Apr 19 19:49 mca_plm_rsh.so 
>>>             -rwxr-xr-x 1 root root    1007 Apr 19 19:47 mca_pmix_pmix112.la 
>>> <http://mca_pmix_pmix112.la/> 
>>>             <http://mca_pmix_pmix112.la> <http://mca_pmix_pmix112.la/> 
>>>             -rwxr-xr-x 1 root root 1400512 Apr 19 19:47 mca_pmix_pmix112.so 
>>>             -rwxr-xr-x 1 root root     971 Apr 19 19:52 mca_pml_cm.la 
>>> <http://mca_pml_cm.la/> 
>>>             <http://mca_pml_cm.la> <http://mca_pml_cm.la/> 
>>>             -rwxr-xr-x 1 root root  342440 Apr 19 19:52 mca_pml_cm.so 
>>>             ... 
>>> 
>>>             SunOS_x86_64/openmpi-2.0.0_64_cc/lib64/openmpi: 
>>>             ... 
>>>             -rwxr-xr-x 1 root root  300096 Apr 19 17:18 mca_plm_rsh.so 
>>>             -rwxr-xr-x 1 root root     970 Apr 19 17:23 mca_pml_cm.la 
>>> <http://mca_pml_cm.la/> 
>>>             <http://mca_pml_cm.la> <http://mca_pml_cm.la/> 
>>>             -rwxr-xr-x 1 root root 1458816 Apr 19 17:23 mca_pml_cm.so 
>>>             ... 
>>> 
>>>             SunOS_x86_64/openmpi-2.0.0_64_gcc/lib64/openmpi: 
>>>             ... 
>>>             -rwxr-xr-x 1 root root  133096 Apr 19 17:42 mca_plm_rsh.so 
>>>             -rwxr-xr-x 1 root root    1007 Apr 19 17:41 mca_pmix_pmix112.la 
>>> <http://mca_pmix_pmix112.la/> 
>>>             <http://mca_pmix_pmix112.la> <http://mca_pmix_pmix112.la/> 
>>>             -rwxr-xr-x 1 root root 1320240 Apr 19 17:41 mca_pmix_pmix112.so 
>>>             -rwxr-xr-x 1 root root     971 Apr 19 17:46 mca_pml_cm.la 
>>> <http://mca_pml_cm.la/> 
>>>             <http://mca_pml_cm.la> <http://mca_pml_cm.la/> 
>>>             -rwxr-xr-x 1 root root  419848 Apr 19 17:46 mca_pml_cm.so 
>>>             ... 
>>> 
>>> 
>>>             Yesterday I've installed openmpi-v2.x-dev-1290-gbd0e4e1 so that 
>>> we 
>>>             have a current version for the investigation of the problem. 
>>> Once 
>>>             more mca_pmix_pmix112.so isn't available on Solaris if I use 
>>> the 
>>>             Sun C compiler. 
>>> 
>>>             "config.log" for gcc-5.1.0 shows the following. 
>>> 
>>>             ... 
>>>             configure:127799: /bin/bash 
>>>             
>>> '../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/ 
>>>             pmix/configure' succeeded for opal/mca/pmix/pmix112/pmix 
>>>             configure:127916: checking if MCA component pmix:pmix112 can 
>>> compile 
>>>             configure:127918: result: yes 
>>>             configure:5637: --- MCA component pmix:external (m4 
>>> configuration 
>>>             macro) 
>>>             configure:128523: checking for MCA component pmix:external 
>>> compile 
>>>             mode 
>>>             configure:128529: result: dso 
>>>             configure:129054: checking if MCA component pmix:external can 
>>> compile 
>>>             configure:129056: result: no 
>>>             ... 
>>>             config.status:3897: creating opal/mca/pmix/Makefile 
>>>             config.status:3897: creating opal/mca/pmix/s1/Makefile 
>>>             config.status:3897: creating opal/mca/pmix/cray/Makefile 
>>>             config.status:3897: creating opal/mca/pmix/s2/Makefile 
>>>             config.status:3897: creating opal/mca/pmix/pmix112/Makefile 
>>>             config.status:3897: creating opal/mca/pmix/external/Makefile 
>>>             ... 
>>>             MCA_BUILD_opal_pmix_cray_DSO_FALSE='#' 
>>>             MCA_BUILD_opal_pmix_cray_DSO_TRUE='' 
>>>             MCA_BUILD_opal_pmix_external_DSO_FALSE='#' 
>>>             MCA_BUILD_opal_pmix_external_DSO_TRUE='' 
>>>             MCA_BUILD_opal_pmix_pmix112_DSO_FALSE='#' 
>>>             MCA_BUILD_opal_pmix_pmix112_DSO_TRUE='' 
>>>             MCA_BUILD_opal_pmix_s1_DSO_FALSE='#' 
>>>             MCA_BUILD_opal_pmix_s1_DSO_TRUE='' 
>>>             MCA_BUILD_opal_pmix_s2_DSO_FALSE='#' 
>>>             MCA_BUILD_opal_pmix_s2_DSO_TRUE='' 
>>>             ... 
>>>             MCA_opal_FRAMEWORKS='common  allocator backtrace btl dl event 
>>> hwloc if 
>>>             installdirs memchecker memcpy memory mpool pmix pstat rcache 
>>> sec 
>>>             shmem timer' 
>>>             MCA_opal_FRAMEWORKS_SUBDIRS='mca/common  mca/allocator 
>>>             mca/backtrace mca/btl 
>>>             mca/dl mca/event mca/hwloc mca/if mca/installdirs 
>>> mca/memchecker 
>>>             mca/memcpy 
>>>             mca/memory mca/mpool mca/pmix mca/pstat mca/rcache mca/sec 
>>>             mca/shmem mca/timer' 
>>>             
>>> MCA_opal_FRAMEWORK_COMPONENT_ALL_SUBDIRS='$(MCA_opal_common_ALL_SUBDIRS) 
>>>             $(MCA_opal_allocator_ALL_SUBDIRS) 
>>> $(MCA_opal_backtrace_ALL_SUBDIRS) 
>>>             $(MCA_opal_btl_ALL_SUBDIRS) $(MCA_opal_dl_ALL_SUBDIRS) 
>>>             $(MCA_opal_event_ALL_SUBDIRS) $(MCA_opal_hwloc_ALL_SUBDIRS) 
>>>             $(MCA_opal_if_ALL_SUBDIRS) $(MCA_opal_installdirs_ALL_SUBDIRS) 
>>>             $(MCA_opal_memchecker_ALL_SUBDIRS) 
>>> $(MCA_opal_memcpy_ALL_SUBDIRS) 
>>>             $(MCA_opal_memory_ALL_SUBDIRS) $(MCA_opal_mpool_ALL_SUBDIRS) 
>>>             $(MCA_opal_pmix_ALL_SUBDIRS) $(MCA_opal_pstat_ALL_SUBDIRS) 
>>>             $(MCA_opal_rcache_ALL_SUBDIRS) $(MCA_opal_sec_ALL_SUBDIRS) 
>>>             $(MCA_opal_shmem_ALL_SUBDIRS) $(MCA_opal_timer_ALL_SUBDIRS)' 
>>>             
>>> MCA_opal_FRAMEWORK_COMPONENT_DSO_SUBDIRS='$(MCA_opal_common_DSO_SUBDIRS) 
>>>             $(MCA_opal_allocator_DSO_SUBDIRS) 
>>> $(MCA_opal_backtrace_DSO_SUBDIRS) 
>>>             $(MCA_opal_btl_DSO_SUBDIRS) $(MCA_opal_dl_DSO_SUBDIRS) 
>>>             $(MCA_opal_event_DSO_SUBDIRS) $(MCA_opal_hwloc_DSO_SUBDIRS) 
>>>             $(MCA_opal_if_DSO_SUBDIRS) $(MCA_opal_installdirs_DSO_SUBDIRS) 
>>>             $(MCA_opal_memchecker_DSO_SUBDIRS) 
>>> $(MCA_opal_memcpy_DSO_SUBDIRS) 
>>>             $(MCA_opal_memory_DSO_SUBDIRS) $(MCA_opal_mpool_DSO_SUBDIRS) 
>>>             $(MCA_opal_pmix_DSO_SUBDIRS) $(MCA_opal_pstat_DSO_SUBDIRS) 
>>>             $(MCA_opal_rcache_DSO_SUBDIRS) $(MCA_opal_sec_DSO_SUBDIRS) 
>>>             $(MCA_opal_shmem_DSO_SUBDIRS) $(MCA_opal_timer_DSO_SUBDIRS)' 
>>>             
>>> MCA_opal_FRAMEWORK_COMPONENT_STATIC_SUBDIRS='$(MCA_opal_common_STATIC_SUBDIRS)
>>>              $(MCA_opal_allocator_STATIC_SUBDIRS) 
>>>             $(MCA_opal_backtrace_STATIC_SUBDIRS) 
>>>             $(MCA_opal_btl_STATIC_SUBDIRS) $(MCA_opal_dl_STATIC_SUBDIRS) 
>>>             $(MCA_opal_event_STATIC_SUBDIRS) 
>>> $(MCA_opal_hwloc_STATIC_SUBDIRS) 
>>>             $(MCA_opal_if_STATIC_SUBDIRS) 
>>> $(MCA_opal_installdirs_STATIC_SUBDIRS) 
>>>             $(MCA_opal_memchecker_STATIC_SUBDIRS) 
>>>             $(MCA_opal_memcpy_STATIC_SUBDIRS) 
>>>             $(MCA_opal_memory_STATIC_SUBDIRS) 
>>> $(MCA_opal_mpool_STATIC_SUBDIRS) 
>>>             $(MCA_opal_pmix_STATIC_SUBDIRS) 
>>> $(MCA_opal_pstat_STATIC_SUBDIRS) 
>>>             $(MCA_opal_rcache_STATIC_SUBDIRS) 
>>> $(MCA_opal_sec_STATIC_SUBDIRS) 
>>>             $(MCA_opal_shmem_STATIC_SUBDIRS) 
>>> $(MCA_opal_timer_STATIC_SUBDIRS)' 
>>>             MCA_opal_FRAMEWORK_LIBS=' $(MCA_opal_common_STATIC_LTLIBS) 
>>>             mca/allocator/libmca_allocator.la <http://libmca_allocator.la/> 
>>> <http://libmca_allocator.la> <http://libmca_allocator.la/> 
>>>             $(MCA_opal_allocator_STATIC_LTLIBS) 
>>>             mca/backtrace/libmca_backtrace.la <http://libmca_backtrace.la/> 
>>> <http://libmca_backtrace.la> <http://libmca_backtrace.la/> 
>>>             $(MCA_opal_backtrace_STATIC_LTLIBS) 
>>>             mca/btl/libmca_btl.la <http://libmca_btl.la/> 
>>> <http://libmca_btl.la> <http://libmca_btl.la/> 
>>>             $(MCA_opal_btl_STATIC_LTLIBS) mca/dl/libmca_dl.la 
>>> <http://libmca_dl.la/> 
>>>             <http://libmca_dl.la> <http://libmca_dl.la/> 
>>>             $(MCA_opal_dl_STATIC_LTLIBS) mca/event/libmca_event.la 
>>> <http://libmca_event.la/> 
>>>             <http://libmca_event.la> <http://libmca_event.la/> 
>>>             $(MCA_opal_event_STATIC_LTLIBS) mca/hwloc/libmca_hwloc.la 
>>> <http://libmca_hwloc.la/> 
>>>             <http://libmca_hwloc.la> <http://libmca_hwloc.la/> 
>>>             $(MCA_opal_hwloc_STATIC_LTLIBS) mca/if/libmca_if.la 
>>> <http://libmca_if.la/> 
>>>             <http://libmca_if.la> <http://libmca_if.la/> 
>>>             $(MCA_opal_if_STATIC_LTLIBS) 
>>> mca/installdirs/libmca_installdirs.la <http://libmca_installdirs.la/> 
>>>             <http://libmca_installdirs.la> <http://libmca_installdirs.la/> 
>>>             $(MCA_opal_installdirs_STATIC_LTLIBS) 
>>>             mca/memchecker/libmca_memchecker.la 
>>> <http://libmca_memchecker.la/> <http://libmca_memchecker.la> 
>>> <http://libmca_memchecker.la/> 
>>>             $(MCA_opal_memchecker_STATIC_LTLIBS) 
>>> mca/memcpy/libmca_memcpy.la <http://libmca_memcpy.la/> 
>>>             <http://libmca_memcpy.la> <http://libmca_memcpy.la/> 
>>>             $(MCA_opal_memcpy_STATIC_LTLIBS) mca/memory/libmca_memory.la 
>>> <http://libmca_memory.la/> 
>>>             <http://libmca_memory.la> <http://libmca_memory.la/> 
>>>             $(MCA_opal_memory_STATIC_LTLIBS) mca/mpool/libmca_mpool.la 
>>> <http://libmca_mpool.la/> 
>>>             <http://libmca_mpool.la> <http://libmca_mpool.la/> 
>>>             $(MCA_opal_mpool_STATIC_LTLIBS) mca/pmix/libmca_pmix.la 
>>> <http://libmca_pmix.la/> 
>>>             <http://libmca_pmix.la> <http://libmca_pmix.la/> 
>>>             $(MCA_opal_pmix_STATIC_LTLIBS) mca/pstat/libmca_pstat.la 
>>> <http://libmca_pstat.la/> 
>>>             <http://libmca_pstat.la> <http://libmca_pstat.la/> 
>>>             $(MCA_opal_pstat_STATIC_LTLIBS) mca/rcache/libmca_rcache.la 
>>> <http://libmca_rcache.la/> 
>>>             <http://libmca_rcache.la> <http://libmca_rcache.la/> 
>>>             $(MCA_opal_rcache_STATIC_LTLIBS) mca/sec/libmca_sec.la 
>>> <http://libmca_sec.la/> 
>>>             <http://libmca_sec.la> <http://libmca_sec.la/> 
>>>             $(MCA_opal_sec_STATIC_LTLIBS) mca/shmem/libmca_shmem.la 
>>> <http://libmca_shmem.la/> 
>>>             <http://libmca_shmem.la> <http://libmca_shmem.la/> 
>>>             $(MCA_opal_shmem_STATIC_LTLIBS) mca/timer/libmca_timer.la 
>>> <http://libmca_timer.la/> 
>>>             <http://libmca_timer.la> <http://libmca_timer.la/> 
>>>             $(MCA_opal_timer_STATIC_LTLIBS)' 
>>>             ... 
>>>             MCA_opal_pmix_ALL_COMPONENTS=' s1 cray s2 pmix112 external' 
>>>             MCA_opal_pmix_ALL_SUBDIRS=' mca/pmix/s1 mca/pmix/cray 
>>> mca/pmix/s2 
>>>             mca/pmix/pmix112 mca/pmix/external' 
>>>             MCA_opal_pmix_DSO_COMPONENTS=' pmix112' 
>>>             MCA_opal_pmix_DSO_SUBDIRS=' mca/pmix/pmix112' 
>>>             MCA_opal_pmix_STATIC_COMPONENTS='' 
>>>             MCA_opal_pmix_STATIC_LTLIBS='' 
>>>             MCA_opal_pmix_STATIC_SUBDIRS='' 
>>>             ... 
>>>             opal_pmix_ext_CPPFLAGS='' 
>>>             opal_pmix_ext_LDFLAGS='' 
>>>             opal_pmix_ext_LIBS='' 
>>>             
>>> opal_pmix_pmix112_CPPFLAGS='-I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/include/pmix
>>>             -I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/include 
>>>             -I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix 
>>>             -I$(OPAL_TOP_SRCDIR)/opal/mca/pmix/pmix112/pmix' 
>>>             
>>> opal_pmix_pmix112_LIBS='$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/libpmix.la
>>>  <http://libpmix.la/>
>>>             <http://libpmix.la> <http://libpmix.la/>' 
>>> 
>>>             ... 
>>> 
>>> 
>>> 
>>>             "config.log" for Sun C 5.13 shows the following. 
>>> 
>>>             ... 
>>>             configure:127803: /bin/bash 
>>>             
>>> '../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/ 
>>>             pmix/configure' *failed* for opal/mca/pmix/pmix112/pmix 
>>>             configure:128379: checking if MCA component pmix:pmix112 can 
>>> compile 
>>>             configure:128381: result: no 
>>>             configure:5637: --- MCA component pmix:external (m4 
>>> configuration 
>>>             macro) 
>>>             configure:128523: checking for MCA component pmix:external 
>>> compile 
>>>             mode 
>>>             configure:128529: result: dso 
>>>             configure:129054: checking if MCA component pmix:external can 
>>> compile 
>>>             configure:129056: result: no 
>>>             ... 
>>>             config.status:3887: creating opal/mca/pmix/Makefile 
>>>             config.status:3887: creating opal/mca/pmix/s1/Makefile 
>>>             config.status:3887: creating opal/mca/pmix/cray/Makefile 
>>>             config.status:3887: creating opal/mca/pmix/s2/Makefile 
>>>             config.status:3887: creating opal/mca/pmix/pmix112/Makefile 
>>>             config.status:3887: creating opal/mca/pmix/external/Makefile 
>>>             ... 
>>>             MCA_BUILD_opal_pmix_cray_DSO_FALSE='#' 
>>>             MCA_BUILD_opal_pmix_cray_DSO_TRUE='' 
>>>             MCA_BUILD_opal_pmix_external_DSO_FALSE='#' 
>>>             MCA_BUILD_opal_pmix_external_DSO_TRUE='' 
>>>             MCA_BUILD_opal_pmix_pmix112_DSO_FALSE='#' 
>>>             MCA_BUILD_opal_pmix_pmix112_DSO_TRUE='' 
>>>             MCA_BUILD_opal_pmix_s1_DSO_FALSE='#' 
>>>             MCA_BUILD_opal_pmix_s1_DSO_TRUE='' 
>>>             MCA_BUILD_opal_pmix_s2_DSO_FALSE='#' 
>>>             MCA_BUILD_opal_pmix_s2_DSO_TRUE='' 
>>>             ... 
>>>             MCA_opal_FRAMEWORKS='common  allocator backtrace btl dl event 
>>> hwloc if 
>>>             installdirs memchecker memcpy memory mpool pmix pstat rcache 
>>> sec 
>>>             shmem timer' 
>>>             MCA_opal_FRAMEWORKS_SUBDIRS='mca/common  mca/allocator 
>>>             mca/backtrace mca/btl 
>>>             mca/dl mca/event mca/hwloc mca/if mca/installdirs 
>>> mca/memchecker 
>>>             mca/memcpy 
>>>             mca/memory mca/mpool mca/pmix mca/pstat mca/rcache mca/sec 
>>>             mca/shmem mca/timer' 
>>>             
>>> MCA_opal_FRAMEWORK_COMPONENT_ALL_SUBDIRS='$(MCA_opal_common_ALL_SUBDIRS) 
>>>             $(MCA_opal_allocator_ALL_SUBDIRS) 
>>> $(MCA_opal_backtrace_ALL_SUBDIRS) 
>>>             $(MCA_opal_btl_ALL_SUBDIRS) $(MCA_opal_dl_ALL_SUBDIRS) 
>>>             $(MCA_opal_event_ALL_SUBDIRS) $(MCA_opal_hwloc_ALL_SUBDIRS) 
>>>             $(MCA_opal_if_ALL_SUBDIRS) $(MCA_opal_installdirs_ALL_SUBDIRS) 
>>>             $(MCA_opal_memchecker_ALL_SUBDIRS) 
>>> $(MCA_opal_memcpy_ALL_SUBDIRS) 
>>>             $(MCA_opal_memory_ALL_SUBDIRS) $(MCA_opal_mpool_ALL_SUBDIRS) 
>>>             $(MCA_opal_pmix_ALL_SUBDIRS) $(MCA_opal_pstat_ALL_SUBDIRS) 
>>>             $(MCA_opal_rcache_ALL_SUBDIRS) $(MCA_opal_sec_ALL_SUBDIRS) 
>>>             $(MCA_opal_shmem_ALL_SUBDIRS) $(MCA_opal_timer_ALL_SUBDIRS)' 
>>>             
>>> MCA_opal_FRAMEWORK_COMPONENT_DSO_SUBDIRS='$(MCA_opal_common_DSO_SUBDIRS) 
>>>             $(MCA_opal_allocator_DSO_SUBDIRS) 
>>> $(MCA_opal_backtrace_DSO_SUBDIRS) 
>>>             $(MCA_opal_btl_DSO_SUBDIRS) $(MCA_opal_dl_DSO_SUBDIRS) 
>>>             $(MCA_opal_event_DSO_SUBDIRS) $(MCA_opal_hwloc_DSO_SUBDIRS) 
>>>             $(MCA_opal_if_DSO_SUBDIRS) $(MCA_opal_installdirs_DSO_SUBDIRS) 
>>>             $(MCA_opal_memchecker_DSO_SUBDIRS) 
>>> $(MCA_opal_memcpy_DSO_SUBDIRS) 
>>>             $(MCA_opal_memory_DSO_SUBDIRS) $(MCA_opal_mpool_DSO_SUBDIRS) 
>>>             $(MCA_opal_pmix_DSO_SUBDIRS) $(MCA_opal_pstat_DSO_SUBDIRS) 
>>>             $(MCA_opal_rcache_DSO_SUBDIRS) $(MCA_opal_sec_DSO_SUBDIRS) 
>>>             $(MCA_opal_shmem_DSO_SUBDIRS) $(MCA_opal_timer_DSO_SUBDIRS)' 
>>>             
>>> MCA_opal_FRAMEWORK_COMPONENT_STATIC_SUBDIRS='$(MCA_opal_common_STATIC_SUBDIRS)
>>>              $(MCA_opal_allocator_STATIC_SUBDIRS) 
>>>             $(MCA_opal_backtrace_STATIC_SUBDIRS) 
>>>             $(MCA_opal_btl_STATIC_SUBDIRS) $(MCA_opal_dl_STATIC_SUBDIRS) 
>>>             $(MCA_opal_event_STATIC_SUBDIRS) 
>>> $(MCA_opal_hwloc_STATIC_SUBDIRS) 
>>>             $(MCA_opal_if_STATIC_SUBDIRS) 
>>> $(MCA_opal_installdirs_STATIC_SUBDIRS) 
>>>             $(MCA_opal_memchecker_STATIC_SUBDIRS) 
>>>             $(MCA_opal_memcpy_STATIC_SUBDIRS) 
>>>             $(MCA_opal_memory_STATIC_SUBDIRS) 
>>> $(MCA_opal_mpool_STATIC_SUBDIRS) 
>>>             $(MCA_opal_pmix_STATIC_SUBDIRS) 
>>> $(MCA_opal_pstat_STATIC_SUBDIRS) 
>>>             $(MCA_opal_rcache_STATIC_SUBDIRS) 
>>> $(MCA_opal_sec_STATIC_SUBDIRS) 
>>>             $(MCA_opal_shmem_STATIC_SUBDIRS) 
>>> $(MCA_opal_timer_STATIC_SUBDIRS)' 
>>>             MCA_opal_FRAMEWORK_LIBS=' $(MCA_opal_common_STATIC_LTLIBS) 
>>>             mca/allocator/libmca_allocator.la <http://libmca_allocator.la/> 
>>> <http://libmca_allocator.la> <http://libmca_allocator.la/> 
>>>             $(MCA_opal_allocator_STATIC_LTLIBS) 
>>>             mca/backtrace/libmca_backtrace.la <http://libmca_backtrace.la/> 
>>> <http://libmca_backtrace.la> <http://libmca_backtrace.la/> 
>>>             $(MCA_opal_backtrace_STATIC_LTLIBS) 
>>>             mca/btl/libmca_btl.la <http://libmca_btl.la/> 
>>> <http://libmca_btl.la> <http://libmca_btl.la/> 
>>>             $(MCA_opal_btl_STATIC_LTLIBS) mca/dl/libmca_dl.la 
>>> <http://libmca_dl.la/> 
>>>             <http://libmca_dl.la> <http://libmca_dl.la/> 
>>>             $(MCA_opal_dl_STATIC_LTLIBS) mca/event/libmca_event.la 
>>> <http://libmca_event.la/> 
>>>             <http://libmca_event.la> <http://libmca_event.la/> 
>>>             $(MCA_opal_event_STATIC_LTLIBS) mca/hwloc/libmca_hwloc.la 
>>> <http://libmca_hwloc.la/> 
>>>             <http://libmca_hwloc.la> <http://libmca_hwloc.la/> 
>>>             $(MCA_opal_hwloc_STATIC_LTLIBS) mca/if/libmca_if.la 
>>> <http://libmca_if.la/> 
>>>             <http://libmca_if.la> <http://libmca_if.la/> 
>>>             $(MCA_opal_if_STATIC_LTLIBS) 
>>> mca/installdirs/libmca_installdirs.la <http://libmca_installdirs.la/> 
>>>             <http://libmca_installdirs.la> <http://libmca_installdirs.la/> 
>>>             $(MCA_opal_installdirs_STATIC_LTLIBS) 
>>>             mca/memchecker/libmca_memchecker.la 
>>> <http://libmca_memchecker.la/> <http://libmca_memchecker.la> 
>>> <http://libmca_memchecker.la/> 
>>>             $(MCA_opal_memchecker_STATIC_LTLIBS) 
>>> mca/memcpy/libmca_memcpy.la <http://libmca_memcpy.la/> 
>>>             <http://libmca_memcpy.la> <http://libmca_memcpy.la/> 
>>>             $(MCA_opal_memcpy_STATIC_LTLIBS) mca/memory/libmca_memory.la 
>>> <http://libmca_memory.la/> 
>>>             <http://libmca_memory.la> <http://libmca_memory.la/> 
>>>             $(MCA_opal_memory_STATIC_LTLIBS) mca/mpool/libmca_mpool.la 
>>> <http://libmca_mpool.la/> 
>>>             <http://libmca_mpool.la> <http://libmca_mpool.la/> 
>>>             $(MCA_opal_mpool_STATIC_LTLIBS) mca/pmix/libmca_pmix.la 
>>> <http://libmca_pmix.la/> 
>>>             <http://libmca_pmix.la> <http://libmca_pmix.la/> 
>>>             $(MCA_opal_pmix_STATIC_LTLIBS) mca/pstat/libmca_pstat.la 
>>> <http://libmca_pstat.la/> 
>>>             <http://libmca_pstat.la> <http://libmca_pstat.la/> 
>>>             $(MCA_opal_pstat_STATIC_LTLIBS) mca/rcache/libmca_rcache.la 
>>> <http://libmca_rcache.la/> 
>>>             <http://libmca_rcache.la> <http://libmca_rcache.la/> 
>>>             $(MCA_opal_rcache_STATIC_LTLIBS) mca/sec/libmca_sec.la 
>>> <http://libmca_sec.la/> 
>>>             <http://libmca_sec.la> <http://libmca_sec.la/> 
>>>             $(MCA_opal_sec_STATIC_LTLIBS) mca/shmem/libmca_shmem.la 
>>> <http://libmca_shmem.la/> 
>>>             <http://libmca_shmem.la> <http://libmca_shmem.la/> 
>>>             $(MCA_opal_shmem_STATIC_LTLIBS) mca/timer/libmca_timer.la 
>>> <http://libmca_timer.la/> 
>>>             <http://libmca_timer.la> <http://libmca_timer.la/> 
>>>             $(MCA_opal_timer_STATIC_LTLIBS)' 
>>>             ... 
>>>             MCA_opal_pmix_ALL_COMPONENTS=' s1 cray s2 pmix112 external' 
>>>             MCA_opal_pmix_ALL_SUBDIRS=' mca/pmix/s1 mca/pmix/cray 
>>> mca/pmix/s2 
>>>             mca/pmix/pmix112 mca/pmix/external' 
>>>             MCA_opal_pmix_DSO_COMPONENTS='' 
>>>             MCA_opal_pmix_DSO_SUBDIRS='' 
>>>             MCA_opal_pmix_STATIC_COMPONENTS='' 
>>>             MCA_opal_pmix_STATIC_LTLIBS='' 
>>>             MCA_opal_pmix_STATIC_SUBDIRS='' 
>>>             ... 
>>>             opal_pmix_ext_CPPFLAGS='' 
>>>             opal_pmix_ext_LDFLAGS='' 
>>>             opal_pmix_ext_LIBS='' 
>>>             opal_pmix_pmix112_CPPFLAGS='' 
>>>             opal_pmix_pmix112_LIBS='' 
>>>             ... 
>>> 
>>> 
>>> 
>>> 
>>>             I've attached the config.log files for pmix. 
>>> 
>>>             tyr openmpi-2.0.0 142 tar zvft pmix_config.log.tar.gz 
>>>             -rw-r--r-- root/root    136291 2016-04-25 08:05:34 
>>>             
>>> openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/pmix/pmix112/pmix/config.log
>>>             -rw-r--r-- root/root    528808 2016-04-25 08:07:54 
>>>             
>>> openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_gcc/opal/mca/pmix/pmix112/pmix/config.log
>>>             tyr openmpi-2.0.0 143 
>>> 
>>> 
>>> 
>>>             I've also attached the output for the broken execution of 
>>>             "spawn_multiple_master" for my gcc-version of Open MPI. 
>>>             "spawn_master" works as expected with my gcc-version of Open 
>>> MPI. 
>>> 
>>>             Hopefully you can fix the problem. 
>>> 
>>> 
>>>             Kind regards and thank you very much for your help 
>>> 
>>>             Siegmar 
>>> 
>>> 
>>> 
>>>             Am 23.04.2016 um 21:34 schrieb Siegmar Gross: 
>>> 
>>>                 Hi Gilles, 
>>> 
>>>                 I don't know what happened, but the files are not available 
>>> now 
>>>                 and they were definitely available when I answered the 
>>> email from 
>>>                 Ralph. The files also have a different timestamp now. This 
>>> is an 
>>>                 extract from my email to Ralph for Solaris Sparc. 
>>> 
>>>                 -rwxr-xr-x 1 root root     977 Apr 19 19:49 mca_plm_rsh.la 
>>> <http://mca_plm_rsh.la/> 
>>>                 <http://mca_plm_rsh.la> <http://mca_plm_rsh.la/> 
>>>                 -rwxr-xr-x 1 root root  153280 Apr 19 19:49 mca_plm_rsh.so 
>>>                 -rwxr-xr-x 1 root root    1007 Apr 19 19:47 
>>>                 mca_pmix_pmix112.la <http://mca_pmix_pmix112.la/> 
>>> <http://mca_pmix_pmix112.la> <http://mca_pmix_pmix112.la/> 
>>>                 -rwxr-xr-x 1 root root 1400512 Apr 19 19:47 
>>> mca_pmix_pmix112.so 
>>>                 -rwxr-xr-x 1 root root     971 Apr 19 19:52 mca_pml_cm.la 
>>> <http://mca_pml_cm.la/> 
>>>                 <http://mca_pml_cm.la> <http://mca_pml_cm.la/> 
>>>                 -rwxr-xr-x 1 root root  342440 Apr 19 19:52 mca_pml_cm.so 
>>> 
>>>                 Now I have the following output for these files. 
>>> 
>>>                 -rwxr-xr-x 1 root root     976 Apr 19 19:58 mca_plm_rsh.la 
>>> <http://mca_plm_rsh.la/> 
>>>                 <http://mca_plm_rsh.la> <http://mca_plm_rsh.la/> 
>>>                 -rwxr-xr-x 1 root root  319816 Apr 19 19:58 mca_plm_rsh.so 
>>>                 -rwxr-xr-x 1 root root     970 Apr 19 20:00 mca_pml_cm.la 
>>> <http://mca_pml_cm.la/> 
>>>                 <http://mca_pml_cm.la> <http://mca_pml_cm.la/> 
>>>                 -rwxr-xr-x 1 root root 1507440 Apr 19 20:00 mca_pml_cm.so 
>>> 
>>>                 I'll try to find out what happened next week when I'm back 
>>> in 
>>>                 my office. 
>>> 
>>> 
>>>                 Kind regards 
>>> 
>>>                 Siegmar 
>>> 
>>> 
>>> 
>>> 
>>> 
>>>                 Am 23.04.16 um 02:12 schrieb Gilles Gouaillardet: 
>>> 
>>>                     Siegmar, 
>>> 
>>>                     I will try to reproduce this on my solaris11 x86_64 vm 
>>> 
>>>                     In the mean time, can you please double check 
>>>                     mca_pmix_pmix_pmix112.so 
>>>                     is a 64 bits library ? 
>>>                     (E.g, confirm "-m64" was correctly passed to pmix) 
>>> 
>>>                     Cheers, 
>>> 
>>>                     Gilles 
>>> 
>>>                     On Friday, April 22, 2016, Siegmar Gross 
>>>                     <siegmar.gr...@informatik.hs-fulda.de <> 
>>>                     <mailto:siegmar.gr...@informatik.hs-fulda.de> <>> 
>>> wrote: 
>>> 
>>>                         Hi Ralph, 
>>> 
>>>                         I've already used "-enable-debug". "SYSTEM_ENV" is 
>>>                     "SunOS" or 
>>>                         "Linux" and "MACHINE_ENV" is "sparc" or "x86_84". 
>>> 
>>>                         mkdir 
>>>                     
>>> openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_gcc 
>>>                         cd 
>>>                     
>>> openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_gcc 
>>> 
>>>                         ../openmpi-v2.x-dev-1280-gc110ae8/configure \ 
>>>                           --prefix=/usr/local/openmpi-2.0.0_64_gcc \ 
>>>                           --libdir=/usr/local/openmpi-2.0.0_64_gcc/lib64 \ 
>>>                           --with-jdk-bindir=/usr/local/jdk1.8.0/bin \ 
>>>                           --with-jdk-headers=/usr/local/jdk1.8.0/include \ 
>>>                           JAVA_HOME=/usr/local/jdk1.8.0 \ 
>>>                           LDFLAGS="-m64" CC="gcc" CXX="g++" FC="gfortran" \ 
>>>                           CFLAGS="-m64" CXXFLAGS="-m64" FCFLAGS="-m64" \ 
>>>                           CPP="cpp" CXXCPP="cpp" \ 
>>>                           --enable-mpi-cxx \ 
>>>                           --enable-cxx-exceptions \ 
>>>                           --enable-mpi-java \ 
>>>                           --enable-heterogeneous \ 
>>>                           --enable-mpi-thread-multiple \ 
>>>                           --with-hwloc=internal \ 
>>>                           --without-verbs \ 
>>>                           --with-wrapper-cflags="-std=c11 -m64" \ 
>>>                           --with-wrapper-cxxflags="-m64" \ 
>>>                           --with-wrapper-fcflags="-m64" \ 
>>>                           --enable-debug \ 
>>>                           |& tee 
>>> log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_gcc 
>>> 
>>> 
>>>                         mkdir 
>>>                     
>>> openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_cc 
>>>                         cd 
>>>                     
>>> openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_cc 
>>> 
>>>                         ../openmpi-v2.x-dev-1280-gc110ae8/configure \ 
>>>                           --prefix=/usr/local/openmpi-2.0.0_64_cc \ 
>>>                           --libdir=/usr/local/openmpi-2.0.0_64_cc/lib64 \ 
>>>                           --with-jdk-bindir=/usr/local/jdk1.8.0/bin \ 
>>>                           --with-jdk-headers=/usr/local/jdk1.8.0/include \ 
>>>                           JAVA_HOME=/usr/local/jdk1.8.0 \ 
>>>                           LDFLAGS="-m64" CC="cc" CXX="CC" FC="f95" \ 
>>>                           CFLAGS="-m64" CXXFLAGS="-m64 -library=stlport4" 
>>>                     FCFLAGS="-m64" \ 
>>>                           CPP="cpp" CXXCPP="cpp" \ 
>>>                           --enable-mpi-cxx \ 
>>>                           --enable-cxx-exceptions \ 
>>>                           --enable-mpi-java \ 
>>>                           --enable-heterogeneous \ 
>>>                           --enable-mpi-thread-multiple \ 
>>>                           --with-hwloc=internal \ 
>>>                           --without-verbs \ 
>>>                           --with-wrapper-cflags="-m64" \ 
>>>                           --with-wrapper-cxxflags="-m64 -library=stlport4" 
>>> \ 
>>>                           --with-wrapper-fcflags="-m64" \ 
>>>                           --with-wrapper-ldflags="" \ 
>>>                           --enable-debug \ 
>>>                           |& tee 
>>> log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_cc 
>>> 
>>> 
>>>                         Kind regards 
>>> 
>>>                         Siegmar 
>>> 
>>>                         Am 21.04.2016 um 18:18 schrieb Ralph Castain: 
>>> 
>>>                             Can you please rebuild OMPI with -enable-debug 
>>> in 
>>>                     the configure 
>>>                             cmd? It will let us see more error output 
>>> 
>>> 
>>>                                 On Apr 21, 2016, at 8:52 AM, Siegmar Gross 
>>>                                 <siegmar.gr...@informatik.hs-fulda.de> <> 
>>> wrote: 
>>> 
>>>                                 Hi Ralph, 
>>> 
>>>                                 I don't see any additional information. 
>>> 
>>>                                 tyr hello_1 108 mpiexec -np 4 --host 
>>>                                 tyr,sunpc1,linpc1,ruester -mca 
>>>                                 mca_base_component_show_load_errors 1 
>>> hello_1_mpi 
>>>                                 [tyr.informatik.hs-fulda.de:06211 
>>> <http://tyr.informatik.hs-fulda.de:06211/> 
>>>                     <http://tyr.informatik.hs-fulda.de:06211> 
>>> <http://tyr.informatik.hs-fulda.de:06211/> 
>>>                                 <http://tyr.informatik.hs-fulda.de:06211> 
>>> <http://tyr.informatik.hs-fulda.de:06211/>] 
>>>                     [[48741,0],0] 
>>>                                 ORTE_ERROR_LOG: Not found in file 
>>> 
>>>                     
>>> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c
>>> 
>>>                                 at line 638 
>>> 
>>>                     
>>> --------------------------------------------------------------------------
>>>                                 It looks like orte_init failed for some 
>>>                     reason; your 
>>>                                 parallel process is 
>>>                                 likely to abort.  There are many reasons 
>>> that 
>>>                     a parallel 
>>>                                 process can 
>>>                                 fail during orte_init; some of which are 
>>> due 
>>>                     to configuration or 
>>>                                 environment problems.  This failure appears 
>>> to 
>>>                     be an 
>>>                                 internal failure; 
>>>                                 here's some additional information (which 
>>> may 
>>>                     only be 
>>>                                 relevant to an 
>>>                                 Open MPI developer): 
>>> 
>>>                                  opal_pmix_base_select failed 
>>>                                  --> Returned value Not found (-13) instead 
>>> of 
>>>                     ORTE_SUCCESS 
>>> 
>>>                     
>>> --------------------------------------------------------------------------
>>> 
>>> 
>>>                                 tyr hello_1 109 mpiexec -np 4 --host 
>>>                                 tyr,sunpc1,linpc1,ruester -mca 
>>>                                 mca_base_component_show_load_errors 1 -mca 
>>>                     pmix_base_verbose 
>>>                                 10 -mca pmix_server_verbose 5 hello_1_mpi 
>>>                                 [tyr.informatik.hs-fulda.de:06212 
>>> <http://tyr.informatik.hs-fulda.de:06212/> 
>>>                     <http://tyr.informatik.hs-fulda.de:06212> 
>>> <http://tyr.informatik.hs-fulda.de:06212/> 
>>>                                 <http://tyr.informatik.hs-fulda.de:06212> 
>>> <http://tyr.informatik.hs-fulda.de:06212/>] 
>>>                     mca: base: 
>>>                                 components_register: registering framework 
>>>                     pmix components 
>>>                                 [tyr.informatik.hs-fulda.de:06212 
>>> <http://tyr.informatik.hs-fulda.de:06212/> 
>>>                     <http://tyr.informatik.hs-fulda.de:06212> 
>>> <http://tyr.informatik.hs-fulda.de:06212/> 
>>>                                 <http://tyr.informatik.hs-fulda.de:06212> 
>>> <http://tyr.informatik.hs-fulda.de:06212/>] 
>>>                     mca: base: 
>>>                                 components_open: opening pmix components 
>>>                                 [tyr.informatik.hs-fulda.de:06212 
>>> <http://tyr.informatik.hs-fulda.de:06212/> 
>>>                     <http://tyr.informatik.hs-fulda.de:06212> 
>>> <http://tyr.informatik.hs-fulda.de:06212/> 
>>>                                 <http://tyr.informatik.hs-fulda.de:06212> 
>>> <http://tyr.informatik.hs-fulda.de:06212/>] 
>>>                     mca:base:select: 
>>>                                 Auto-selecting pmix components 
>>>                                 [tyr.informatik.hs-fulda.de:06212 
>>> <http://tyr.informatik.hs-fulda.de:06212/> 
>>>                     <http://tyr.informatik.hs-fulda.de:06212> 
>>> <http://tyr.informatik.hs-fulda.de:06212/> 
>>>                                 <http://tyr.informatik.hs-fulda.de:06212> 
>>> <http://tyr.informatik.hs-fulda.de:06212/>] 
>>>                     mca:base:select:( 
>>>                                 pmix) No component selected! 
>>>                                 [tyr.informatik.hs-fulda.de:06212 
>>> <http://tyr.informatik.hs-fulda.de:06212/> 
>>>                     <http://tyr.informatik.hs-fulda.de:06212> 
>>> <http://tyr.informatik.hs-fulda.de:06212/> 
>>>                                 <http://tyr.informatik.hs-fulda.de:06212> 
>>> <http://tyr.informatik.hs-fulda.de:06212/>] 
>>>                     [[48738,0],0] 
>>>                                 ORTE_ERROR_LOG: Not found in file 
>>> 
>>>                     
>>> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c
>>> 
>>>                                 at line 638 
>>> 
>>>                     
>>> --------------------------------------------------------------------------
>>>                                 It looks like orte_init failed for some 
>>>                     reason; your 
>>>                                 parallel process is 
>>>                                 likely to abort.  There are many reasons 
>>> that 
>>>                     a parallel 
>>>                                 process can 
>>>                                 fail during orte_init; some of which are 
>>> due 
>>>                     to configuration or 
>>>                                 environment problems.  This failure appears 
>>> to 
>>>                     be an 
>>>                                 internal failure; 
>>>                                 here's some additional information (which 
>>> may 
>>>                     only be 
>>>                                 relevant to an 
>>>                                 Open MPI developer): 
>>> 
>>>                                  opal_pmix_base_select failed 
>>>                                  --> Returned value Not found (-13) instead 
>>> of 
>>>                     ORTE_SUCCESS 
>>> 
>>>                     
>>> --------------------------------------------------------------------------
>>>                                 tyr hello_1 110 
>>> 
>>> 
>>>                                 Kind regards 
>>> 
>>>                                 Siegmar 
>>> 
>>> 
>>>                                 Am 21.04.2016 um 17:24 schrieb Ralph 
>>> Castain: 
>>> 
>>>                                     Hmmm…it looks like you built the right 
>>>                     components, but 
>>>                                     they are not being picked up. Can you 
>>> run 
>>>                     your mpiexec 
>>>                                     command again, adding “-mca 
>>>                                     mca_base_component_show_load_errors 1” 
>>> to 
>>>                     the cmd line? 
>>> 
>>> 
>>>                                         On Apr 21, 2016, at 8:16 AM, 
>>> Siegmar Gross 
>>>                                         
>>> <siegmar.gr...@informatik.hs-fulda.de> <> 
>>>                     wrote: 
>>> 
>>>                                         Hi Ralph, 
>>> 
>>>                                         I have attached ompi_info output 
>>> for 
>>>                     both compilers 
>>>                                         from my 
>>>                                         sparc machine and the listings for 
>>>                     both compilers 
>>>                                         from the 
>>>                                         <prefix>/lib/openmpi directories. 
>>>                     Hopefully that 
>>>                                         helps to 
>>>                                         find the problem. 
>>> 
>>>                                         hermes tmp 3 tar zvft 
>>>                     openmpi-2.x_info.tar.gz 
>>>                                         -rw-r--r-- root/root     10969 
>>>                     2016-04-21 17:06 
>>>                                         ompi_info_SunOS_sparc_cc.txt 
>>>                                         -rw-r--r-- root/root     11044 
>>>                     2016-04-21 17:06 
>>>                                         ompi_info_SunOS_sparc_gcc.txt 
>>>                                         -rw-r--r-- root/root     71252 
>>>                     2016-04-21 17:02 
>>>                                         lib64_openmpi.txt 
>>>                                         hermes tmp 4 
>>> 
>>> 
>>>                                         Kind regards and thank you very 
>>> much 
>>>                     once more for 
>>>                                         your help 
>>> 
>>>                                         Siegmar 
>>> 
>>> 
>>>                                         Am 21.04.2016 um 15:54 schrieb 
>>> Ralph 
>>>                     Castain: 
>>> 
>>>                                             Odd - it would appear that none 
>>> of 
>>>                     the pmix 
>>>                                             components built? Can you send 
>>>                                             along the output from 
>>> ompi_info? 
>>>                     Or just send a 
>>>                                             listing of the files in the 
>>>                                             <prefix>/lib/openmpi directory? 
>>> 
>>> 
>>>                                                 On Apr 21, 2016, at 1:27 
>>> AM, 
>>>                     Siegmar Gross 
>>> 
>>>                     <siegmar.gr...@informatik.hs-fulda.de <> 
>>> 
>>>                     <mailto:siegmar.gr...@informatik.hs-fulda.de> <>> 
>>>                                                 wrote: 
>>> 
>>>                                                 Hi Ralph, 
>>> 
>>>                                                 Am 21.04.2016 um 00:18 
>>> schrieb 
>>>                     Ralph Castain: 
>>> 
>>>                                                     Could you please rerun 
>>>                     these test and 
>>>                                                     add “-mca 
>>> pmix_base_verbose 10 
>>>                                                     -mca 
>>> pmix_server_verbose 
>>>                     5” to your cmd 
>>>                                                     line? I need to see why 
>>> the 
>>>                                                     pmix components failed. 
>>> 
>>> 
>>> 
>>>                                                 tyr spawn 111 mpiexec -np 1 
>>> --host 
>>>                                                 tyr,sunpc1,linpc1,ruester 
>>> -mca 
>>>                                                 pmix_base_verbose 10 -mca 
>>>                                                 pmix_server_verbose 5 
>>>                     spawn_multiple_master 
>>>                                                 [tyr.informatik.hs-fulda.de 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de/> 
>>> <http://tyr.informatik.hs-fulda.de/>:26652] mca: 
>>>                                                 base: components_register: 
>>>                     registering 
>>>                                                 framework pmix components 
>>>                                                 [tyr.informatik.hs-fulda.de 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de/> 
>>> <http://tyr.informatik.hs-fulda.de/>:26652] mca: 
>>>                                                 base: components_open: 
>>> opening 
>>>                     pmix components 
>>>                                                 [tyr.informatik.hs-fulda.de 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de/> 
>>> <http://tyr.informatik.hs-fulda.de/>:26652] 
>>>                                                 mca:base:select: 
>>>                     Auto-selecting pmix components 
>>>                                                 [tyr.informatik.hs-fulda.de 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de/> 
>>> <http://tyr.informatik.hs-fulda.de/>:26652] 
>>>                                                 mca:base:select:( pmix) No 
>>>                     component selected! 
>>>                                                 [tyr.informatik.hs-fulda.de 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de/> 
>>> <http://tyr.informatik.hs-fulda.de/>:26652] 
>>>                                                 [[52794,0],0] 
>>> ORTE_ERROR_LOG: 
>>>                     Not found in file 
>>> 
>>>                     
>>> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c
>>> 
>>>                                                 at line 638 
>>> 
>>>                     
>>> --------------------------------------------------------------------------
>>>                                                 It looks like orte_init 
>>> failed 
>>>                     for some 
>>>                                                 reason; your parallel 
>>> process is 
>>>                                                 likely to abort.  There are 
>>>                     many reasons 
>>>                                                 that a parallel process can 
>>>                                                 fail during orte_init; some 
>>> of 
>>>                     which are due 
>>>                                                 to configuration or 
>>>                                                 environment problems.  This 
>>>                     failure appears 
>>>                                                 to be an internal failure; 
>>>                                                 here's some additional 
>>>                     information (which 
>>>                                                 may only be relevant to an 
>>>                                                 Open MPI developer): 
>>> 
>>>                                                 opal_pmix_base_select 
>>> failed 
>>>                                                 --> Returned value Not 
>>> found 
>>>                     (-13) instead 
>>>                                                 of ORTE_SUCCESS 
>>> 
>>>                     
>>> --------------------------------------------------------------------------
>>>                                                 tyr spawn 112 
>>> 
>>> 
>>> 
>>> 
>>>                                                 tyr hello_1 116 mpiexec -np 
>>> 1 
>>>                     --host 
>>>                                                 tyr,sunpc1,linpc1,ruester 
>>> -mca 
>>>                                                 pmix_base_verbose 10 -mca 
>>>                                                 pmix_server_verbose 5 
>>> hello_1_mpi 
>>>                                                 [tyr.informatik.hs-fulda.de 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de/> 
>>> <http://tyr.informatik.hs-fulda.de/>:27261] mca: 
>>>                                                 base: components_register: 
>>>                     registering 
>>>                                                 framework pmix components 
>>>                                                 [tyr.informatik.hs-fulda.de 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de/> 
>>> <http://tyr.informatik.hs-fulda.de/>:27261] mca: 
>>>                                                 base: components_open: 
>>> opening 
>>>                     pmix components 
>>>                                                 [tyr.informatik.hs-fulda.de 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de/> 
>>> <http://tyr.informatik.hs-fulda.de/>:27261] 
>>>                                                 mca:base:select: 
>>>                     Auto-selecting pmix components 
>>>                                                 [tyr.informatik.hs-fulda.de 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de/> 
>>> <http://tyr.informatik.hs-fulda.de/>:27261] 
>>>                                                 mca:base:select:( pmix) No 
>>>                     component selected! 
>>>                                                 [tyr.informatik.hs-fulda.de 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de/> 
>>> <http://tyr.informatik.hs-fulda.de/>:27261] 
>>>                                                 [[52315,0],0] 
>>> ORTE_ERROR_LOG: 
>>>                     Not found in file 
>>> 
>>>                     
>>> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c
>>> 
>>>                                                 at line 638 
>>> 
>>>                     
>>> --------------------------------------------------------------------------
>>>                                                 It looks like orte_init 
>>> failed 
>>>                     for some 
>>>                                                 reason; your parallel 
>>> process is 
>>>                                                 likely to abort.  There are 
>>>                     many reasons 
>>>                                                 that a parallel process can 
>>>                                                 fail during orte_init; some 
>>> of 
>>>                     which are due 
>>>                                                 to configuration or 
>>>                                                 environment problems.  This 
>>>                     failure appears 
>>>                                                 to be an internal failure; 
>>>                                                 here's some additional 
>>>                     information (which 
>>>                                                 may only be relevant to an 
>>>                                                 Open MPI developer): 
>>> 
>>>                                                 opal_pmix_base_select 
>>> failed 
>>>                                                 --> Returned value Not 
>>> found 
>>>                     (-13) instead 
>>>                                                 of ORTE_SUCCESS 
>>> 
>>>                     
>>> --------------------------------------------------------------------------
>>>                                                 tyr hello_1 117 
>>> 
>>> 
>>> 
>>>                                                 Thank you very much for 
>>> your help. 
>>> 
>>> 
>>>                                                 Kind regards 
>>> 
>>>                                                 Siegmar 
>>> 
>>> 
>>> 
>>> 
>>>                                                     Thanks 
>>>                                                     Ralph 
>>> 
>>>                                                         On Apr 20, 2016, at 
>>>                     10:12 AM, 
>>>                                                         Siegmar Gross 
>>> 
>>>                     <siegmar.gr...@informatik.hs-fulda.de <> 
>>> 
>>>                     <mailto:siegmar.gr...@informatik.hs-fulda.de> <>> 
>>>                                                         wrote: 
>>> 
>>>                                                         Hi, 
>>> 
>>>                                                         I have built 
>>> 
>>>                     openmpi-v2.x-dev-1280-gc110ae8 on my 
>>>                                                         machines 
>>>                                                         (Solaris 10 Sparc, 
>>>                     Solaris 10 
>>>                                                         x86_64, and 
>>> openSUSE Linux 
>>>                                                         12.1 x86_64) with 
>>>                     gcc-5.1.0 and Sun 
>>>                                                         C 5.13. 
>>> Unfortunately 
>>>                     I get 
>>>                                                         runtime errors for 
>>>                     some programs. 
>>> 
>>> 
>>>                                                         Sun C 5.13: 
>>>                                                         =========== 
>>> 
>>>                                                         For all my test 
>>>                     programs I get the 
>>>                                                         same error on 
>>> Solaris 
>>>                     Sparc and 
>>>                                                         Solaris x86_64, 
>>> while 
>>>                     the programs 
>>>                                                         work fine on Linux. 
>>> 
>>>                                                         tyr hello_1 115 
>>>                     mpiexec -np 2 
>>>                                                         hello_1_mpi 
>>> 
>>>                     [tyr.informatik.hs-fulda.de 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/>:22373] 
>>>                                                         [[61763,0],0] 
>>>                     ORTE_ERROR_LOG: Not 
>>>                                                         found in file 
>>> 
>>>                     
>>> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c
>>> 
>>>                                                         at line 638 
>>> 
>>>                     
>>> --------------------------------------------------------------------------
>>>                                                         It looks like 
>>>                     orte_init failed for 
>>>                                                         some reason; your 
>>>                     parallel process is 
>>>                                                         likely to abort. 
>>>                     There are many 
>>>                                                         reasons that a 
>>>                     parallel process can 
>>>                                                         fail during 
>>> orte_init; 
>>>                     some of which 
>>>                                                         are due to 
>>>                     configuration or 
>>>                                                         environment 
>>> problems. 
>>>                     This failure 
>>>                                                         appears to be an 
>>>                     internal failure; 
>>>                                                         here's some 
>>> additional 
>>>                     information 
>>>                                                         (which may only be 
>>>                     relevant to an 
>>>                                                         Open MPI 
>>> developer): 
>>> 
>>>                                                         
>>> opal_pmix_base_select 
>>>                     failed 
>>>                                                         --> Returned value 
>>> Not 
>>>                     found (-13) 
>>>                                                         instead of 
>>> ORTE_SUCCESS 
>>> 
>>>                     
>>> --------------------------------------------------------------------------
>>>                                                         tyr hello_1 116 
>>> 
>>> 
>>> 
>>> 
>>>                                                         GCC-5.1.0: 
>>>                                                         ========== 
>>> 
>>>                                                         tyr spawn 121 
>>> mpiexec 
>>>                     -np 1 --host 
>>>                                                         
>>> tyr,sunpc1,linpc1,ruester 
>>>                                                         
>>> spawn_multiple_master 
>>> 
>>>                                                         Parent process 0 
>>>                     running on 
>>> 
>>>                     tyr.informatik.hs-fulda.de 
>>> <http://tyr.informatik.hs-fulda.de/> <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>>                                                         I create 3 slave 
>>>                     processes. 
>>> 
>>> 
>>>                     [tyr.informatik.hs-fulda.de 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/>:25366] 
>>>                                                         PMIX ERROR: 
>>>                     UNPACK-PAST-END in file 
>>> 
>>>                     
>>> ../../../../../../openmpi-v2.x-dev-1280-gc110ae8/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c
>>> 
>>> 
>>>                                                         at line 829 
>>> 
>>>                     [tyr.informatik.hs-fulda.de 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/> 
>>> 
>>>                     <http://tyr.informatik.hs-fulda.de> 
>>> <http://tyr.informatik.hs-fulda.de/>:25366] 
>>>                                                         PMIX ERROR: 
>>>                     UNPACK-PAST-END in file 
>>> 
>>>                     
>>> ../../../../../../openmpi-v2.x-dev-1280-gc110ae8/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c
>>> 
>>> 
>>>                                                         at line 2176 
>>>                                                         [tyr:25377] *** An 
>>>                     error occurred in 
>>>                                                         
>>> MPI_Comm_spawn_multiple 
>>>                                                         [tyr:25377] *** 
>>>                     reported by process 
>>>                                                         [3308257281,0] 
>>>                                                         [tyr:25377] *** on 
>>>                     communicator 
>>>                                                         MPI_COMM_WORLD 
>>>                                                         [tyr:25377] *** 
>>>                     MPI_ERR_SPAWN: could 
>>>                                                         not spawn processes 
>>>                                                         [tyr:25377] *** 
>>>                     MPI_ERRORS_ARE_FATAL 
>>>                                                         (processes in this 
>>>                     communicator will 
>>>                                                         now abort, 
>>>                                                         [tyr:25377] ***    
>>> and 
>>>                     potentially 
>>>                                                         your MPI job) 
>>>                                                         tyr spawn 122 
>>> 
>>> 
>>>                                                         I would be grateful 
>>> if 
>>>                     somebody can 
>>>                                                         fix the problems. 
>>>                     Thank you very 
>>>                                                         much for any help 
>>> in 
>>>                     advance. 
>>> 
>>> 
>>>                                                         Kind regards 
>>> 
>>>                                                         Siegmar 
>>> 
>>>                     
>>> <hello_1_mpi.c><spawn_multiple_master.c>_______________________________________________
>>> 
>>> 
>>>                                                         users mailing list 
>>>                                                         us...@open-mpi.org 
>>> <> 
>>>                     <mailto:us...@open-mpi.org> <> 
>>>                                                         Subscription: 
>>> 
>>>                     http://www.open-mpi.org/mailman/listinfo.cgi/users 
>>> <http://www.open-mpi.org/mailman/listinfo.cgi/users> 
>>>                                                         Link to this post: 
>>> 
>>>                     
>>> <http://www.open-mpi.org/community/lists/users/2016/04/28983.php>_______________________________________________
> users mailing list
> us...@open-mpi.org <mailto:us...@open-mpi.org>
> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users 
> <https://www.open-mpi.org/mailman/listinfo.cgi/users>
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2016/04/29047.php 
> <http://www.open-mpi.org/community/lists/users/2016/04/29047.php>

Reply via email to