Hmmm…well, I may have to wait and let Gilles fix this. So far as I can see, the code in the current OMPI 2.x tarball (and upstream) is correct:
int pmix_bfrop_pack_app(pmix_buffer_t *buffer, const void *src, int32_t num_vals, pmix_data_type_t type) { pmix_app_t *app; int32_t i, j, nvals; int ret; app = (pmix_app_t *) src; for (i = 0; i < num_vals; ++i) { if (PMIX_SUCCESS != (ret = pmix_bfrop_pack_string(buffer, &app[i].cmd, 1, PMIX_STRING))) { return ret; } /* argv */ if (PMIX_SUCCESS != (ret = pmix_bfrop_pack_int(buffer, &app[i].argc, 1, PMIX_INT))) { return ret; } for (j=0; j < app->argc; j++) { if (PMIX_SUCCESS != (ret = pmix_bfrop_pack_string(buffer, &app[i].argv[j], 1, PMIX_STRING))) { return ret; } } /* env */ nvals = pmix_argv_count(app[i].env); if (PMIX_SUCCESS != (ret = pmix_bfrop_pack_int32(buffer, &nvals, 1, PMIX_INT32))) { return ret; } for (j=0; j < nvals; j++) { if (PMIX_SUCCESS != (ret = pmix_bfrop_pack_string(buffer, &app[i].env[j], 1, PMIX_STRING))) { return ret; } } /* maxprocs */ if (PMIX_SUCCESS != (ret = pmix_bfrop_pack_int(buffer, &app[i].maxprocs, 1, PMIX_INT))) { return ret; } /* info array */ if (PMIX_SUCCESS != (ret = pmix_bfrop_pack_sizet(buffer, &app[i].ninfo, 1, PMIX_SIZE))) { return ret; } if (0 < app[i].ninfo) { if (PMIX_SUCCESS != (ret = pmix_bfrop_pack_info(buffer, app[i].info, app[i].ninfo, PMIX_INFO))) { return ret; } } } return PMIX_SUCCESS; } Siegmar: have you tried the latest release candidate? > On Apr 28, 2016, at 11:08 PM, Siegmar Gross > <siegmar.gr...@informatik.hs-fulda.de> wrote: > > Hi Gilles, > > thank you very much for identifying the reason for the problem > and fixing it. > > Have a nice weekend > > Siegmar > > Am 29.04.2016 um 03:38 schrieb Gilles Gouaillardet: >> Siegmar, >> >> in pmix_bfrop_pack_app, >> app->argc >> must be replaced with >> app[i].argc >> >> I will PR to pmix, ompi and ompi-release when I am back at work on Monday >> >> Cheers, >> >> Gilles >> >> On Thursday, April 28, 2016, Gilles Gouaillardet <gil...@rist.or.jp >> <mailto:gil...@rist.or.jp <mailto:gil...@rist.or.jp>>> wrote: >> >> Siegmar, >> >> >> can you please also post the source of spawn_slave ? >> >> >> Cheers, >> >> Gilles >> >> >> On 4/28/2016 1:17 AM, Siegmar Gross wrote: >>> Hi Gilles, >>> >>> it is not necessary to have a heterogeneous environment to reproduce >>> the error as you can see below. All machines are 64 bit. >>> >>> tyr spawn 119 ompi_info | grep -e "OPAL repo revision" -e "C compiler >>> absolute" >>> OPAL repo revision: v2.x-dev-1290-gbd0e4e1 >>> C compiler absolute: /usr/local/gcc-5.1.0/bin/gcc >>> tyr spawn 120 uname -a >>> SunOS tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/> >>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>> 5.10 Generic_150400-11 sun4u sparc SUNW,A70 Solaris >>> tyr spawn 121 mpiexec -np 1 --host tyr,tyr,tyr,tyr spawn_multiple_master >>> >>> Parent process 0 running on tyr.informatik.hs-fulda.de >>> <http://tyr.informatik.hs-fulda.de/> >>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>> I create 3 slave processes. >>> >>> [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>:27286 >>> <http://tyr.informatik.hs-fulda.de:27286 >>> <http://tyr.informatik.hs-fulda.de:27286/>>] PMIX ERROR: UNPACK-PAST-END >>> in file >>> >>> ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c >>> at line 829 >>> [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>:27286 >>> <http://tyr.informatik.hs-fulda.de:27286 >>> <http://tyr.informatik.hs-fulda.de:27286/>>] PMIX ERROR: UNPACK-PAST-END >>> in file >>> >>> ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c >>> at line 2176 >>> [tyr:27288] *** An error occurred in MPI_Comm_spawn_multiple >>> [tyr:27288] *** reported by process [3434086401,0] >>> [tyr:27288] *** on communicator MPI_COMM_WORLD >>> [tyr:27288] *** MPI_ERR_SPAWN: could not spawn processes >>> [tyr:27288] *** MPI_ERRORS_ARE_FATAL (processes in this communicator >>> will now abort, >>> [tyr:27288] *** and potentially your MPI job) >>> tyr spawn 122 >>> >>> >>> >>> >>> >>> >>> sunpc1 fd1026 105 ompi_info | grep -e "OPAL repo revision" -e "C >>> compiler absolute" >>> OPAL repo revision: v2.x-dev-1290-gbd0e4e1 >>> C compiler absolute: /usr/local/gcc-5.1.0/bin/gcc >>> sunpc1 fd1026 106 uname -a >>> SunOS sunpc1 5.10 Generic_147441-21 i86pc i386 i86pc Solaris >>> sunpc1 fd1026 107 mpiexec -np 1 --host sunpc1,sunpc1,sunpc1,sunpc1 >>> spawn_multiple_master >>> >>> Parent process 0 running on sunpc1 >>> I create 3 slave processes. >>> >>> [sunpc1:00368] PMIX ERROR: UNPACK-PAST-END in file >>> >>> ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c >>> at line 829 >>> [sunpc1:00368] PMIX ERROR: UNPACK-PAST-END in file >>> >>> ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c >>> at line 2176 >>> [sunpc1:370] *** An error occurred in MPI_Comm_spawn_multiple >>> [sunpc1:370] *** reported by process [43909121,0] >>> [sunpc1:370] *** on communicator MPI_COMM_WORLD >>> [sunpc1:370] *** MPI_ERR_SPAWN: could not spawn processes >>> [sunpc1:370] *** MPI_ERRORS_ARE_FATAL (processes in this communicator >>> will now abort, >>> [sunpc1:370] *** and potentially your MPI job) >>> sunpc1 fd1026 108 >>> >>> >>> >>> >>> >>> linpc1 fd1026 105 ompi_info | grep -e "OPAL repo revision" -e "C >>> compiler absolute" >>> OPAL repo revision: v2.x-dev-1290-gbd0e4e1 >>> C compiler absolute: /usr/local/gcc-5.1.0/bin/gcc >>> linpc1 fd1026 106 uname -a >>> Linux linpc1 3.1.10-1.29-desktop #1 SMP PREEMPT Fri May 31 20:10:04 UTC >>> 2013 (2529847) x86_64 x86_64 x86_64 GNU/Linux >>> linpc1 fd1026 107 mpiexec -np 1 --host linpc1,linpc1,linpc1,linpc1 >>> spawn_multiple_master >>> >>> Parent process 0 running on linpc1 >>> I create 3 slave processes. >>> >>> [linpc1:21502] PMIX ERROR: UNPACK-PAST-END in file >>> >>> ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c >>> at line 829 >>> [linpc1:21502] PMIX ERROR: UNPACK-PAST-END in file >>> >>> ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c >>> at line 2176 >>> [linpc1:21507] *** An error occurred in MPI_Comm_spawn_multiple >>> [linpc1:21507] *** reported by process [1005518849,0] >>> [linpc1:21507] *** on communicator MPI_COMM_WORLD >>> [linpc1:21507] *** MPI_ERR_SPAWN: could not spawn processes >>> [linpc1:21507] *** MPI_ERRORS_ARE_FATAL (processes in this communicator >>> will now abort, >>> [linpc1:21507] *** and potentially your MPI job) >>> linpc1 fd1026 108 >>> >>> >>> I used the following configure command. >>> >>> ../openmpi-v2.x-dev-1290-gbd0e4e1/configure \ >>> --prefix=/usr/local/openmpi-2.0.0_64_gcc \ >>> --libdir=/usr/local/openmpi-2.0.0_64_gcc/lib64 \ >>> --with-jdk-bindir=/usr/local/jdk1.8.0/bin \ >>> --with-jdk-headers=/usr/local/jdk1.8.0/include \ >>> JAVA_HOME=/usr/local/jdk1.8.0 \ >>> LDFLAGS="-m64" CC="gcc" CXX="g++" FC="gfortran" \ >>> CFLAGS="-m64" CXXFLAGS="-m64" FCFLAGS="-m64" \ >>> CPP="cpp" CXXCPP="cpp" \ >>> --enable-mpi-cxx \ >>> --enable-cxx-exceptions \ >>> --enable-mpi-java \ >>> --enable-heterogeneous \ >>> --enable-mpi-thread-multiple \ >>> --with-hwloc=internal \ >>> --without-verbs \ >>> --with-wrapper-cflags="-std=c11 -m64" \ >>> --with-wrapper-cxxflags="-m64" \ >>> --with-wrapper-fcflags="-m64" \ >>> --enable-debug \ >>> |& tee log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_gcc >>> >>> >>> Kind regards >>> >>> Siegmar >>> >>> >>> >>> Am 27.04.2016 um 13:21 schrieb Gilles Gouaillardet: >>>> Siegmar, >>>> >>>> please add this to your CFLAGS for the time being. >>>> >>>> configure tries to detect which flags must be added for C99 support, >>>> and it seems >>>> the test is not working for Solaris 10 and Oracle compilers. >>>> this is no more a widely used environment, and I am not sure I can find >>>> the >>>> time to fix this >>>> in a near future. >>>> >>>> >>>> regarding the runtime issue, can you please describe your 4 hosts (os, >>>> endianness and bitness) >>>> >>>> Cheers, >>>> >>>> Gilles >>>> >>>> On Wednesday, April 27, 2016, Siegmar Gross >>>> <siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de> >>>> <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de>');> >>>> <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de>');>> >>>> wrote: >>>> >>>> Hi Gilles, >>>> >>>> adding "-std=c99" to CFLAGS solves the problem with the missing >>>> library. >>>> Shall I add it permanently to my configure command or will you add >>>> it, >>>> so that I will not run into problems if you need the C11 standard >>>> later? >>>> >>>> "spawn_multiple_master" breaks with the same error that I reported >>>> yesterday for my gcc-version of Open MPI. Hopefully you can solve >>>> the >>>> problem as well. >>>> >>>> >>>> Kind regards and thank you very much for your help >>>> >>>> Siegmar >>>> >>>> >>>> Am 27.04.2016 um 08:05 schrieb Gilles Gouaillardet: >>>> >>>> Siegmar, >>>> >>>> >>>> here is the error : >>>> >>>> configure:17969: cc -o conftest -m64 -D_REENTRANT -g -g >>>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1 >>>> >>>> >>>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc >>>> >>>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/include >>>> >>>> >>>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/include >>>> -D_REENTRANT >>>> >>>> >>>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/hwloc/hwloc1112/hwloc/include >>>> >>>> >>>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/hwloc/hwloc1112/hwloc/include >>>> >>>> >>>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/event/libevent2022/libevent >>>> >>>> >>>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/event/libevent2022/libevent/include >>>> >>>> >>>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/event/libevent2022/libevent/include >>>> -m64 conftest.c >&5 >>>> "/usr/include/stdbool.h", line 42: #error: "Use of <stdbool.h> >>>> is >>>> valid only >>>> in a c99 compilation environment." >>>> >>>> >>>> i cannot reproduce this on solaris 11 with oracle studio 5.3 >>>> compiler, >>>> and i >>>> do not have solaris 10 yet. >>>> >>>> could you please re-configure with '-std=c99' appended to your >>>> CFLAGS >>>> and see >>>> if it helps ? >>>> >>>> >>>> Cheers, >>>> >>>> >>>> Gilles >>>> >>>> >>>> On 4/26/2016 7:57 PM, Siegmar Gross wrote: >>>> >>>> Hi Gilles and Ralph, >>>> >>>> I was able to sort out my mess. In my last email I compared >>>> the >>>> files from "SunOS_sparc/openmpi-2.0.0_64_gcc/lib64/openmpi" >>>> from >>>> the attachment of my email to Ralph with the files from >>>> "SunOS_sparc/openmpi-2.0.0_64_cc/lib64/openmpi" from my >>>> current >>>> file system. That's the reason while I have had different >>>> timestamps. The other problem was that Ralph didn't >>>> recognize >>>> that "mca_pmix_pmix112.so" wasn't built on Solaris with the >>>> Sun C compiler. I've removed most of the files from the >>>> attachment >>>> of my email so that it is easier to see the relevant files. >>>> Below >>>> I try to give you more information that may be relevant to >>>> track >>>> down the problem. I still get an error running one of my >>>> small >>>> test programs, when I use my gcc-version of Open MPI. >>>> "mca_pmix_pmix112.so" is a 64 bits library. >>>> >>>> Linux_x86_64/openmpi-2.0.0_64_cc/lib64/openmpi: >>>> ... >>>> -rwxr-xr-x 1 root root 261327 Apr 19 16:46 mca_plm_slurm.so >>>> -rwxr-xr-x 1 root root 1002 Apr 19 16:45 >>>> mca_pmix_pmix112.la <http://mca_pmix_pmix112.la >>>> <http://mca_pmix_pmix112.la/>> >>>> <http://mca_pmix_pmix112.la <http://mca_pmix_pmix112.la/>> >>>> <http://mca_pmix_pmix112.la <http://mca_pmix_pmix112.la/>> >>>> -rwxr-xr-x 1 root root 3906526 Apr 19 16:45 >>>> mca_pmix_pmix112.so >>>> -rwxr-xr-x 1 root root 966 Apr 19 16:51 mca_pml_cm.la >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> -rwxr-xr-x 1 root root 1574265 Apr 19 16:51 mca_pml_cm.so >>>> ... >>>> >>>> Linux_x86_64/openmpi-2.0.0_64_gcc/lib64/openmpi: >>>> ... >>>> -rwxr-xr-x 1 root root 70371 Apr 19 16:43 mca_plm_slurm.so >>>> -rwxr-xr-x 1 root root 1008 Apr 19 16:42 >>>> mca_pmix_pmix112.la <http://mca_pmix_pmix112.la >>>> <http://mca_pmix_pmix112.la/>> >>>> <http://mca_pmix_pmix112.la <http://mca_pmix_pmix112.la/>> >>>> <http://mca_pmix_pmix112.la <http://mca_pmix_pmix112.la/>> >>>> -rwxr-xr-x 1 root root 1029005 Apr 19 16:42 >>>> mca_pmix_pmix112.so >>>> -rwxr-xr-x 1 root root 972 Apr 19 16:46 mca_pml_cm.la >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> -rwxr-xr-x 1 root root 284858 Apr 19 16:46 mca_pml_cm.so >>>> ... >>>> >>>> SunOS_sparc/openmpi-2.0.0_64_cc/lib64/openmpi: >>>> ... >>>> -rwxr-xr-x 1 root root 319816 Apr 19 19:58 mca_plm_rsh.so >>>> -rwxr-xr-x 1 root root 970 Apr 19 20:00 mca_pml_cm.la >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> -rwxr-xr-x 1 root root 1507440 Apr 19 20:00 mca_pml_cm.so >>>> ... >>>> >>>> SunOS_sparc/openmpi-2.0.0_64_gcc/lib64/openmpi: >>>> ... >>>> -rwxr-xr-x 1 root root 153280 Apr 19 19:49 mca_plm_rsh.so >>>> -rwxr-xr-x 1 root root 1007 Apr 19 19:47 >>>> mca_pmix_pmix112.la <http://mca_pmix_pmix112.la >>>> <http://mca_pmix_pmix112.la/>> >>>> <http://mca_pmix_pmix112.la <http://mca_pmix_pmix112.la/>> >>>> <http://mca_pmix_pmix112.la <http://mca_pmix_pmix112.la/>> >>>> -rwxr-xr-x 1 root root 1400512 Apr 19 19:47 >>>> mca_pmix_pmix112.so >>>> -rwxr-xr-x 1 root root 971 Apr 19 19:52 mca_pml_cm.la >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> -rwxr-xr-x 1 root root 342440 Apr 19 19:52 mca_pml_cm.so >>>> ... >>>> >>>> SunOS_x86_64/openmpi-2.0.0_64_cc/lib64/openmpi: >>>> ... >>>> -rwxr-xr-x 1 root root 300096 Apr 19 17:18 mca_plm_rsh.so >>>> -rwxr-xr-x 1 root root 970 Apr 19 17:23 mca_pml_cm.la >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> -rwxr-xr-x 1 root root 1458816 Apr 19 17:23 mca_pml_cm.so >>>> ... >>>> >>>> SunOS_x86_64/openmpi-2.0.0_64_gcc/lib64/openmpi: >>>> ... >>>> -rwxr-xr-x 1 root root 133096 Apr 19 17:42 mca_plm_rsh.so >>>> -rwxr-xr-x 1 root root 1007 Apr 19 17:41 >>>> mca_pmix_pmix112.la <http://mca_pmix_pmix112.la >>>> <http://mca_pmix_pmix112.la/>> >>>> <http://mca_pmix_pmix112.la <http://mca_pmix_pmix112.la/>> >>>> <http://mca_pmix_pmix112.la <http://mca_pmix_pmix112.la/>> >>>> -rwxr-xr-x 1 root root 1320240 Apr 19 17:41 >>>> mca_pmix_pmix112.so >>>> -rwxr-xr-x 1 root root 971 Apr 19 17:46 mca_pml_cm.la >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> -rwxr-xr-x 1 root root 419848 Apr 19 17:46 mca_pml_cm.so >>>> ... >>>> >>>> >>>> Yesterday I've installed openmpi-v2.x-dev-1290-gbd0e4e1 so >>>> that we >>>> have a current version for the investigation of the >>>> problem. Once >>>> more mca_pmix_pmix112.so isn't available on Solaris if I >>>> use the >>>> Sun C compiler. >>>> >>>> "config.log" for gcc-5.1.0 shows the following. >>>> >>>> ... >>>> configure:127799: /bin/bash >>>> >>>> '../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/ >>>> pmix/configure' succeeded for opal/mca/pmix/pmix112/pmix >>>> configure:127916: checking if MCA component pmix:pmix112 >>>> can compile >>>> configure:127918: result: yes >>>> configure:5637: --- MCA component pmix:external (m4 >>>> configuration >>>> macro) >>>> configure:128523: checking for MCA component pmix:external >>>> compile >>>> mode >>>> configure:128529: result: dso >>>> configure:129054: checking if MCA component pmix:external >>>> can compile >>>> configure:129056: result: no >>>> ... >>>> config.status:3897: creating opal/mca/pmix/Makefile >>>> config.status:3897: creating opal/mca/pmix/s1/Makefile >>>> config.status:3897: creating opal/mca/pmix/cray/Makefile >>>> config.status:3897: creating opal/mca/pmix/s2/Makefile >>>> config.status:3897: creating opal/mca/pmix/pmix112/Makefile >>>> config.status:3897: creating opal/mca/pmix/external/Makefile >>>> ... >>>> MCA_BUILD_opal_pmix_cray_DSO_FALSE='#' >>>> MCA_BUILD_opal_pmix_cray_DSO_TRUE='' >>>> MCA_BUILD_opal_pmix_external_DSO_FALSE='#' >>>> MCA_BUILD_opal_pmix_external_DSO_TRUE='' >>>> MCA_BUILD_opal_pmix_pmix112_DSO_FALSE='#' >>>> MCA_BUILD_opal_pmix_pmix112_DSO_TRUE='' >>>> MCA_BUILD_opal_pmix_s1_DSO_FALSE='#' >>>> MCA_BUILD_opal_pmix_s1_DSO_TRUE='' >>>> MCA_BUILD_opal_pmix_s2_DSO_FALSE='#' >>>> MCA_BUILD_opal_pmix_s2_DSO_TRUE='' >>>> ... >>>> MCA_opal_FRAMEWORKS='common allocator backtrace btl dl >>>> event hwloc if >>>> installdirs memchecker memcpy memory mpool pmix pstat >>>> rcache sec >>>> shmem timer' >>>> MCA_opal_FRAMEWORKS_SUBDIRS='mca/common mca/allocator >>>> mca/backtrace mca/btl >>>> mca/dl mca/event mca/hwloc mca/if mca/installdirs >>>> mca/memchecker >>>> mca/memcpy >>>> mca/memory mca/mpool mca/pmix mca/pstat mca/rcache mca/sec >>>> mca/shmem mca/timer' >>>> >>>> MCA_opal_FRAMEWORK_COMPONENT_ALL_SUBDIRS='$(MCA_opal_common_ALL_SUBDIRS) >>>> $(MCA_opal_allocator_ALL_SUBDIRS) >>>> $(MCA_opal_backtrace_ALL_SUBDIRS) >>>> $(MCA_opal_btl_ALL_SUBDIRS) $(MCA_opal_dl_ALL_SUBDIRS) >>>> $(MCA_opal_event_ALL_SUBDIRS) $(MCA_opal_hwloc_ALL_SUBDIRS) >>>> $(MCA_opal_if_ALL_SUBDIRS) >>>> $(MCA_opal_installdirs_ALL_SUBDIRS) >>>> $(MCA_opal_memchecker_ALL_SUBDIRS) >>>> $(MCA_opal_memcpy_ALL_SUBDIRS) >>>> $(MCA_opal_memory_ALL_SUBDIRS) $(MCA_opal_mpool_ALL_SUBDIRS) >>>> $(MCA_opal_pmix_ALL_SUBDIRS) $(MCA_opal_pstat_ALL_SUBDIRS) >>>> $(MCA_opal_rcache_ALL_SUBDIRS) $(MCA_opal_sec_ALL_SUBDIRS) >>>> $(MCA_opal_shmem_ALL_SUBDIRS) $(MCA_opal_timer_ALL_SUBDIRS)' >>>> >>>> MCA_opal_FRAMEWORK_COMPONENT_DSO_SUBDIRS='$(MCA_opal_common_DSO_SUBDIRS) >>>> $(MCA_opal_allocator_DSO_SUBDIRS) >>>> $(MCA_opal_backtrace_DSO_SUBDIRS) >>>> $(MCA_opal_btl_DSO_SUBDIRS) $(MCA_opal_dl_DSO_SUBDIRS) >>>> $(MCA_opal_event_DSO_SUBDIRS) $(MCA_opal_hwloc_DSO_SUBDIRS) >>>> $(MCA_opal_if_DSO_SUBDIRS) >>>> $(MCA_opal_installdirs_DSO_SUBDIRS) >>>> $(MCA_opal_memchecker_DSO_SUBDIRS) >>>> $(MCA_opal_memcpy_DSO_SUBDIRS) >>>> $(MCA_opal_memory_DSO_SUBDIRS) $(MCA_opal_mpool_DSO_SUBDIRS) >>>> $(MCA_opal_pmix_DSO_SUBDIRS) $(MCA_opal_pstat_DSO_SUBDIRS) >>>> $(MCA_opal_rcache_DSO_SUBDIRS) $(MCA_opal_sec_DSO_SUBDIRS) >>>> $(MCA_opal_shmem_DSO_SUBDIRS) $(MCA_opal_timer_DSO_SUBDIRS)' >>>> >>>> >>>> MCA_opal_FRAMEWORK_COMPONENT_STATIC_SUBDIRS='$(MCA_opal_common_STATIC_SUBDIRS) >>>> $(MCA_opal_allocator_STATIC_SUBDIRS) >>>> $(MCA_opal_backtrace_STATIC_SUBDIRS) >>>> $(MCA_opal_btl_STATIC_SUBDIRS) $(MCA_opal_dl_STATIC_SUBDIRS) >>>> $(MCA_opal_event_STATIC_SUBDIRS) >>>> $(MCA_opal_hwloc_STATIC_SUBDIRS) >>>> $(MCA_opal_if_STATIC_SUBDIRS) >>>> $(MCA_opal_installdirs_STATIC_SUBDIRS) >>>> $(MCA_opal_memchecker_STATIC_SUBDIRS) >>>> $(MCA_opal_memcpy_STATIC_SUBDIRS) >>>> $(MCA_opal_memory_STATIC_SUBDIRS) >>>> $(MCA_opal_mpool_STATIC_SUBDIRS) >>>> $(MCA_opal_pmix_STATIC_SUBDIRS) >>>> $(MCA_opal_pstat_STATIC_SUBDIRS) >>>> $(MCA_opal_rcache_STATIC_SUBDIRS) >>>> $(MCA_opal_sec_STATIC_SUBDIRS) >>>> $(MCA_opal_shmem_STATIC_SUBDIRS) >>>> $(MCA_opal_timer_STATIC_SUBDIRS)' >>>> MCA_opal_FRAMEWORK_LIBS=' $(MCA_opal_common_STATIC_LTLIBS) >>>> mca/allocator/libmca_allocator.la >>>> <http://libmca_allocator.la <http://libmca_allocator.la/>> >>>> <http://libmca_allocator.la <http://libmca_allocator.la/>> >>>> <http://libmca_allocator.la <http://libmca_allocator.la/>> >>>> $(MCA_opal_allocator_STATIC_LTLIBS) >>>> mca/backtrace/libmca_backtrace.la >>>> <http://libmca_backtrace.la <http://libmca_backtrace.la/>> >>>> <http://libmca_backtrace.la <http://libmca_backtrace.la/>> >>>> <http://libmca_backtrace.la <http://libmca_backtrace.la/>> >>>> $(MCA_opal_backtrace_STATIC_LTLIBS) >>>> mca/btl/libmca_btl.la <http://libmca_btl.la >>>> <http://libmca_btl.la/>> >>>> <http://libmca_btl.la <http://libmca_btl.la/>> <http://libmca_btl.la >>>> <http://libmca_btl.la/>> >>>> $(MCA_opal_btl_STATIC_LTLIBS) mca/dl/libmca_dl.la >>>> <http://libmca_dl.la <http://libmca_dl.la/>> >>>> <http://libmca_dl.la <http://libmca_dl.la/>> >>>> <http://libmca_dl.la <http://libmca_dl.la/>> >>>> $(MCA_opal_dl_STATIC_LTLIBS) mca/event/libmca_event.la >>>> <http://libmca_event.la <http://libmca_event.la/>> >>>> <http://libmca_event.la <http://libmca_event.la/>> >>>> <http://libmca_event.la <http://libmca_event.la/>> >>>> $(MCA_opal_event_STATIC_LTLIBS) mca/hwloc/libmca_hwloc.la >>>> <http://libmca_hwloc.la <http://libmca_hwloc.la/>> >>>> <http://libmca_hwloc.la <http://libmca_hwloc.la/>> >>>> <http://libmca_hwloc.la <http://libmca_hwloc.la/>> >>>> $(MCA_opal_hwloc_STATIC_LTLIBS) mca/if/libmca_if.la >>>> <http://libmca_if.la <http://libmca_if.la/>> >>>> <http://libmca_if.la <http://libmca_if.la/>> >>>> <http://libmca_if.la <http://libmca_if.la/>> >>>> $(MCA_opal_if_STATIC_LTLIBS) >>>> mca/installdirs/libmca_installdirs.la <http://libmca_installdirs.la >>>> <http://libmca_installdirs.la/>> >>>> <http://libmca_installdirs.la >>>> <http://libmca_installdirs.la/>> <http://libmca_installdirs.la >>>> <http://libmca_installdirs.la/>> >>>> $(MCA_opal_installdirs_STATIC_LTLIBS) >>>> mca/memchecker/libmca_memchecker.la >>>> <http://libmca_memchecker.la <http://libmca_memchecker.la/>> >>>> <http://libmca_memchecker.la <http://libmca_memchecker.la/>> >>>> <http://libmca_memchecker.la <http://libmca_memchecker.la/>> >>>> $(MCA_opal_memchecker_STATIC_LTLIBS) >>>> mca/memcpy/libmca_memcpy.la <http://libmca_memcpy.la >>>> <http://libmca_memcpy.la/>> >>>> <http://libmca_memcpy.la <http://libmca_memcpy.la/>> >>>> <http://libmca_memcpy.la <http://libmca_memcpy.la/>> >>>> $(MCA_opal_memcpy_STATIC_LTLIBS) >>>> mca/memory/libmca_memory.la <http://libmca_memory.la >>>> <http://libmca_memory.la/>> >>>> <http://libmca_memory.la <http://libmca_memory.la/>> >>>> <http://libmca_memory.la <http://libmca_memory.la/>> >>>> $(MCA_opal_memory_STATIC_LTLIBS) mca/mpool/libmca_mpool.la >>>> <http://libmca_mpool.la <http://libmca_mpool.la/>> >>>> <http://libmca_mpool.la <http://libmca_mpool.la/>> >>>> <http://libmca_mpool.la <http://libmca_mpool.la/>> >>>> $(MCA_opal_mpool_STATIC_LTLIBS) mca/pmix/libmca_pmix.la >>>> <http://libmca_pmix.la <http://libmca_pmix.la/>> >>>> <http://libmca_pmix.la <http://libmca_pmix.la/>> >>>> <http://libmca_pmix.la <http://libmca_pmix.la/>> >>>> $(MCA_opal_pmix_STATIC_LTLIBS) mca/pstat/libmca_pstat.la >>>> <http://libmca_pstat.la <http://libmca_pstat.la/>> >>>> <http://libmca_pstat.la <http://libmca_pstat.la/>> >>>> <http://libmca_pstat.la <http://libmca_pstat.la/>> >>>> $(MCA_opal_pstat_STATIC_LTLIBS) mca/rcache/libmca_rcache.la >>>> <http://libmca_rcache.la <http://libmca_rcache.la/>> >>>> <http://libmca_rcache.la <http://libmca_rcache.la/>> >>>> <http://libmca_rcache.la <http://libmca_rcache.la/>> >>>> $(MCA_opal_rcache_STATIC_LTLIBS) mca/sec/libmca_sec.la >>>> <http://libmca_sec.la <http://libmca_sec.la/>> >>>> <http://libmca_sec.la <http://libmca_sec.la/>> >>>> <http://libmca_sec.la <http://libmca_sec.la/>> >>>> $(MCA_opal_sec_STATIC_LTLIBS) mca/shmem/libmca_shmem.la >>>> <http://libmca_shmem.la <http://libmca_shmem.la/>> >>>> <http://libmca_shmem.la <http://libmca_shmem.la/>> >>>> <http://libmca_shmem.la <http://libmca_shmem.la/>> >>>> $(MCA_opal_shmem_STATIC_LTLIBS) mca/timer/libmca_timer.la >>>> <http://libmca_timer.la <http://libmca_timer.la/>> >>>> <http://libmca_timer.la <http://libmca_timer.la/>> >>>> <http://libmca_timer.la <http://libmca_timer.la/>> >>>> $(MCA_opal_timer_STATIC_LTLIBS)' >>>> ... >>>> MCA_opal_pmix_ALL_COMPONENTS=' s1 cray s2 pmix112 external' >>>> MCA_opal_pmix_ALL_SUBDIRS=' mca/pmix/s1 mca/pmix/cray >>>> mca/pmix/s2 >>>> mca/pmix/pmix112 mca/pmix/external' >>>> MCA_opal_pmix_DSO_COMPONENTS=' pmix112' >>>> MCA_opal_pmix_DSO_SUBDIRS=' mca/pmix/pmix112' >>>> MCA_opal_pmix_STATIC_COMPONENTS='' >>>> MCA_opal_pmix_STATIC_LTLIBS='' >>>> MCA_opal_pmix_STATIC_SUBDIRS='' >>>> ... >>>> opal_pmix_ext_CPPFLAGS='' >>>> opal_pmix_ext_LDFLAGS='' >>>> opal_pmix_ext_LIBS='' >>>> >>>> >>>> opal_pmix_pmix112_CPPFLAGS='-I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/include/pmix >>>> -I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/include >>>> -I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix >>>> -I$(OPAL_TOP_SRCDIR)/opal/mca/pmix/pmix112/pmix' >>>> >>>> >>>> opal_pmix_pmix112_LIBS='$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/libpmix.la >>>> <http://libpmix.la <http://libpmix.la/>> >>>> <http://libpmix.la <http://libpmix.la/>> <http://libpmix.la >>>> <http://libpmix.la/>>' >>>> >>>> ... >>>> >>>> >>>> >>>> "config.log" for Sun C 5.13 shows the following. >>>> >>>> ... >>>> configure:127803: /bin/bash >>>> >>>> '../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/ >>>> pmix/configure' *failed* for opal/mca/pmix/pmix112/pmix >>>> configure:128379: checking if MCA component pmix:pmix112 >>>> can compile >>>> configure:128381: result: no >>>> configure:5637: --- MCA component pmix:external (m4 >>>> configuration >>>> macro) >>>> configure:128523: checking for MCA component pmix:external >>>> compile >>>> mode >>>> configure:128529: result: dso >>>> configure:129054: checking if MCA component pmix:external >>>> can compile >>>> configure:129056: result: no >>>> ... >>>> config.status:3887: creating opal/mca/pmix/Makefile >>>> config.status:3887: creating opal/mca/pmix/s1/Makefile >>>> config.status:3887: creating opal/mca/pmix/cray/Makefile >>>> config.status:3887: creating opal/mca/pmix/s2/Makefile >>>> config.status:3887: creating opal/mca/pmix/pmix112/Makefile >>>> config.status:3887: creating opal/mca/pmix/external/Makefile >>>> ... >>>> MCA_BUILD_opal_pmix_cray_DSO_FALSE='#' >>>> MCA_BUILD_opal_pmix_cray_DSO_TRUE='' >>>> MCA_BUILD_opal_pmix_external_DSO_FALSE='#' >>>> MCA_BUILD_opal_pmix_external_DSO_TRUE='' >>>> MCA_BUILD_opal_pmix_pmix112_DSO_FALSE='#' >>>> MCA_BUILD_opal_pmix_pmix112_DSO_TRUE='' >>>> MCA_BUILD_opal_pmix_s1_DSO_FALSE='#' >>>> MCA_BUILD_opal_pmix_s1_DSO_TRUE='' >>>> MCA_BUILD_opal_pmix_s2_DSO_FALSE='#' >>>> MCA_BUILD_opal_pmix_s2_DSO_TRUE='' >>>> ... >>>> MCA_opal_FRAMEWORKS='common allocator backtrace btl dl >>>> event hwloc if >>>> installdirs memchecker memcpy memory mpool pmix pstat >>>> rcache sec >>>> shmem timer' >>>> MCA_opal_FRAMEWORKS_SUBDIRS='mca/common mca/allocator >>>> mca/backtrace mca/btl >>>> mca/dl mca/event mca/hwloc mca/if mca/installdirs >>>> mca/memchecker >>>> mca/memcpy >>>> mca/memory mca/mpool mca/pmix mca/pstat mca/rcache mca/sec >>>> mca/shmem mca/timer' >>>> >>>> MCA_opal_FRAMEWORK_COMPONENT_ALL_SUBDIRS='$(MCA_opal_common_ALL_SUBDIRS) >>>> $(MCA_opal_allocator_ALL_SUBDIRS) >>>> $(MCA_opal_backtrace_ALL_SUBDIRS) >>>> $(MCA_opal_btl_ALL_SUBDIRS) $(MCA_opal_dl_ALL_SUBDIRS) >>>> $(MCA_opal_event_ALL_SUBDIRS) $(MCA_opal_hwloc_ALL_SUBDIRS) >>>> $(MCA_opal_if_ALL_SUBDIRS) >>>> $(MCA_opal_installdirs_ALL_SUBDIRS) >>>> $(MCA_opal_memchecker_ALL_SUBDIRS) >>>> $(MCA_opal_memcpy_ALL_SUBDIRS) >>>> $(MCA_opal_memory_ALL_SUBDIRS) $(MCA_opal_mpool_ALL_SUBDIRS) >>>> $(MCA_opal_pmix_ALL_SUBDIRS) $(MCA_opal_pstat_ALL_SUBDIRS) >>>> $(MCA_opal_rcache_ALL_SUBDIRS) $(MCA_opal_sec_ALL_SUBDIRS) >>>> $(MCA_opal_shmem_ALL_SUBDIRS) $(MCA_opal_timer_ALL_SUBDIRS)' >>>> >>>> MCA_opal_FRAMEWORK_COMPONENT_DSO_SUBDIRS='$(MCA_opal_common_DSO_SUBDIRS) >>>> $(MCA_opal_allocator_DSO_SUBDIRS) >>>> $(MCA_opal_backtrace_DSO_SUBDIRS) >>>> $(MCA_opal_btl_DSO_SUBDIRS) $(MCA_opal_dl_DSO_SUBDIRS) >>>> $(MCA_opal_event_DSO_SUBDIRS) $(MCA_opal_hwloc_DSO_SUBDIRS) >>>> $(MCA_opal_if_DSO_SUBDIRS) >>>> $(MCA_opal_installdirs_DSO_SUBDIRS) >>>> $(MCA_opal_memchecker_DSO_SUBDIRS) >>>> $(MCA_opal_memcpy_DSO_SUBDIRS) >>>> $(MCA_opal_memory_DSO_SUBDIRS) $(MCA_opal_mpool_DSO_SUBDIRS) >>>> $(MCA_opal_pmix_DSO_SUBDIRS) $(MCA_opal_pstat_DSO_SUBDIRS) >>>> $(MCA_opal_rcache_DSO_SUBDIRS) $(MCA_opal_sec_DSO_SUBDIRS) >>>> $(MCA_opal_shmem_DSO_SUBDIRS) $(MCA_opal_timer_DSO_SUBDIRS)' >>>> >>>> >>>> MCA_opal_FRAMEWORK_COMPONENT_STATIC_SUBDIRS='$(MCA_opal_common_STATIC_SUBDIRS) >>>> $(MCA_opal_allocator_STATIC_SUBDIRS) >>>> $(MCA_opal_backtrace_STATIC_SUBDIRS) >>>> $(MCA_opal_btl_STATIC_SUBDIRS) $(MCA_opal_dl_STATIC_SUBDIRS) >>>> $(MCA_opal_event_STATIC_SUBDIRS) >>>> $(MCA_opal_hwloc_STATIC_SUBDIRS) >>>> $(MCA_opal_if_STATIC_SUBDIRS) >>>> $(MCA_opal_installdirs_STATIC_SUBDIRS) >>>> $(MCA_opal_memchecker_STATIC_SUBDIRS) >>>> $(MCA_opal_memcpy_STATIC_SUBDIRS) >>>> $(MCA_opal_memory_STATIC_SUBDIRS) >>>> $(MCA_opal_mpool_STATIC_SUBDIRS) >>>> $(MCA_opal_pmix_STATIC_SUBDIRS) >>>> $(MCA_opal_pstat_STATIC_SUBDIRS) >>>> $(MCA_opal_rcache_STATIC_SUBDIRS) >>>> $(MCA_opal_sec_STATIC_SUBDIRS) >>>> $(MCA_opal_shmem_STATIC_SUBDIRS) >>>> $(MCA_opal_timer_STATIC_SUBDIRS)' >>>> MCA_opal_FRAMEWORK_LIBS=' $(MCA_opal_common_STATIC_LTLIBS) >>>> mca/allocator/libmca_allocator.la >>>> <http://libmca_allocator.la <http://libmca_allocator.la/>> >>>> <http://libmca_allocator.la <http://libmca_allocator.la/>> >>>> <http://libmca_allocator.la <http://libmca_allocator.la/>> >>>> $(MCA_opal_allocator_STATIC_LTLIBS) >>>> mca/backtrace/libmca_backtrace.la >>>> <http://libmca_backtrace.la <http://libmca_backtrace.la/>> >>>> <http://libmca_backtrace.la <http://libmca_backtrace.la/>> >>>> <http://libmca_backtrace.la <http://libmca_backtrace.la/>> >>>> $(MCA_opal_backtrace_STATIC_LTLIBS) >>>> mca/btl/libmca_btl.la <http://libmca_btl.la >>>> <http://libmca_btl.la/>> >>>> <http://libmca_btl.la <http://libmca_btl.la/>> <http://libmca_btl.la >>>> <http://libmca_btl.la/>> >>>> $(MCA_opal_btl_STATIC_LTLIBS) mca/dl/libmca_dl.la >>>> <http://libmca_dl.la <http://libmca_dl.la/>> >>>> <http://libmca_dl.la <http://libmca_dl.la/>> >>>> <http://libmca_dl.la <http://libmca_dl.la/>> >>>> $(MCA_opal_dl_STATIC_LTLIBS) mca/event/libmca_event.la >>>> <http://libmca_event.la <http://libmca_event.la/>> >>>> <http://libmca_event.la <http://libmca_event.la/>> >>>> <http://libmca_event.la <http://libmca_event.la/>> >>>> $(MCA_opal_event_STATIC_LTLIBS) mca/hwloc/libmca_hwloc.la >>>> <http://libmca_hwloc.la <http://libmca_hwloc.la/>> >>>> <http://libmca_hwloc.la <http://libmca_hwloc.la/>> >>>> <http://libmca_hwloc.la <http://libmca_hwloc.la/>> >>>> $(MCA_opal_hwloc_STATIC_LTLIBS) mca/if/libmca_if.la >>>> <http://libmca_if.la <http://libmca_if.la/>> >>>> <http://libmca_if.la <http://libmca_if.la/>> >>>> <http://libmca_if.la <http://libmca_if.la/>> >>>> $(MCA_opal_if_STATIC_LTLIBS) >>>> mca/installdirs/libmca_installdirs.la <http://libmca_installdirs.la >>>> <http://libmca_installdirs.la/>> >>>> <http://libmca_installdirs.la >>>> <http://libmca_installdirs.la/>> <http://libmca_installdirs.la >>>> <http://libmca_installdirs.la/>> >>>> $(MCA_opal_installdirs_STATIC_LTLIBS) >>>> mca/memchecker/libmca_memchecker.la >>>> <http://libmca_memchecker.la <http://libmca_memchecker.la/>> >>>> <http://libmca_memchecker.la <http://libmca_memchecker.la/>> >>>> <http://libmca_memchecker.la <http://libmca_memchecker.la/>> >>>> $(MCA_opal_memchecker_STATIC_LTLIBS) >>>> mca/memcpy/libmca_memcpy.la <http://libmca_memcpy.la >>>> <http://libmca_memcpy.la/>> >>>> <http://libmca_memcpy.la <http://libmca_memcpy.la/>> >>>> <http://libmca_memcpy.la <http://libmca_memcpy.la/>> >>>> $(MCA_opal_memcpy_STATIC_LTLIBS) >>>> mca/memory/libmca_memory.la <http://libmca_memory.la >>>> <http://libmca_memory.la/>> >>>> <http://libmca_memory.la <http://libmca_memory.la/>> >>>> <http://libmca_memory.la <http://libmca_memory.la/>> >>>> $(MCA_opal_memory_STATIC_LTLIBS) mca/mpool/libmca_mpool.la >>>> <http://libmca_mpool.la <http://libmca_mpool.la/>> >>>> <http://libmca_mpool.la <http://libmca_mpool.la/>> >>>> <http://libmca_mpool.la <http://libmca_mpool.la/>> >>>> $(MCA_opal_mpool_STATIC_LTLIBS) mca/pmix/libmca_pmix.la >>>> <http://libmca_pmix.la <http://libmca_pmix.la/>> >>>> <http://libmca_pmix.la <http://libmca_pmix.la/>> >>>> <http://libmca_pmix.la <http://libmca_pmix.la/>> >>>> $(MCA_opal_pmix_STATIC_LTLIBS) mca/pstat/libmca_pstat.la >>>> <http://libmca_pstat.la <http://libmca_pstat.la/>> >>>> <http://libmca_pstat.la <http://libmca_pstat.la/>> >>>> <http://libmca_pstat.la <http://libmca_pstat.la/>> >>>> $(MCA_opal_pstat_STATIC_LTLIBS) mca/rcache/libmca_rcache.la >>>> <http://libmca_rcache.la <http://libmca_rcache.la/>> >>>> <http://libmca_rcache.la <http://libmca_rcache.la/>> >>>> <http://libmca_rcache.la <http://libmca_rcache.la/>> >>>> $(MCA_opal_rcache_STATIC_LTLIBS) mca/sec/libmca_sec.la >>>> <http://libmca_sec.la <http://libmca_sec.la/>> >>>> <http://libmca_sec.la <http://libmca_sec.la/>> >>>> <http://libmca_sec.la <http://libmca_sec.la/>> >>>> $(MCA_opal_sec_STATIC_LTLIBS) mca/shmem/libmca_shmem.la >>>> <http://libmca_shmem.la <http://libmca_shmem.la/>> >>>> <http://libmca_shmem.la <http://libmca_shmem.la/>> >>>> <http://libmca_shmem.la <http://libmca_shmem.la/>> >>>> $(MCA_opal_shmem_STATIC_LTLIBS) mca/timer/libmca_timer.la >>>> <http://libmca_timer.la <http://libmca_timer.la/>> >>>> <http://libmca_timer.la <http://libmca_timer.la/>> >>>> <http://libmca_timer.la <http://libmca_timer.la/>> >>>> $(MCA_opal_timer_STATIC_LTLIBS)' >>>> ... >>>> MCA_opal_pmix_ALL_COMPONENTS=' s1 cray s2 pmix112 external' >>>> MCA_opal_pmix_ALL_SUBDIRS=' mca/pmix/s1 mca/pmix/cray >>>> mca/pmix/s2 >>>> mca/pmix/pmix112 mca/pmix/external' >>>> MCA_opal_pmix_DSO_COMPONENTS='' >>>> MCA_opal_pmix_DSO_SUBDIRS='' >>>> MCA_opal_pmix_STATIC_COMPONENTS='' >>>> MCA_opal_pmix_STATIC_LTLIBS='' >>>> MCA_opal_pmix_STATIC_SUBDIRS='' >>>> ... >>>> opal_pmix_ext_CPPFLAGS='' >>>> opal_pmix_ext_LDFLAGS='' >>>> opal_pmix_ext_LIBS='' >>>> opal_pmix_pmix112_CPPFLAGS='' >>>> opal_pmix_pmix112_LIBS='' >>>> ... >>>> >>>> >>>> >>>> >>>> I've attached the config.log files for pmix. >>>> >>>> tyr openmpi-2.0.0 142 tar zvft pmix_config.log.tar.gz >>>> -rw-r--r-- root/root 136291 2016-04-25 08:05:34 >>>> >>>> >>>> openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/pmix/pmix112/pmix/config.log >>>> -rw-r--r-- root/root 528808 2016-04-25 08:07:54 >>>> >>>> >>>> openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_gcc/opal/mca/pmix/pmix112/pmix/config.log >>>> tyr openmpi-2.0.0 143 >>>> >>>> >>>> >>>> I've also attached the output for the broken execution of >>>> "spawn_multiple_master" for my gcc-version of Open MPI. >>>> "spawn_master" works as expected with my gcc-version of >>>> Open MPI. >>>> >>>> Hopefully you can fix the problem. >>>> >>>> >>>> Kind regards and thank you very much for your help >>>> >>>> Siegmar >>>> >>>> >>>> >>>> Am 23.04.2016 um 21:34 schrieb Siegmar Gross: >>>> >>>> Hi Gilles, >>>> >>>> I don't know what happened, but the files are not >>>> available now >>>> and they were definitely available when I answered the >>>> email from >>>> Ralph. The files also have a different timestamp now. >>>> This is an >>>> extract from my email to Ralph for Solaris Sparc. >>>> >>>> -rwxr-xr-x 1 root root 977 Apr 19 19:49 >>>> mca_plm_rsh.la <http://mca_plm_rsh.la <http://mca_plm_rsh.la/>> >>>> <http://mca_plm_rsh.la <http://mca_plm_rsh.la/>> >>>> <http://mca_plm_rsh.la <http://mca_plm_rsh.la/>> >>>> -rwxr-xr-x 1 root root 153280 Apr 19 19:49 >>>> mca_plm_rsh.so >>>> -rwxr-xr-x 1 root root 1007 Apr 19 19:47 >>>> mca_pmix_pmix112.la <http://mca_pmix_pmix112.la >>>> <http://mca_pmix_pmix112.la/>> >>>> <http://mca_pmix_pmix112.la <http://mca_pmix_pmix112.la/>> >>>> <http://mca_pmix_pmix112.la <http://mca_pmix_pmix112.la/>> >>>> -rwxr-xr-x 1 root root 1400512 Apr 19 19:47 >>>> mca_pmix_pmix112.so >>>> -rwxr-xr-x 1 root root 971 Apr 19 19:52 >>>> mca_pml_cm.la <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> -rwxr-xr-x 1 root root 342440 Apr 19 19:52 >>>> mca_pml_cm.so >>>> >>>> Now I have the following output for these files. >>>> >>>> -rwxr-xr-x 1 root root 976 Apr 19 19:58 >>>> mca_plm_rsh.la <http://mca_plm_rsh.la <http://mca_plm_rsh.la/>> >>>> <http://mca_plm_rsh.la <http://mca_plm_rsh.la/>> >>>> <http://mca_plm_rsh.la <http://mca_plm_rsh.la/>> >>>> -rwxr-xr-x 1 root root 319816 Apr 19 19:58 >>>> mca_plm_rsh.so >>>> -rwxr-xr-x 1 root root 970 Apr 19 20:00 >>>> mca_pml_cm.la <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> <http://mca_pml_cm.la <http://mca_pml_cm.la/>> >>>> -rwxr-xr-x 1 root root 1507440 Apr 19 20:00 >>>> mca_pml_cm.so >>>> >>>> I'll try to find out what happened next week when I'm >>>> back in >>>> my office. >>>> >>>> >>>> Kind regards >>>> >>>> Siegmar >>>> >>>> >>>> >>>> >>>> >>>> Am 23.04.16 um 02:12 schrieb Gilles Gouaillardet: >>>> >>>> Siegmar, >>>> >>>> I will try to reproduce this on my solaris11 x86_64 >>>> vm >>>> >>>> In the mean time, can you please double check >>>> mca_pmix_pmix_pmix112.so >>>> is a 64 bits library ? >>>> (E.g, confirm "-m64" was correctly passed to pmix) >>>> >>>> Cheers, >>>> >>>> Gilles >>>> >>>> On Friday, April 22, 2016, Siegmar Gross >>>> <siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de> >>>> <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de>');> >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de>> >>>> <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de>');>> >>>> wrote: >>>> >>>> Hi Ralph, >>>> >>>> I've already used "-enable-debug". "SYSTEM_ENV" >>>> is >>>> "SunOS" or >>>> "Linux" and "MACHINE_ENV" is "sparc" or >>>> "x86_84". >>>> >>>> mkdir >>>> >>>> openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_gcc >>>> cd >>>> >>>> openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_gcc >>>> >>>> ../openmpi-v2.x-dev-1280-gc110ae8/configure \ >>>> --prefix=/usr/local/openmpi-2.0.0_64_gcc \ >>>> >>>> --libdir=/usr/local/openmpi-2.0.0_64_gcc/lib64 \ >>>> --with-jdk-bindir=/usr/local/jdk1.8.0/bin \ >>>> >>>> --with-jdk-headers=/usr/local/jdk1.8.0/include \ >>>> JAVA_HOME=/usr/local/jdk1.8.0 \ >>>> LDFLAGS="-m64" CC="gcc" CXX="g++" >>>> FC="gfortran" \ >>>> CFLAGS="-m64" CXXFLAGS="-m64" FCFLAGS="-m64" \ >>>> CPP="cpp" CXXCPP="cpp" \ >>>> --enable-mpi-cxx \ >>>> --enable-cxx-exceptions \ >>>> --enable-mpi-java \ >>>> --enable-heterogeneous \ >>>> --enable-mpi-thread-multiple \ >>>> --with-hwloc=internal \ >>>> --without-verbs \ >>>> --with-wrapper-cflags="-std=c11 -m64" \ >>>> --with-wrapper-cxxflags="-m64" \ >>>> --with-wrapper-fcflags="-m64" \ >>>> --enable-debug \ >>>> |& tee >>>> log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_gcc >>>> >>>> >>>> mkdir >>>> >>>> openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_cc >>>> cd >>>> >>>> openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_cc >>>> >>>> ../openmpi-v2.x-dev-1280-gc110ae8/configure \ >>>> --prefix=/usr/local/openmpi-2.0.0_64_cc \ >>>> --libdir=/usr/local/openmpi-2.0.0_64_cc/lib64 >>>> \ >>>> --with-jdk-bindir=/usr/local/jdk1.8.0/bin \ >>>> >>>> --with-jdk-headers=/usr/local/jdk1.8.0/include \ >>>> JAVA_HOME=/usr/local/jdk1.8.0 \ >>>> LDFLAGS="-m64" CC="cc" CXX="CC" FC="f95" \ >>>> CFLAGS="-m64" CXXFLAGS="-m64 >>>> -library=stlport4" >>>> FCFLAGS="-m64" \ >>>> CPP="cpp" CXXCPP="cpp" \ >>>> --enable-mpi-cxx \ >>>> --enable-cxx-exceptions \ >>>> --enable-mpi-java \ >>>> --enable-heterogeneous \ >>>> --enable-mpi-thread-multiple \ >>>> --with-hwloc=internal \ >>>> --without-verbs \ >>>> --with-wrapper-cflags="-m64" \ >>>> --with-wrapper-cxxflags="-m64 >>>> -library=stlport4" \ >>>> --with-wrapper-fcflags="-m64" \ >>>> --with-wrapper-ldflags="" \ >>>> --enable-debug \ >>>> |& tee >>>> log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_cc >>>> >>>> >>>> Kind regards >>>> >>>> Siegmar >>>> >>>> Am 21.04.2016 um 18:18 schrieb Ralph Castain: >>>> >>>> Can you please rebuild OMPI with >>>> -enable-debug in >>>> the configure >>>> cmd? It will let us see more error output >>>> >>>> >>>> On Apr 21, 2016, at 8:52 AM, Siegmar >>>> Gross >>>> <siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de>> >>>> <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de>');> >>>> wrote: >>>> >>>> Hi Ralph, >>>> >>>> I don't see any additional information. >>>> >>>> tyr hello_1 108 mpiexec -np 4 --host >>>> tyr,sunpc1,linpc1,ruester -mca >>>> mca_base_component_show_load_errors 1 >>>> hello_1_mpi >>>> [tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>:06211 >>>> <http://tyr.informatik.hs-fulda.de:06211 >>>> <http://tyr.informatik.hs-fulda.de:06211/>> >>>> <http://tyr.informatik.hs-fulda.de:06211 >>>> <http://tyr.informatik.hs-fulda.de:06211/>> >>>> <http://tyr.informatik.hs-fulda.de:06211 >>>> <http://tyr.informatik.hs-fulda.de:06211/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de:06211 >>>> <http://tyr.informatik.hs-fulda.de:06211/>> >>>> <http://tyr.informatik.hs-fulda.de:06211 >>>> <http://tyr.informatik.hs-fulda.de:06211/>>] >>>> [[48741,0],0] >>>> ORTE_ERROR_LOG: Not found in file >>>> >>>> >>>> >>>> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c >>>> >>>> at line 638 >>>> >>>> >>>> >>>> -------------------------------------------------------------------------- >>>> It looks like orte_init failed for some >>>> reason; your >>>> parallel process is >>>> likely to abort. There are many >>>> reasons that >>>> a parallel >>>> process can >>>> fail during orte_init; some of which >>>> are due >>>> to configuration or >>>> environment problems. This failure >>>> appears to >>>> be an >>>> internal failure; >>>> here's some additional information >>>> (which may >>>> only be >>>> relevant to an >>>> Open MPI developer): >>>> >>>> opal_pmix_base_select failed >>>> --> Returned value Not found (-13) >>>> instead of >>>> ORTE_SUCCESS >>>> >>>> >>>> >>>> -------------------------------------------------------------------------- >>>> >>>> >>>> tyr hello_1 109 mpiexec -np 4 --host >>>> tyr,sunpc1,linpc1,ruester -mca >>>> mca_base_component_show_load_errors 1 >>>> -mca >>>> pmix_base_verbose >>>> 10 -mca pmix_server_verbose 5 >>>> hello_1_mpi >>>> [tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>>] >>>> mca: base: >>>> components_register: registering >>>> framework >>>> pmix components >>>> [tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>>] >>>> mca: base: >>>> components_open: opening pmix components >>>> [tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>>] >>>> mca:base:select: >>>> Auto-selecting pmix components >>>> [tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>>] >>>> mca:base:select:( >>>> pmix) No component selected! >>>> [tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>> >>>> <http://tyr.informatik.hs-fulda.de:06212 >>>> <http://tyr.informatik.hs-fulda.de:06212/>>] >>>> [[48738,0],0] >>>> ORTE_ERROR_LOG: Not found in file >>>> >>>> >>>> >>>> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c >>>> >>>> at line 638 >>>> >>>> >>>> >>>> -------------------------------------------------------------------------- >>>> It looks like orte_init failed for some >>>> reason; your >>>> parallel process is >>>> likely to abort. There are many >>>> reasons that >>>> a parallel >>>> process can >>>> fail during orte_init; some of which >>>> are due >>>> to configuration or >>>> environment problems. This failure >>>> appears to >>>> be an >>>> internal failure; >>>> here's some additional information >>>> (which may >>>> only be >>>> relevant to an >>>> Open MPI developer): >>>> >>>> opal_pmix_base_select failed >>>> --> Returned value Not found (-13) >>>> instead of >>>> ORTE_SUCCESS >>>> >>>> >>>> >>>> -------------------------------------------------------------------------- >>>> tyr hello_1 110 >>>> >>>> >>>> Kind regards >>>> >>>> Siegmar >>>> >>>> >>>> Am 21.04.2016 um 17:24 schrieb Ralph >>>> Castain: >>>> >>>> Hmmm…it looks like you built the >>>> right >>>> components, but >>>> they are not being picked up. Can >>>> you run >>>> your mpiexec >>>> command again, adding “-mca >>>> mca_base_component_show_load_errors >>>> 1” to >>>> the cmd line? >>>> >>>> >>>> On Apr 21, 2016, at 8:16 AM, >>>> Siegmar Gross >>>> >>>> <siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de>> >>>> <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de>');> >>>> wrote: >>>> >>>> Hi Ralph, >>>> >>>> I have attached ompi_info >>>> output for >>>> both compilers >>>> from my >>>> sparc machine and the listings >>>> for >>>> both compilers >>>> from the >>>> <prefix>/lib/openmpi >>>> directories. >>>> Hopefully that >>>> helps to >>>> find the problem. >>>> >>>> hermes tmp 3 tar zvft >>>> openmpi-2.x_info.tar.gz >>>> -rw-r--r-- root/root 10969 >>>> 2016-04-21 17:06 >>>> ompi_info_SunOS_sparc_cc.txt >>>> -rw-r--r-- root/root 11044 >>>> 2016-04-21 17:06 >>>> ompi_info_SunOS_sparc_gcc.txt >>>> -rw-r--r-- root/root 71252 >>>> 2016-04-21 17:02 >>>> lib64_openmpi.txt >>>> hermes tmp 4 >>>> >>>> >>>> Kind regards and thank you very >>>> much >>>> once more for >>>> your help >>>> >>>> Siegmar >>>> >>>> >>>> Am 21.04.2016 um 15:54 schrieb >>>> Ralph >>>> Castain: >>>> >>>> Odd - it would appear that >>>> none of >>>> the pmix >>>> components built? Can you >>>> send >>>> along the output from >>>> ompi_info? >>>> Or just send a >>>> listing of the files in the >>>> <prefix>/lib/openmpi >>>> directory? >>>> >>>> >>>> On Apr 21, 2016, at >>>> 1:27 AM, >>>> Siegmar Gross >>>> >>>> <siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de> >>>> <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de>');> >>>> >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de>> >>>> <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de>');>> >>>> wrote: >>>> >>>> Hi Ralph, >>>> >>>> Am 21.04.2016 um 00:18 >>>> schrieb >>>> Ralph Castain: >>>> >>>> Could you please >>>> rerun >>>> these test and >>>> add “-mca >>>> pmix_base_verbose 10 >>>> -mca >>>> pmix_server_verbose >>>> 5” to your cmd >>>> line? I need to see >>>> why the >>>> pmix components >>>> failed. >>>> >>>> >>>> >>>> tyr spawn 111 mpiexec >>>> -np 1 --host >>>> >>>> tyr,sunpc1,linpc1,ruester -mca >>>> pmix_base_verbose 10 >>>> -mca >>>> pmix_server_verbose 5 >>>> spawn_multiple_master >>>> >>>> [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>>:26652] mca: >>>> base: >>>> components_register: >>>> registering >>>> framework pmix >>>> components >>>> >>>> [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>>:26652] mca: >>>> base: components_open: >>>> opening >>>> pmix components >>>> >>>> [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>>:26652] >>>> mca:base:select: >>>> Auto-selecting pmix components >>>> >>>> [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>>:26652] >>>> mca:base:select:( pmix) >>>> No >>>> component selected! >>>> >>>> [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>>:26652] >>>> [[52794,0],0] >>>> ORTE_ERROR_LOG: >>>> Not found in file >>>> >>>> >>>> >>>> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c >>>> >>>> at line 638 >>>> >>>> >>>> >>>> -------------------------------------------------------------------------- >>>> It looks like orte_init >>>> failed >>>> for some >>>> reason; your parallel >>>> process is >>>> likely to abort. There >>>> are >>>> many reasons >>>> that a parallel process >>>> can >>>> fail during orte_init; >>>> some of >>>> which are due >>>> to configuration or >>>> environment problems. >>>> This >>>> failure appears >>>> to be an internal >>>> failure; >>>> here's some additional >>>> information (which >>>> may only be relevant to >>>> an >>>> Open MPI developer): >>>> >>>> opal_pmix_base_select >>>> failed >>>> --> Returned value Not >>>> found >>>> (-13) instead >>>> of ORTE_SUCCESS >>>> >>>> >>>> >>>> -------------------------------------------------------------------------- >>>> tyr spawn 112 >>>> >>>> >>>> >>>> >>>> tyr hello_1 116 mpiexec >>>> -np 1 >>>> --host >>>> >>>> tyr,sunpc1,linpc1,ruester -mca >>>> pmix_base_verbose 10 >>>> -mca >>>> pmix_server_verbose 5 >>>> hello_1_mpi >>>> >>>> [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>>:27261] mca: >>>> base: >>>> components_register: >>>> registering >>>> framework pmix >>>> components >>>> >>>> [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>>:27261] mca: >>>> base: components_open: >>>> opening >>>> pmix components >>>> >>>> [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>>:27261] >>>> mca:base:select: >>>> Auto-selecting pmix components >>>> >>>> [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>>:27261] >>>> mca:base:select:( pmix) >>>> No >>>> component selected! >>>> >>>> [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de/ >>>> <http://tyr.informatik.hs-fulda.de/>>:27261] >>>> [[52315,0],0] >>>> ORTE_ERROR_LOG: >>>> Not found in file >>>> >>>> >>>> >>>> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c >>>> >>>> at line 638 >>>> >>>> >>>> >>>> -------------------------------------------------------------------------- >>>> It looks like orte_init >>>> failed >>>> for some >>>> reason; your parallel >>>> process is >>>> likely to abort. There >>>> are >>>> many reasons >>>> that a parallel process >>>> can >>>> fail during orte_init; >>>> some of >>>> which are due >>>> to configuration or >>>> environment problems. >>>> This >>>> failure appears >>>> to be an internal >>>> failure; >>>> here's some additional >>>> information (which >>>> may only be relevant to >>>> an >>>> Open MPI developer): >>>> >>>> opal_pmix_base_select >>>> failed >>>> --> Returned value Not >>>> found >>>> (-13) instead >>>> of ORTE_SUCCESS >>>> >>>> >>>> >>>> -------------------------------------------------------------------------- >>>> tyr hello_1 117 >>>> >>>> >>>> >>>> Thank you very much for >>>> your help. >>>> >>>> >>>> Kind regards >>>> >>>> Siegmar >>>> >>>> >>>> >>>> >>>> Thanks >>>> Ralph >>>> >>>> On Apr 20, >>>> 2016, at >>>> 10:12 AM, >>>> Siegmar Gross >>>> >>>> <siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de> >>>> <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de>');> >>>> >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de>> >>>> <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de >>>> <mailto:siegmar.gr...@informatik.hs-fulda.de>');>> >>>> wrote: >>>> >>>> Hi, >>>> >>>> I have built >>>> >>>> openmpi-v2.x-dev-1280-gc110ae8 on my >>>> machines >>>> (Solaris 10 >>>> Sparc, >>>> Solaris 10 >>>> x86_64, and >>>> openSUSE Linux >>>> 12.1 x86_64) >>>> with >>>> gcc-5.1.0 and Sun >>>> C 5.13. >>>> Unfortunately >>>> I get >>>> runtime errors >>>> for >>>> some programs. >>>> >>>> >>>> Sun C 5.13: >>>> =========== >>>> >>>> For all my test >>>> programs I get the >>>> same error on >>>> Solaris >>>> Sparc and >>>> Solaris x86_64, >>>> while >>>> the programs >>>> work fine on >>>> Linux. >>>> >>>> tyr hello_1 115 >>>> mpiexec -np 2 >>>> hello_1_mpi >>>> >>>> [tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>>:22373] >>>> [[61763,0],0] >>>> ORTE_ERROR_LOG: Not >>>> found in file >>>> >>>> >>>> >>>> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c >>>> >>>> at line 638 >>>> >>>> >>>> >>>> -------------------------------------------------------------------------- >>>> It looks like >>>> orte_init failed for >>>> some reason; >>>> your >>>> parallel process is >>>> likely to abort. >>>> There are many >>>> reasons that a >>>> parallel process can >>>> fail during >>>> orte_init; >>>> some of which >>>> are due to >>>> configuration or >>>> environment >>>> problems. >>>> This failure >>>> appears to be an >>>> internal failure; >>>> here's some >>>> additional >>>> information >>>> (which may only >>>> be >>>> relevant to an >>>> Open MPI >>>> developer): >>>> >>>> >>>> opal_pmix_base_select >>>> failed >>>> --> Returned >>>> value Not >>>> found (-13) >>>> instead of >>>> ORTE_SUCCESS >>>> >>>> >>>> >>>> -------------------------------------------------------------------------- >>>> tyr hello_1 116 >>>> >>>> >>>> >>>> >>>> GCC-5.1.0: >>>> ========== >>>> >>>> tyr spawn 121 >>>> mpiexec >>>> -np 1 --host >>>> >>>> tyr,sunpc1,linpc1,ruester >>>> >>>> spawn_multiple_master >>>> >>>> Parent process 0 >>>> running on >>>> >>>> tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> I create 3 slave >>>> processes. >>>> >>>> >>>> [tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>>:25366] >>>> PMIX ERROR: >>>> UNPACK-PAST-END in file >>>> >>>> >>>> >>>> ../../../../../../openmpi-v2.x-dev-1280-gc110ae8/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c >>>> >>>> >>>> at line 829 >>>> >>>> [tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>> >>>> >>>> <http://tyr.informatik.hs-fulda.de >>>> <http://tyr.informatik.hs-fulda.de/>> >>>> <http://tyr.informatik.hs-fulda.de> >>>> >>>> <http://tyr.informatik.hs-fulda.de> >>>> <http://tyr.informatik.hs-fulda.de>:25366] >>>> PMIX ERROR: >>>> UNPACK-PAST-END in file >>>> >>>> >>>> >>>> ../../../../../../openmpi-v2.x-dev-1280-gc110ae8/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c >>>> >>>> >>>> at line 2176 >>>> [tyr:25377] *** >>>> An >>>> error occurred in >>>> >>>> MPI_Comm_spawn_multiple >>>> [tyr:25377] *** >>>> reported by process >>>> [3308257281,0] >>>> [tyr:25377] *** >>>> on >>>> communicator >>>> MPI_COMM_WORLD >>>> [tyr:25377] *** >>>> MPI_ERR_SPAWN: could >>>> not spawn >>>> processes >>>> [tyr:25377] *** >>>> MPI_ERRORS_ARE_FATAL >>>> (processes in >>>> this >>>> communicator will >>>> now abort, >>>> [tyr:25377] >>>> *** and >>>> potentially >>>> your MPI job) >>>> tyr spawn 122 >>>> >>>> >>>> I would be >>>> grateful if >>>> somebody can >>>> fix the >>>> problems. >>>> Thank you very >>>> much for any >>>> help in >>>> advance. >>>> >>>> >>>> Kind regards >>>> >>>> Siegmar >>>> >>>> >>>> >>>> <hello_1_mpi.c><spawn_multiple_master.c>_______________________________________________ >>>> >>>> >>>> users mailing >>>> list >>>> >>>> us...@open-mpi.org >>>> <mailto:us...@open-mpi.org> >>>> Subscription: >>>> >>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>> Link to this >>>> post: >>>> >>>> >>>> <http://www.open-mpi.org/community/lists/users/2016/04/28983.php >>>> <http://www.open-mpi.org/community/lists/users/2016/04/28983.php>> >> >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org <mailto:us...@open-mpi.org> >> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users >> <https://www.open-mpi.org/mailman/listinfo.cgi/users> >> Link to this post: >> http://www.open-mpi.org/community/lists/users/2016/04/29047.php >> <http://www.open-mpi.org/community/lists/users/2016/04/29047.php> >> > _______________________________________________ > users mailing list > us...@open-mpi.org <mailto:us...@open-mpi.org> > Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users > <https://www.open-mpi.org/mailman/listinfo.cgi/users> > Link to this post: > http://www.open-mpi.org/community/lists/users/2016/04/29049.php > <http://www.open-mpi.org/community/lists/users/2016/04/29049.php>