Siegmar, in pmix_bfrop_pack_app, app->argc must be replaced with app[i].argc
I will PR to pmix, ompi and ompi-release when I am back at work on Monday Cheers, Gilles On Thursday, April 28, 2016, Gilles Gouaillardet <gil...@rist.or.jp> wrote: > Siegmar, > > > can you please also post the source of spawn_slave ? > > > Cheers, > > Gilles > > On 4/28/2016 1:17 AM, Siegmar Gross wrote: > > Hi Gilles, > > it is not necessary to have a heterogeneous environment to reproduce > the error as you can see below. All machines are 64 bit. > > tyr spawn 119 ompi_info | grep -e "OPAL repo revision" -e "C compiler > absolute" > OPAL repo revision: v2.x-dev-1290-gbd0e4e1 > C compiler absolute: /usr/local/gcc-5.1.0/bin/gcc > tyr spawn 120 uname -a > SunOS tyr.informatik.hs-fulda.de 5.10 Generic_150400-11 sun4u sparc > SUNW,A70 Solaris > tyr spawn 121 mpiexec -np 1 --host tyr,tyr,tyr,tyr spawn_multiple_master > > Parent process 0 running on tyr.informatik.hs-fulda.de > I create 3 slave processes. > > [tyr.informatik.hs-fulda.de:27286] PMIX ERROR: UNPACK-PAST-END in file > ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c > at line 829 > [tyr.informatik.hs-fulda.de:27286] PMIX ERROR: UNPACK-PAST-END in file > ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c > at line 2176 > [tyr:27288] *** An error occurred in MPI_Comm_spawn_multiple > [tyr:27288] *** reported by process [3434086401,0] > [tyr:27288] *** on communicator MPI_COMM_WORLD > [tyr:27288] *** MPI_ERR_SPAWN: could not spawn processes > [tyr:27288] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will > now abort, > [tyr:27288] *** and potentially your MPI job) > tyr spawn 122 > > > > > > > sunpc1 fd1026 105 ompi_info | grep -e "OPAL repo revision" -e "C compiler > absolute" > OPAL repo revision: v2.x-dev-1290-gbd0e4e1 > C compiler absolute: /usr/local/gcc-5.1.0/bin/gcc > sunpc1 fd1026 106 uname -a > SunOS sunpc1 5.10 Generic_147441-21 i86pc i386 i86pc Solaris > sunpc1 fd1026 107 mpiexec -np 1 --host sunpc1,sunpc1,sunpc1,sunpc1 > spawn_multiple_master > > Parent process 0 running on sunpc1 > I create 3 slave processes. > > [sunpc1:00368] PMIX ERROR: UNPACK-PAST-END in file > ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c > at line 829 > [sunpc1:00368] PMIX ERROR: UNPACK-PAST-END in file > ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c > at line 2176 > [sunpc1:370] *** An error occurred in MPI_Comm_spawn_multiple > [sunpc1:370] *** reported by process [43909121,0] > [sunpc1:370] *** on communicator MPI_COMM_WORLD > [sunpc1:370] *** MPI_ERR_SPAWN: could not spawn processes > [sunpc1:370] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will > now abort, > [sunpc1:370] *** and potentially your MPI job) > sunpc1 fd1026 108 > > > > > > linpc1 fd1026 105 ompi_info | grep -e "OPAL repo revision" -e "C compiler > absolute" > OPAL repo revision: v2.x-dev-1290-gbd0e4e1 > C compiler absolute: /usr/local/gcc-5.1.0/bin/gcc > linpc1 fd1026 106 uname -a > Linux linpc1 3.1.10-1.29-desktop #1 SMP PREEMPT Fri May 31 20:10:04 UTC > 2013 (2529847) x86_64 x86_64 x86_64 GNU/Linux > linpc1 fd1026 107 mpiexec -np 1 --host linpc1,linpc1,linpc1,linpc1 > spawn_multiple_master > > Parent process 0 running on linpc1 > I create 3 slave processes. > > [linpc1:21502] PMIX ERROR: UNPACK-PAST-END in file > ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c > at line 829 > [linpc1:21502] PMIX ERROR: UNPACK-PAST-END in file > ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c > at line 2176 > [linpc1:21507] *** An error occurred in MPI_Comm_spawn_multiple > [linpc1:21507] *** reported by process [1005518849,0] > [linpc1:21507] *** on communicator MPI_COMM_WORLD > [linpc1:21507] *** MPI_ERR_SPAWN: could not spawn processes > [linpc1:21507] *** MPI_ERRORS_ARE_FATAL (processes in this communicator > will now abort, > [linpc1:21507] *** and potentially your MPI job) > linpc1 fd1026 108 > > > I used the following configure command. > > ../openmpi-v2.x-dev-1290-gbd0e4e1/configure \ > --prefix=/usr/local/openmpi-2.0.0_64_gcc \ > --libdir=/usr/local/openmpi-2.0.0_64_gcc/lib64 \ > --with-jdk-bindir=/usr/local/jdk1.8.0/bin \ > --with-jdk-headers=/usr/local/jdk1.8.0/include \ > JAVA_HOME=/usr/local/jdk1.8.0 \ > LDFLAGS="-m64" CC="gcc" CXX="g++" FC="gfortran" \ > CFLAGS="-m64" CXXFLAGS="-m64" FCFLAGS="-m64" \ > CPP="cpp" CXXCPP="cpp" \ > --enable-mpi-cxx \ > --enable-cxx-exceptions \ > --enable-mpi-java \ > --enable-heterogeneous \ > --enable-mpi-thread-multiple \ > --with-hwloc=internal \ > --without-verbs \ > --with-wrapper-cflags="-std=c11 -m64" \ > --with-wrapper-cxxflags="-m64" \ > --with-wrapper-fcflags="-m64" \ > --enable-debug \ > |& tee log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_gcc > > > Kind regards > > Siegmar > > > > Am 27.04.2016 um 13:21 schrieb Gilles Gouaillardet: > > Siegmar, > > please add this to your CFLAGS for the time being. > > configure tries to detect which flags must be added for C99 support, and > it seems > the test is not working for Solaris 10 and Oracle compilers. > this is no more a widely used environment, and I am not sure I can find > the > time to fix this > in a near future. > > > regarding the runtime issue, can you please describe your 4 hosts (os, > endianness and bitness) > > Cheers, > > Gilles > > On Wednesday, April 27, 2016, Siegmar Gross > <siegmar.gr...@informatik.hs-fulda.de > <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de');> > <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de');>> > wrote: > > Hi Gilles, > > adding "-std=c99" to CFLAGS solves the problem with the missing > library. > Shall I add it permanently to my configure command or will you add it, > so that I will not run into problems if you need the C11 standard > later? > > "spawn_multiple_master" breaks with the same error that I reported > yesterday for my gcc-version of Open MPI. Hopefully you can solve the > problem as well. > > > Kind regards and thank you very much for your help > > Siegmar > > > Am 27.04.2016 um 08:05 schrieb Gilles Gouaillardet: > > Siegmar, > > > here is the error : > > configure:17969: cc -o conftest -m64 -D_REENTRANT -g -g > -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1 > > -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc > > -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/include > > -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/include > -D_REENTRANT > > -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/hwloc/hwloc1112/hwloc/include > > -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/hwloc/hwloc1112/hwloc/include > > -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/event/libevent2022/libevent > > -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/event/libevent2022/libevent/include > > -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/event/libevent2022/libevent/include > -m64 conftest.c >&5 > "/usr/include/stdbool.h", line 42: #error: "Use of <stdbool.h> is > valid only > in a c99 compilation environment." > > > i cannot reproduce this on solaris 11 with oracle studio 5.3 > compiler, > and i > do not have solaris 10 yet. > > could you please re-configure with '-std=c99' appended to your > CFLAGS > and see > if it helps ? > > > Cheers, > > > Gilles > > > On 4/26/2016 7:57 PM, Siegmar Gross wrote: > > Hi Gilles and Ralph, > > I was able to sort out my mess. In my last email I compared > the > files from "SunOS_sparc/openmpi-2.0.0_64_gcc/lib64/openmpi" > from > the attachment of my email to Ralph with the files from > "SunOS_sparc/openmpi-2.0.0_64_cc/lib64/openmpi" from my > current > file system. That's the reason while I have had different > timestamps. The other problem was that Ralph didn't recognize > that "mca_pmix_pmix112.so" wasn't built on Solaris with the > Sun C compiler. I've removed most of the files from the > attachment > of my email so that it is easier to see the relevant files. > Below > I try to give you more information that may be relevant to > track > down the problem. I still get an error running one of my small > test programs, when I use my gcc-version of Open MPI. > "mca_pmix_pmix112.so" is a 64 bits library. > > Linux_x86_64/openmpi-2.0.0_64_cc/lib64/openmpi: > ... > -rwxr-xr-x 1 root root 261327 Apr 19 16:46 mca_plm_slurm.so > -rwxr-xr-x 1 root root 1002 Apr 19 16:45 > mca_pmix_pmix112.la > <http://mca_pmix_pmix112.la> <http://mca_pmix_pmix112.la> > -rwxr-xr-x 1 root root 3906526 Apr 19 16:45 > mca_pmix_pmix112.so > -rwxr-xr-x 1 root root 966 Apr 19 16:51 mca_pml_cm.la > <http://mca_pml_cm.la> <http://mca_pml_cm.la> > -rwxr-xr-x 1 root root 1574265 Apr 19 16:51 mca_pml_cm.so > ... > > Linux_x86_64/openmpi-2.0.0_64_gcc/lib64/openmpi: > ... > -rwxr-xr-x 1 root root 70371 Apr 19 16:43 mca_plm_slurm.so > -rwxr-xr-x 1 root root 1008 Apr 19 16:42 > mca_pmix_pmix112.la > <http://mca_pmix_pmix112.la> <http://mca_pmix_pmix112.la> > -rwxr-xr-x 1 root root 1029005 Apr 19 16:42 > mca_pmix_pmix112.so > -rwxr-xr-x 1 root root 972 Apr 19 16:46 mca_pml_cm.la > <http://mca_pml_cm.la> <http://mca_pml_cm.la> > -rwxr-xr-x 1 root root 284858 Apr 19 16:46 mca_pml_cm.so > ... > > SunOS_sparc/openmpi-2.0.0_64_cc/lib64/openmpi: > ... > -rwxr-xr-x 1 root root 319816 Apr 19 19:58 mca_plm_rsh.so > -rwxr-xr-x 1 root root 970 Apr 19 20:00 mca_pml_cm.la > <http://mca_pml_cm.la> <http://mca_pml_cm.la> > -rwxr-xr-x 1 root root 1507440 Apr 19 20:00 mca_pml_cm.so > ... > > SunOS_sparc/openmpi-2.0.0_64_gcc/lib64/openmpi: > ... > -rwxr-xr-x 1 root root 153280 Apr 19 19:49 mca_plm_rsh.so > -rwxr-xr-x 1 root root 1007 Apr 19 19:47 > mca_pmix_pmix112.la > <http://mca_pmix_pmix112.la> <http://mca_pmix_pmix112.la> > -rwxr-xr-x 1 root root 1400512 Apr 19 19:47 > mca_pmix_pmix112.so > -rwxr-xr-x 1 root root 971 Apr 19 19:52 mca_pml_cm.la > <http://mca_pml_cm.la> <http://mca_pml_cm.la> > -rwxr-xr-x 1 root root 342440 Apr 19 19:52 mca_pml_cm.so > ... > > SunOS_x86_64/openmpi-2.0.0_64_cc/lib64/openmpi: > ... > -rwxr-xr-x 1 root root 300096 Apr 19 17:18 mca_plm_rsh.so > -rwxr-xr-x 1 root root 970 Apr 19 17:23 mca_pml_cm.la > <http://mca_pml_cm.la> <http://mca_pml_cm.la> > -rwxr-xr-x 1 root root 1458816 Apr 19 17:23 mca_pml_cm.so > ... > > SunOS_x86_64/openmpi-2.0.0_64_gcc/lib64/openmpi: > ... > -rwxr-xr-x 1 root root 133096 Apr 19 17:42 mca_plm_rsh.so > -rwxr-xr-x 1 root root 1007 Apr 19 17:41 > mca_pmix_pmix112.la > <http://mca_pmix_pmix112.la> <http://mca_pmix_pmix112.la> > -rwxr-xr-x 1 root root 1320240 Apr 19 17:41 > mca_pmix_pmix112.so > -rwxr-xr-x 1 root root 971 Apr 19 17:46 mca_pml_cm.la > <http://mca_pml_cm.la> <http://mca_pml_cm.la> > -rwxr-xr-x 1 root root 419848 Apr 19 17:46 mca_pml_cm.so > ... > > > Yesterday I've installed openmpi-v2.x-dev-1290-gbd0e4e1 so > that we > have a current version for the investigation of the problem. > Once > more mca_pmix_pmix112.so isn't available on Solaris if I use > the > Sun C compiler. > > "config.log" for gcc-5.1.0 shows the following. > > ... > configure:127799: /bin/bash > > '../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/ > pmix/configure' succeeded for opal/mca/pmix/pmix112/pmix > configure:127916: checking if MCA component pmix:pmix112 can > compile > configure:127918: result: yes > configure:5637: --- MCA component pmix:external (m4 > configuration > macro) > configure:128523: checking for MCA component pmix:external > compile > mode > configure:128529: result: dso > configure:129054: checking if MCA component pmix:external can > compile > configure:129056: result: no > ... > config.status:3897: creating opal/mca/pmix/Makefile > config.status:3897: creating opal/mca/pmix/s1/Makefile > config.status:3897: creating opal/mca/pmix/cray/Makefile > config.status:3897: creating opal/mca/pmix/s2/Makefile > config.status:3897: creating opal/mca/pmix/pmix112/Makefile > config.status:3897: creating opal/mca/pmix/external/Makefile > ... > MCA_BUILD_opal_pmix_cray_DSO_FALSE='#' > MCA_BUILD_opal_pmix_cray_DSO_TRUE='' > MCA_BUILD_opal_pmix_external_DSO_FALSE='#' > MCA_BUILD_opal_pmix_external_DSO_TRUE='' > MCA_BUILD_opal_pmix_pmix112_DSO_FALSE='#' > MCA_BUILD_opal_pmix_pmix112_DSO_TRUE='' > MCA_BUILD_opal_pmix_s1_DSO_FALSE='#' > MCA_BUILD_opal_pmix_s1_DSO_TRUE='' > MCA_BUILD_opal_pmix_s2_DSO_FALSE='#' > MCA_BUILD_opal_pmix_s2_DSO_TRUE='' > ... > MCA_opal_FRAMEWORKS='common allocator backtrace btl dl event > hwloc if > installdirs memchecker memcpy memory mpool pmix pstat rcache > sec > shmem timer' > MCA_opal_FRAMEWORKS_SUBDIRS='mca/common mca/allocator > mca/backtrace mca/btl > mca/dl mca/event mca/hwloc mca/if mca/installdirs > mca/memchecker > mca/memcpy > mca/memory mca/mpool mca/pmix mca/pstat mca/rcache mca/sec > mca/shmem mca/timer' > > MCA_opal_FRAMEWORK_COMPONENT_ALL_SUBDIRS='$(MCA_opal_common_ALL_SUBDIRS) > $(MCA_opal_allocator_ALL_SUBDIRS) > $(MCA_opal_backtrace_ALL_SUBDIRS) > $(MCA_opal_btl_ALL_SUBDIRS) $(MCA_opal_dl_ALL_SUBDIRS) > $(MCA_opal_event_ALL_SUBDIRS) $(MCA_opal_hwloc_ALL_SUBDIRS) > $(MCA_opal_if_ALL_SUBDIRS) $(MCA_opal_installdirs_ALL_SUBDIRS) > $(MCA_opal_memchecker_ALL_SUBDIRS) > $(MCA_opal_memcpy_ALL_SUBDIRS) > $(MCA_opal_memory_ALL_SUBDIRS) $(MCA_opal_mpool_ALL_SUBDIRS) > $(MCA_opal_pmix_ALL_SUBDIRS) $(MCA_opal_pstat_ALL_SUBDIRS) > $(MCA_opal_rcache_ALL_SUBDIRS) $(MCA_opal_sec_ALL_SUBDIRS) > $(MCA_opal_shmem_ALL_SUBDIRS) $(MCA_opal_timer_ALL_SUBDIRS)' > > MCA_opal_FRAMEWORK_COMPONENT_DSO_SUBDIRS='$(MCA_opal_common_DSO_SUBDIRS) > $(MCA_opal_allocator_DSO_SUBDIRS) > $(MCA_opal_backtrace_DSO_SUBDIRS) > $(MCA_opal_btl_DSO_SUBDIRS) $(MCA_opal_dl_DSO_SUBDIRS) > $(MCA_opal_event_DSO_SUBDIRS) $(MCA_opal_hwloc_DSO_SUBDIRS) > $(MCA_opal_if_DSO_SUBDIRS) $(MCA_opal_installdirs_DSO_SUBDIRS) > $(MCA_opal_memchecker_DSO_SUBDIRS) > $(MCA_opal_memcpy_DSO_SUBDIRS) > $(MCA_opal_memory_DSO_SUBDIRS) $(MCA_opal_mpool_DSO_SUBDIRS) > $(MCA_opal_pmix_DSO_SUBDIRS) $(MCA_opal_pstat_DSO_SUBDIRS) > $(MCA_opal_rcache_DSO_SUBDIRS) $(MCA_opal_sec_DSO_SUBDIRS) > $(MCA_opal_shmem_DSO_SUBDIRS) $(MCA_opal_timer_DSO_SUBDIRS)' > > MCA_opal_FRAMEWORK_COMPONENT_STATIC_SUBDIRS='$(MCA_opal_common_STATIC_SUBDIRS) > $(MCA_opal_allocator_STATIC_SUBDIRS) > $(MCA_opal_backtrace_STATIC_SUBDIRS) > $(MCA_opal_btl_STATIC_SUBDIRS) $(MCA_opal_dl_STATIC_SUBDIRS) > $(MCA_opal_event_STATIC_SUBDIRS) > $(MCA_opal_hwloc_STATIC_SUBDIRS) > $(MCA_opal_if_STATIC_SUBDIRS) > $(MCA_opal_installdirs_STATIC_SUBDIRS) > $(MCA_opal_memchecker_STATIC_SUBDIRS) > $(MCA_opal_memcpy_STATIC_SUBDIRS) > $(MCA_opal_memory_STATIC_SUBDIRS) > $(MCA_opal_mpool_STATIC_SUBDIRS) > $(MCA_opal_pmix_STATIC_SUBDIRS) > $(MCA_opal_pstat_STATIC_SUBDIRS) > $(MCA_opal_rcache_STATIC_SUBDIRS) > $(MCA_opal_sec_STATIC_SUBDIRS) > $(MCA_opal_shmem_STATIC_SUBDIRS) > $(MCA_opal_timer_STATIC_SUBDIRS)' > MCA_opal_FRAMEWORK_LIBS=' $(MCA_opal_common_STATIC_LTLIBS) > mca/allocator/libmca_allocator.la <http://libmca_allocator.la> > <http://libmca_allocator.la> > $(MCA_opal_allocator_STATIC_LTLIBS) > mca/backtrace/libmca_backtrace.la <http://libmca_backtrace.la> > <http://libmca_backtrace.la> > $(MCA_opal_backtrace_STATIC_LTLIBS) > mca/btl/libmca_btl.la <http://libmca_btl.la> > <http://libmca_btl.la> > $(MCA_opal_btl_STATIC_LTLIBS) mca/dl/libmca_dl.la > <http://libmca_dl.la> <http://libmca_dl.la> > $(MCA_opal_dl_STATIC_LTLIBS) mca/event/libmca_event.la > <http://libmca_event.la> <http://libmca_event.la> > $(MCA_opal_event_STATIC_LTLIBS) mca/hwloc/libmca_hwloc.la > <http://libmca_hwloc.la> <http://libmca_hwloc.la> > $(MCA_opal_hwloc_STATIC_LTLIBS) mca/if/libmca_if.la > <http://libmca_if.la> <http://libmca_if.la> > $(MCA_opal_if_STATIC_LTLIBS) mca/installdirs/ > libmca_installdirs.la > <http://libmca_installdirs.la> <http://libmca_installdirs.la> > $(MCA_opal_installdirs_STATIC_LTLIBS) > mca/memchecker/libmca_memchecker.la > <http://libmca_memchecker.la> <http://libmca_memchecker.la> > $(MCA_opal_memchecker_STATIC_LTLIBS) mca/memcpy/ > libmca_memcpy.la > <http://libmca_memcpy.la> <http://libmca_memcpy.la> > $(MCA_opal_memcpy_STATIC_LTLIBS) mca/memory/libmca_memory.la > <http://libmca_memory.la> <http://libmca_memory.la> > $(MCA_opal_memory_STATIC_LTLIBS) mca/mpool/libmca_mpool.la > <http://libmca_mpool.la> <http://libmca_mpool.la> > $(MCA_opal_mpool_STATIC_LTLIBS) mca/pmix/libmca_pmix.la > <http://libmca_pmix.la> <http://libmca_pmix.la> > $(MCA_opal_pmix_STATIC_LTLIBS) mca/pstat/libmca_pstat.la > <http://libmca_pstat.la> <http://libmca_pstat.la> > $(MCA_opal_pstat_STATIC_LTLIBS) mca/rcache/libmca_rcache.la > <http://libmca_rcache.la> <http://libmca_rcache.la> > $(MCA_opal_rcache_STATIC_LTLIBS) mca/sec/libmca_sec.la > <http://libmca_sec.la> <http://libmca_sec.la> > $(MCA_opal_sec_STATIC_LTLIBS) mca/shmem/libmca_shmem.la > <http://libmca_shmem.la> <http://libmca_shmem.la> > $(MCA_opal_shmem_STATIC_LTLIBS) mca/timer/libmca_timer.la > <http://libmca_timer.la> <http://libmca_timer.la> > $(MCA_opal_timer_STATIC_LTLIBS)' > ... > MCA_opal_pmix_ALL_COMPONENTS=' s1 cray s2 pmix112 external' > MCA_opal_pmix_ALL_SUBDIRS=' mca/pmix/s1 mca/pmix/cray > mca/pmix/s2 > mca/pmix/pmix112 mca/pmix/external' > MCA_opal_pmix_DSO_COMPONENTS=' pmix112' > MCA_opal_pmix_DSO_SUBDIRS=' mca/pmix/pmix112' > MCA_opal_pmix_STATIC_COMPONENTS='' > MCA_opal_pmix_STATIC_LTLIBS='' > MCA_opal_pmix_STATIC_SUBDIRS='' > ... > opal_pmix_ext_CPPFLAGS='' > opal_pmix_ext_LDFLAGS='' > opal_pmix_ext_LIBS='' > > opal_pmix_pmix112_CPPFLAGS='-I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/include/pmix > -I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/include > -I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix > -I$(OPAL_TOP_SRCDIR)/opal/mca/pmix/pmix112/pmix' > > opal_pmix_pmix112_LIBS='$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/ > libpmix.la > <http://libpmix.la> <http://libpmix.la>' > > ... > > > > "config.log" for Sun C 5.13 shows the following. > > ... > configure:127803: /bin/bash > > '../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/ > pmix/configure' *failed* for opal/mca/pmix/pmix112/pmix > configure:128379: checking if MCA component pmix:pmix112 can > compile > configure:128381: result: no > configure:5637: --- MCA component pmix:external (m4 > configuration > macro) > configure:128523: checking for MCA component pmix:external > compile > mode > configure:128529: result: dso > configure:129054: checking if MCA component pmix:external can > compile > configure:129056: result: no > ... > config.status:3887: creating opal/mca/pmix/Makefile > config.status:3887: creating opal/mca/pmix/s1/Makefile > config.status:3887: creating opal/mca/pmix/cray/Makefile > config.status:3887: creating opal/mca/pmix/s2/Makefile > config.status:3887: creating opal/mca/pmix/pmix112/Makefile > config.status:3887: creating opal/mca/pmix/external/Makefile > ... > MCA_BUILD_opal_pmix_cray_DSO_FALSE='#' > MCA_BUILD_opal_pmix_cray_DSO_TRUE='' > MCA_BUILD_opal_pmix_external_DSO_FALSE='#' > MCA_BUILD_opal_pmix_external_DSO_TRUE='' > MCA_BUILD_opal_pmix_pmix112_DSO_FALSE='#' > MCA_BUILD_opal_pmix_pmix112_DSO_TRUE='' > MCA_BUILD_opal_pmix_s1_DSO_FALSE='#' > MCA_BUILD_opal_pmix_s1_DSO_TRUE='' > MCA_BUILD_opal_pmix_s2_DSO_FALSE='#' > MCA_BUILD_opal_pmix_s2_DSO_TRUE='' > ... > MCA_opal_FRAMEWORKS='common allocator backtrace btl dl event > hwloc if > installdirs memchecker memcpy memory mpool pmix pstat rcache > sec > shmem timer' > MCA_opal_FRAMEWORKS_SUBDIRS='mca/common mca/allocator > mca/backtrace mca/btl > mca/dl mca/event mca/hwloc mca/if mca/installdirs > mca/memchecker > mca/memcpy > mca/memory mca/mpool mca/pmix mca/pstat mca/rcache mca/sec > mca/shmem mca/timer' > > MCA_opal_FRAMEWORK_COMPONENT_ALL_SUBDIRS='$(MCA_opal_common_ALL_SUBDIRS) > $(MCA_opal_allocator_ALL_SUBDIRS) > $(MCA_opal_backtrace_ALL_SUBDIRS) > $(MCA_opal_btl_ALL_SUBDIRS) $(MCA_opal_dl_ALL_SUBDIRS) > $(MCA_opal_event_ALL_SUBDIRS) $(MCA_opal_hwloc_ALL_SUBDIRS) > $(MCA_opal_if_ALL_SUBDIRS) $(MCA_opal_installdirs_ALL_SUBDIRS) > $(MCA_opal_memchecker_ALL_SUBDIRS) > $(MCA_opal_memcpy_ALL_SUBDIRS) > $(MCA_opal_memory_ALL_SUBDIRS) $(MCA_opal_mpool_ALL_SUBDIRS) > $(MCA_opal_pmix_ALL_SUBDIRS) $(MCA_opal_pstat_ALL_SUBDIRS) > $(MCA_opal_rcache_ALL_SUBDIRS) $(MCA_opal_sec_ALL_SUBDIRS) > $(MCA_opal_shmem_ALL_SUBDIRS) $(MCA_opal_timer_ALL_SUBDIRS)' > > MCA_opal_FRAMEWORK_COMPONENT_DSO_SUBDIRS='$(MCA_opal_common_DSO_SUBDIRS) > $(MCA_opal_allocator_DSO_SUBDIRS) > $(MCA_opal_backtrace_DSO_SUBDIRS) > $(MCA_opal_btl_DSO_SUBDIRS) $(MCA_opal_dl_DSO_SUBDIRS) > $(MCA_opal_event_DSO_SUBDIRS) $(MCA_opal_hwloc_DSO_SUBDIRS) > $(MCA_opal_if_DSO_SUBDIRS) $(MCA_opal_installdirs_DSO_SUBDIRS) > $(MCA_opal_memchecker_DSO_SUBDIRS) > $(MCA_opal_memcpy_DSO_SUBDIRS) > $(MCA_opal_memory_DSO_SUBDIRS) $(MCA_opal_mpool_DSO_SUBDIRS) > $(MCA_opal_pmix_DSO_SUBDIRS) $(MCA_opal_pstat_DSO_SUBDIRS) > $(MCA_opal_rcache_DSO_SUBDIRS) $(MCA_opal_sec_DSO_SUBDIRS) > $(MCA_opal_shmem_DSO_SUBDIRS) $(MCA_opal_timer_DSO_SUBDIRS)' > > MCA_opal_FRAMEWORK_COMPONENT_STATIC_SUBDIRS='$(MCA_opal_common_STATIC_SUBDIRS) > $(MCA_opal_allocator_STATIC_SUBDIRS) > $(MCA_opal_backtrace_STATIC_SUBDIRS) > $(MCA_opal_btl_STATIC_SUBDIRS) $(MCA_opal_dl_STATIC_SUBDIRS) > $(MCA_opal_event_STATIC_SUBDIRS) > $(MCA_opal_hwloc_STATIC_SUBDIRS) > $(MCA_opal_if_STATIC_SUBDIRS) > $(MCA_opal_installdirs_STATIC_SUBDIRS) > $(MCA_opal_memchecker_STATIC_SUBDIRS) > $(MCA_opal_memcpy_STATIC_SUBDIRS) > $(MCA_opal_memory_STATIC_SUBDIRS) > $(MCA_opal_mpool_STATIC_SUBDIRS) > $(MCA_opal_pmix_STATIC_SUBDIRS) > $(MCA_opal_pstat_STATIC_SUBDIRS) > $(MCA_opal_rcache_STATIC_SUBDIRS) > $(MCA_opal_sec_STATIC_SUBDIRS) > $(MCA_opal_shmem_STATIC_SUBDIRS) > $(MCA_opal_timer_STATIC_SUBDIRS)' > MCA_opal_FRAMEWORK_LIBS=' $(MCA_opal_common_STATIC_LTLIBS) > mca/allocator/libmca_allocator.la <http://libmca_allocator.la> > <http://libmca_allocator.la> > $(MCA_opal_allocator_STATIC_LTLIBS) > mca/backtrace/libmca_backtrace.la <http://libmca_backtrace.la> > <http://libmca_backtrace.la> > $(MCA_opal_backtrace_STATIC_LTLIBS) > mca/btl/libmca_btl.la <http://libmca_btl.la> > <http://libmca_btl.la> > $(MCA_opal_btl_STATIC_LTLIBS) mca/dl/libmca_dl.la > <http://libmca_dl.la> <http://libmca_dl.la> > $(MCA_opal_dl_STATIC_LTLIBS) mca/event/libmca_event.la > <http://libmca_event.la> <http://libmca_event.la> > $(MCA_opal_event_STATIC_LTLIBS) mca/hwloc/libmca_hwloc.la > <http://libmca_hwloc.la> <http://libmca_hwloc.la> > $(MCA_opal_hwloc_STATIC_LTLIBS) mca/if/libmca_if.la > <http://libmca_if.la> <http://libmca_if.la> > $(MCA_opal_if_STATIC_LTLIBS) mca/installdirs/ > libmca_installdirs.la > <http://libmca_installdirs.la> <http://libmca_installdirs.la> > $(MCA_opal_installdirs_STATIC_LTLIBS) > mca/memchecker/libmca_memchecker.la > <http://libmca_memchecker.la> <http://libmca_memchecker.la> > $(MCA_opal_memchecker_STATIC_LTLIBS) mca/memcpy/ > libmca_memcpy.la > <http://libmca_memcpy.la> <http://libmca_memcpy.la> > $(MCA_opal_memcpy_STATIC_LTLIBS) mca/memory/libmca_memory.la > <http://libmca_memory.la> <http://libmca_memory.la> > $(MCA_opal_memory_STATIC_LTLIBS) mca/mpool/libmca_mpool.la > <http://libmca_mpool.la> <http://libmca_mpool.la> > $(MCA_opal_mpool_STATIC_LTLIBS) mca/pmix/libmca_pmix.la > <http://libmca_pmix.la> <http://libmca_pmix.la> > $(MCA_opal_pmix_STATIC_LTLIBS) mca/pstat/libmca_pstat.la > <http://libmca_pstat.la> <http://libmca_pstat.la> > $(MCA_opal_pstat_STATIC_LTLIBS) mca/rcache/libmca_rcache.la > <http://libmca_rcache.la> <http://libmca_rcache.la> > $(MCA_opal_rcache_STATIC_LTLIBS) mca/sec/libmca_sec.la > <http://libmca_sec.la> <http://libmca_sec.la> > $(MCA_opal_sec_STATIC_LTLIBS) mca/shmem/libmca_shmem.la > <http://libmca_shmem.la> <http://libmca_shmem.la> > $(MCA_opal_shmem_STATIC_LTLIBS) mca/timer/libmca_timer.la > <http://libmca_timer.la> <http://libmca_timer.la> > $(MCA_opal_timer_STATIC_LTLIBS)' > ... > MCA_opal_pmix_ALL_COMPONENTS=' s1 cray s2 pmix112 external' > MCA_opal_pmix_ALL_SUBDIRS=' mca/pmix/s1 mca/pmix/cray > mca/pmix/s2 > mca/pmix/pmix112 mca/pmix/external' > MCA_opal_pmix_DSO_COMPONENTS='' > MCA_opal_pmix_DSO_SUBDIRS='' > MCA_opal_pmix_STATIC_COMPONENTS='' > MCA_opal_pmix_STATIC_LTLIBS='' > MCA_opal_pmix_STATIC_SUBDIRS='' > ... > opal_pmix_ext_CPPFLAGS='' > opal_pmix_ext_LDFLAGS='' > opal_pmix_ext_LIBS='' > opal_pmix_pmix112_CPPFLAGS='' > opal_pmix_pmix112_LIBS='' > ... > > > > > I've attached the config.log files for pmix. > > tyr openmpi-2.0.0 142 tar zvft pmix_config.log.tar.gz > -rw-r--r-- root/root 136291 2016-04-25 08:05:34 > > openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/pmix/pmix112/pmix/config.log > -rw-r--r-- root/root 528808 2016-04-25 08:07:54 > > openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_gcc/opal/mca/pmix/pmix112/pmix/config.log > tyr openmpi-2.0.0 143 > > > > I've also attached the output for the broken execution of > "spawn_multiple_master" for my gcc-version of Open MPI. > "spawn_master" works as expected with my gcc-version of Open > MPI. > > Hopefully you can fix the problem. > > > Kind regards and thank you very much for your help > > Siegmar > > > > Am 23.04.2016 um 21:34 schrieb Siegmar Gross: > > Hi Gilles, > > I don't know what happened, but the files are not > available now > and they were definitely available when I answered the > email from > Ralph. The files also have a different timestamp now. This > is an > extract from my email to Ralph for Solaris Sparc. > > -rwxr-xr-x 1 root root 977 Apr 19 19:49 mca_plm_rsh.la > <http://mca_plm_rsh.la> <http://mca_plm_rsh.la> > -rwxr-xr-x 1 root root 153280 Apr 19 19:49 mca_plm_rsh.so > -rwxr-xr-x 1 root root 1007 Apr 19 19:47 > mca_pmix_pmix112.la <http://mca_pmix_pmix112.la> > <http://mca_pmix_pmix112.la> > -rwxr-xr-x 1 root root 1400512 Apr 19 19:47 > mca_pmix_pmix112.so > -rwxr-xr-x 1 root root 971 Apr 19 19:52 mca_pml_cm.la > <http://mca_pml_cm.la> <http://mca_pml_cm.la> > -rwxr-xr-x 1 root root 342440 Apr 19 19:52 mca_pml_cm.so > > Now I have the following output for these files. > > -rwxr-xr-x 1 root root 976 Apr 19 19:58 mca_plm_rsh.la > <http://mca_plm_rsh.la> <http://mca_plm_rsh.la> > -rwxr-xr-x 1 root root 319816 Apr 19 19:58 mca_plm_rsh.so > -rwxr-xr-x 1 root root 970 Apr 19 20:00 mca_pml_cm.la > <http://mca_pml_cm.la> <http://mca_pml_cm.la> > -rwxr-xr-x 1 root root 1507440 Apr 19 20:00 mca_pml_cm.so > > I'll try to find out what happened next week when I'm back > in > my office. > > > Kind regards > > Siegmar > > > > > > Am 23.04.16 um 02:12 schrieb Gilles Gouaillardet: > > Siegmar, > > I will try to reproduce this on my solaris11 x86_64 vm > > In the mean time, can you please double check > mca_pmix_pmix_pmix112.so > is a 64 bits library ? > (E.g, confirm "-m64" was correctly passed to pmix) > > Cheers, > > Gilles > > On Friday, April 22, 2016, Siegmar Gross > <siegmar.gr...@informatik.hs-fulda.de > <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de');> > <mailto:siegmar.gr...@informatik.hs-fulda.de> > <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de');>> > wrote: > > Hi Ralph, > > I've already used "-enable-debug". "SYSTEM_ENV" is > "SunOS" or > "Linux" and "MACHINE_ENV" is "sparc" or "x86_84". > > mkdir > > openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_gcc > cd > > openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_gcc > > ../openmpi-v2.x-dev-1280-gc110ae8/configure \ > --prefix=/usr/local/openmpi-2.0.0_64_gcc \ > --libdir=/usr/local/openmpi-2.0.0_64_gcc/lib64 \ > --with-jdk-bindir=/usr/local/jdk1.8.0/bin \ > --with-jdk-headers=/usr/local/jdk1.8.0/include \ > JAVA_HOME=/usr/local/jdk1.8.0 \ > LDFLAGS="-m64" CC="gcc" CXX="g++" FC="gfortran" > \ > CFLAGS="-m64" CXXFLAGS="-m64" FCFLAGS="-m64" \ > CPP="cpp" CXXCPP="cpp" \ > --enable-mpi-cxx \ > --enable-cxx-exceptions \ > --enable-mpi-java \ > --enable-heterogeneous \ > --enable-mpi-thread-multiple \ > --with-hwloc=internal \ > --without-verbs \ > --with-wrapper-cflags="-std=c11 -m64" \ > --with-wrapper-cxxflags="-m64" \ > --with-wrapper-fcflags="-m64" \ > --enable-debug \ > |& tee > log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_gcc > > > mkdir > > openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_cc > cd > > openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_cc > > ../openmpi-v2.x-dev-1280-gc110ae8/configure \ > --prefix=/usr/local/openmpi-2.0.0_64_cc \ > --libdir=/usr/local/openmpi-2.0.0_64_cc/lib64 \ > --with-jdk-bindir=/usr/local/jdk1.8.0/bin \ > --with-jdk-headers=/usr/local/jdk1.8.0/include \ > JAVA_HOME=/usr/local/jdk1.8.0 \ > LDFLAGS="-m64" CC="cc" CXX="CC" FC="f95" \ > CFLAGS="-m64" CXXFLAGS="-m64 -library=stlport4" > FCFLAGS="-m64" \ > CPP="cpp" CXXCPP="cpp" \ > --enable-mpi-cxx \ > --enable-cxx-exceptions \ > --enable-mpi-java \ > --enable-heterogeneous \ > --enable-mpi-thread-multiple \ > --with-hwloc=internal \ > --without-verbs \ > --with-wrapper-cflags="-m64" \ > --with-wrapper-cxxflags="-m64 -library=stlport4" > \ > --with-wrapper-fcflags="-m64" \ > --with-wrapper-ldflags="" \ > --enable-debug \ > |& tee > log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_cc > > > Kind regards > > Siegmar > > Am 21.04.2016 um 18:18 schrieb Ralph Castain: > > Can you please rebuild OMPI with -enable-debug > in > the configure > cmd? It will let us see more error output > > > On Apr 21, 2016, at 8:52 AM, Siegmar Gross > <siegmar.gr...@informatik.hs-fulda.de> > <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de');> > wrote: > > Hi Ralph, > > I don't see any additional information. > > tyr hello_1 108 mpiexec -np 4 --host > tyr,sunpc1,linpc1,ruester -mca > mca_base_component_show_load_errors 1 > hello_1_mpi > [tyr.informatik.hs-fulda.de:06211 > <http://tyr.informatik.hs-fulda.de:06211> > <http://tyr.informatik.hs-fulda.de:06211> > <http://tyr.informatik.hs-fulda.de:06211> > <http://tyr.informatik.hs-fulda.de:06211>] > [[48741,0],0] > ORTE_ERROR_LOG: Not found in file > > > ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c > > at line 638 > > > -------------------------------------------------------------------------- > It looks like orte_init failed for some > reason; your > parallel process is > likely to abort. There are many reasons > that > a parallel > process can > fail during orte_init; some of which are > due > to configuration or > environment problems. This failure > appears to > be an > internal failure; > here's some additional information (which > may > only be > relevant to an > Open MPI developer): > > opal_pmix_base_select failed > --> Returned value Not found (-13) > instead of > ORTE_SUCCESS > > > -------------------------------------------------------------------------- > > > tyr hello_1 109 mpiexec -np 4 --host > tyr,sunpc1,linpc1,ruester -mca > mca_base_component_show_load_errors 1 -mca > pmix_base_verbose > 10 -mca pmix_server_verbose 5 hello_1_mpi > [tyr.informatik.hs-fulda.de:06212 > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212>] > mca: base: > components_register: registering framework > pmix components > [tyr.informatik.hs-fulda.de:06212 > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212>] > mca: base: > components_open: opening pmix components > [tyr.informatik.hs-fulda.de:06212 > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212>] > mca:base:select: > Auto-selecting pmix components > [tyr.informatik.hs-fulda.de:06212 > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212>] > mca:base:select:( > pmix) No component selected! > [tyr.informatik.hs-fulda.de:06212 > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212>] > [[48738,0],0] > ORTE_ERROR_LOG: Not found in file > > > ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c > > at line 638 > > > -------------------------------------------------------------------------- > It looks like orte_init failed for some > reason; your > parallel process is > likely to abort. There are many reasons > that > a parallel > process can > fail during orte_init; some of which are > due > to configuration or > environment problems. This failure > appears to > be an > internal failure; > here's some additional information (which > may > only be > relevant to an > Open MPI developer): > > opal_pmix_base_select failed > --> Returned value Not found (-13) > instead of > ORTE_SUCCESS > > > -------------------------------------------------------------------------- > tyr hello_1 110 > > > Kind regards > > Siegmar > > > Am 21.04.2016 um 17:24 schrieb Ralph > Castain: > > Hmmm…it looks like you built the right > components, but > they are not being picked up. Can you > run > your mpiexec > command again, adding “-mca > mca_base_component_show_load_errors 1” > to > the cmd line? > > > On Apr 21, 2016, at 8:16 AM, > Siegmar Gross > > <siegmar.gr...@informatik.hs-fulda.de> > <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de');> > wrote: > > Hi Ralph, > > I have attached ompi_info output > for > both compilers > from my > sparc machine and the listings for > both compilers > from the > <prefix>/lib/openmpi directories. > Hopefully that > helps to > find the problem. > > hermes tmp 3 tar zvft > openmpi-2.x_info.tar.gz > -rw-r--r-- root/root 10969 > 2016-04-21 17:06 > ompi_info_SunOS_sparc_cc.txt > -rw-r--r-- root/root 11044 > 2016-04-21 17:06 > ompi_info_SunOS_sparc_gcc.txt > -rw-r--r-- root/root 71252 > 2016-04-21 17:02 > lib64_openmpi.txt > hermes tmp 4 > > > Kind regards and thank you very > much > once more for > your help > > Siegmar > > > Am 21.04.2016 um 15:54 schrieb > Ralph > Castain: > > Odd - it would appear that > none of > the pmix > components built? Can you send > along the output from > ompi_info? > Or just send a > listing of the files in the > <prefix>/lib/openmpi > directory? > > > On Apr 21, 2016, at 1:27 > AM, > Siegmar Gross > > <siegmar.gr...@informatik.hs-fulda.de > <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de');> > > <mailto:siegmar.gr...@informatik.hs-fulda.de> > <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de');>> > wrote: > > Hi Ralph, > > Am 21.04.2016 um 00:18 > schrieb > Ralph Castain: > > Could you please rerun > these test and > add “-mca > pmix_base_verbose 10 > -mca > pmix_server_verbose > 5” to your cmd > line? I need to see > why the > pmix components > failed. > > > > tyr spawn 111 mpiexec -np > 1 --host > tyr,sunpc1,linpc1,ruester > -mca > pmix_base_verbose 10 -mca > pmix_server_verbose 5 > spawn_multiple_master > [ > tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/> > <http://tyr.informatik.hs-fulda.de/>:26652] mca: > base: components_register: > registering > framework pmix components > [ > tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/> > <http://tyr.informatik.hs-fulda.de/>:26652] mca: > base: components_open: > opening > pmix components > [ > tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/> > <http://tyr.informatik.hs-fulda.de/>:26652] > mca:base:select: > Auto-selecting pmix components > [ > tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/> > <http://tyr.informatik.hs-fulda.de/>:26652] > mca:base:select:( pmix) No > component selected! > [ > tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/> > <http://tyr.informatik.hs-fulda.de/>:26652] > [[52794,0],0] > ORTE_ERROR_LOG: > Not found in file > > > ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c > > at line 638 > > > -------------------------------------------------------------------------- > It looks like orte_init > failed > for some > reason; your parallel > process is > likely to abort. There > are > many reasons > that a parallel process > can > fail during orte_init; > some of > which are due > to configuration or > environment problems. > This > failure appears > to be an internal failure; > here's some additional > information (which > may only be relevant to an > Open MPI developer): > > opal_pmix_base_select > failed > --> Returned value Not > found > (-13) instead > of ORTE_SUCCESS > > > -------------------------------------------------------------------------- > tyr spawn 112 > > > > > tyr hello_1 116 mpiexec > -np 1 > --host > tyr,sunpc1,linpc1,ruester > -mca > pmix_base_verbose 10 -mca > pmix_server_verbose 5 > hello_1_mpi > [ > tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/> > <http://tyr.informatik.hs-fulda.de/>:27261] mca: > base: components_register: > registering > framework pmix components > [ > tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/> > <http://tyr.informatik.hs-fulda.de/>:27261] mca: > base: components_open: > opening > pmix components > [ > tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/> > <http://tyr.informatik.hs-fulda.de/>:27261] > mca:base:select: > Auto-selecting pmix components > [ > tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/> > <http://tyr.informatik.hs-fulda.de/>:27261] > mca:base:select:( pmix) No > component selected! > [ > tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/> > <http://tyr.informatik.hs-fulda.de/>:27261] > [[52315,0],0] > ORTE_ERROR_LOG: > Not found in file > > > ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c > > at line 638 > > > -------------------------------------------------------------------------- > It looks like orte_init > failed > for some > reason; your parallel > process is > likely to abort. There > are > many reasons > that a parallel process > can > fail during orte_init; > some of > which are due > to configuration or > environment problems. > This > failure appears > to be an internal failure; > here's some additional > information (which > may only be relevant to an > Open MPI developer): > > opal_pmix_base_select > failed > --> Returned value Not > found > (-13) instead > of ORTE_SUCCESS > > > -------------------------------------------------------------------------- > tyr hello_1 117 > > > > Thank you very much for > your help. > > > Kind regards > > Siegmar > > > > > Thanks > Ralph > > On Apr 20, 2016, > at > 10:12 AM, > Siegmar Gross > > <siegmar.gr...@informatik.hs-fulda.de > <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de');> > > <mailto:siegmar.gr...@informatik.hs-fulda.de> > <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de');>> > wrote: > > Hi, > > I have built > > openmpi-v2.x-dev-1280-gc110ae8 on my > machines > (Solaris 10 Sparc, > Solaris 10 > x86_64, and > openSUSE Linux > 12.1 x86_64) with > gcc-5.1.0 and Sun > C 5.13. > Unfortunately > I get > runtime errors for > some programs. > > > Sun C 5.13: > =========== > > For all my test > programs I get the > same error on > Solaris > Sparc and > Solaris x86_64, > while > the programs > work fine on > Linux. > > tyr hello_1 115 > mpiexec -np 2 > hello_1_mpi > > [tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de>:22373] > [[61763,0],0] > ORTE_ERROR_LOG: Not > found in file > > > ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c > > at line 638 > > > -------------------------------------------------------------------------- > It looks like > orte_init failed for > some reason; your > parallel process is > likely to abort. > There are many > reasons that a > parallel process can > fail during > orte_init; > some of which > are due to > configuration or > environment > problems. > This failure > appears to be an > internal failure; > here's some > additional > information > (which may only be > relevant to an > Open MPI > developer): > > > opal_pmix_base_select > failed > --> Returned value > Not > found (-13) > instead of > ORTE_SUCCESS > > > -------------------------------------------------------------------------- > tyr hello_1 116 > > > > > GCC-5.1.0: > ========== > > tyr spawn 121 > mpiexec > -np 1 --host > > tyr,sunpc1,linpc1,ruester > > spawn_multiple_master > > Parent process 0 > running on > > tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > I create 3 slave > processes. > > > [tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de>:25366] > PMIX ERROR: > UNPACK-PAST-END in file > > > ../../../../../../openmpi-v2.x-dev-1280-gc110ae8/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c > > > at line 829 > > [tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > <http://tyr.informatik.hs-fulda.de>:25366] > PMIX ERROR: > UNPACK-PAST-END in file > > > ../../../../../../openmpi-v2.x-dev-1280-gc110ae8/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c > > > at line 2176 > [tyr:25377] *** An > error occurred in > > MPI_Comm_spawn_multiple > [tyr:25377] *** > reported by process > [3308257281,0] > [tyr:25377] *** on > communicator > MPI_COMM_WORLD > [tyr:25377] *** > MPI_ERR_SPAWN: could > not spawn > processes > [tyr:25377] *** > MPI_ERRORS_ARE_FATAL > (processes in this > communicator will > now abort, > [tyr:25377] *** > and > potentially > your MPI job) > tyr spawn 122 > > > I would be > grateful if > somebody can > fix the problems. > Thank you very > much for any help > in > advance. > > > Kind regards > > Siegmar > > > <hello_1_mpi.c><spawn_multiple_master.c>_______________________________________________ > > > users mailing list > us...@open-mpi.org > <mailto:us...@open-mpi.org> > Subscription: > > http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > > > <http://www.open-mpi.org/community/lists/users/2016/04/28983.php> > >