Thanks Rowena, greg -----Original Message----- From: users [mailto:users-boun...@open-mpi.org] On Behalf Of users-requ...@open-mpi.org Sent: Thursday, 28 April 2016 11:34 AM To: us...@open-mpi.org Subject: users Digest, Vol 3486, Issue 3
Send users mailing list submissions to us...@open-mpi.org To subscribe or unsubscribe via the World Wide Web, visit https://www.open-mpi.org/mailman/listinfo.cgi/users or, via email, send a message with subject or body 'help' to users-requ...@open-mpi.org You can reach the person managing the list at users-ow...@open-mpi.org When replying, please edit your Subject line so it is more specific than "Re: Contents of users digest..." Today's Topics: 1. Re: runtime errors for openmpi-v2.x-dev-1280-gc110ae8 (Siegmar Gross) 2. Re: runtime errors for openmpi-v2.x-dev-1280-gc110ae8 (Gilles Gouaillardet) ---------------------------------------------------------------------- Message: 1 List-Post: users@lists.open-mpi.org Date: Wed, 27 Apr 2016 18:17:35 +0200 From: Siegmar Gross <siegmar.gr...@informatik.hs-fulda.de> To: Open MPI Users <us...@open-mpi.org> Subject: Re: [OMPI users] runtime errors for openmpi-v2.x-dev-1280-gc110ae8 Message-ID: <a8d0facb-7b67-2a79-8fba-015819983...@informatik.hs-fulda.de> Content-Type: text/plain; charset="windows-1252"; Format="flowed" Hi Gilles, it is not necessary to have a heterogeneous environment to reproduce the error as you can see below. All machines are 64 bit. tyr spawn 119 ompi_info | grep -e "OPAL repo revision" -e "C compiler absolute" OPAL repo revision: v2.x-dev-1290-gbd0e4e1 C compiler absolute: /usr/local/gcc-5.1.0/bin/gcc tyr spawn 120 uname -a SunOS tyr.informatik.hs-fulda.de 5.10 Generic_150400-11 sun4u sparc SUNW,A70 Solaris tyr spawn 121 mpiexec -np 1 --host tyr,tyr,tyr,tyr spawn_multiple_master Parent process 0 running on tyr.informatik.hs-fulda.de I create 3 slave processes. [tyr.informatik.hs-fulda.de:27286] PMIX ERROR: UNPACK-PAST-END in file ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c at line 829 [tyr.informatik.hs-fulda.de:27286] PMIX ERROR: UNPACK-PAST-END in file ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c at line 2176 [tyr:27288] *** An error occurred in MPI_Comm_spawn_multiple [tyr:27288] *** reported by process [3434086401,0] [tyr:27288] *** on communicator MPI_COMM_WORLD [tyr:27288] *** MPI_ERR_SPAWN: could not spawn processes [tyr:27288] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort, [tyr:27288] *** and potentially your MPI job) tyr spawn 122 sunpc1 fd1026 105 ompi_info | grep -e "OPAL repo revision" -e "C compiler absolute" OPAL repo revision: v2.x-dev-1290-gbd0e4e1 C compiler absolute: /usr/local/gcc-5.1.0/bin/gcc sunpc1 fd1026 106 uname -a SunOS sunpc1 5.10 Generic_147441-21 i86pc i386 i86pc Solaris sunpc1 fd1026 107 mpiexec -np 1 --host sunpc1,sunpc1,sunpc1,sunpc1 spawn_multiple_master Parent process 0 running on sunpc1 I create 3 slave processes. [sunpc1:00368] PMIX ERROR: UNPACK-PAST-END in file ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c at line 829 [sunpc1:00368] PMIX ERROR: UNPACK-PAST-END in file ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c at line 2176 [sunpc1:370] *** An error occurred in MPI_Comm_spawn_multiple [sunpc1:370] *** reported by process [43909121,0] [sunpc1:370] *** on communicator MPI_COMM_WORLD [sunpc1:370] *** MPI_ERR_SPAWN: could not spawn processes [sunpc1:370] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort, [sunpc1:370] *** and potentially your MPI job) sunpc1 fd1026 108 linpc1 fd1026 105 ompi_info | grep -e "OPAL repo revision" -e "C compiler absolute" OPAL repo revision: v2.x-dev-1290-gbd0e4e1 C compiler absolute: /usr/local/gcc-5.1.0/bin/gcc linpc1 fd1026 106 uname -a Linux linpc1 3.1.10-1.29-desktop #1 SMP PREEMPT Fri May 31 20:10:04 UTC 2013 (2529847) x86_64 x86_64 x86_64 GNU/Linux linpc1 fd1026 107 mpiexec -np 1 --host linpc1,linpc1,linpc1,linpc1 spawn_multiple_master Parent process 0 running on linpc1 I create 3 slave processes. [linpc1:21502] PMIX ERROR: UNPACK-PAST-END in file ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c at line 829 [linpc1:21502] PMIX ERROR: UNPACK-PAST-END in file ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c at line 2176 [linpc1:21507] *** An error occurred in MPI_Comm_spawn_multiple [linpc1:21507] *** reported by process [1005518849,0] [linpc1:21507] *** on communicator MPI_COMM_WORLD [linpc1:21507] *** MPI_ERR_SPAWN: could not spawn processes [linpc1:21507] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort, [linpc1:21507] *** and potentially your MPI job) linpc1 fd1026 108 I used the following configure command. ../openmpi-v2.x-dev-1290-gbd0e4e1/configure \ --prefix=/usr/local/openmpi-2.0.0_64_gcc \ --libdir=/usr/local/openmpi-2.0.0_64_gcc/lib64 \ --with-jdk-bindir=/usr/local/jdk1.8.0/bin \ --with-jdk-headers=/usr/local/jdk1.8.0/include \ JAVA_HOME=/usr/local/jdk1.8.0 \ LDFLAGS="-m64" CC="gcc" CXX="g++" FC="gfortran" \ CFLAGS="-m64" CXXFLAGS="-m64" FCFLAGS="-m64" \ CPP="cpp" CXXCPP="cpp" \ --enable-mpi-cxx \ --enable-cxx-exceptions \ --enable-mpi-java \ --enable-heterogeneous \ --enable-mpi-thread-multiple \ --with-hwloc=internal \ --without-verbs \ --with-wrapper-cflags="-std=c11 -m64" \ --with-wrapper-cxxflags="-m64" \ --with-wrapper-fcflags="-m64" \ --enable-debug \ |& tee log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_gcc Kind regards Siegmar Am 27.04.2016 um 13:21 schrieb Gilles Gouaillardet: > Siegmar, > > please add this to your CFLAGS for the time being. > > configure tries to detect which flags must be added for C99 support, and it > seems > the test is not working for Solaris 10 and Oracle compilers. > this is no more a widely used environment, and I am not sure I can find the > time to fix this > in a near future. > > > regarding the runtime issue, can you please describe your 4 hosts (os, > endianness and bitness) > > Cheers, > > Gilles > > On Wednesday, April 27, 2016, Siegmar Gross > <siegmar.gr...@informatik.hs-fulda.de > <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de');>> wrote: > > Hi Gilles, > > adding "-std=c99" to CFLAGS solves the problem with the missing library. > Shall I add it permanently to my configure command or will you add it, > so that I will not run into problems if you need the C11 standard later? > > "spawn_multiple_master" breaks with the same error that I reported > yesterday for my gcc-version of Open MPI. Hopefully you can solve the > problem as well. > > > Kind regards and thank you very much for your help > > Siegmar > > > Am 27.04.2016 um 08:05 schrieb Gilles Gouaillardet: > > Siegmar, > > > here is the error : > > configure:17969: cc -o conftest -m64 -D_REENTRANT -g -g > -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1 > > -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc > > -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/include > > -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/include > -D_REENTRANT > > -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/hwloc/hwloc1112/hwloc/include > > -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/hwloc/hwloc1112/hwloc/include > > -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/event/libevent2022/libevent > > -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/event/libevent2022/libevent/include > > -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/event/libevent2022/libevent/include > -m64 conftest.c >&5 > "/usr/include/stdbool.h", line 42: #error: "Use of <stdbool.h> is > valid only > in a c99 compilation environment." > > > i cannot reproduce this on solaris 11 with oracle studio 5.3 compiler, > and i > do not have solaris 10 yet. > > could you please re-configure with '-std=c99' appended to your CFLAGS > and see > if it helps ? > > > Cheers, > > > Gilles > > > On 4/26/2016 7:57 PM, Siegmar Gross wrote: > > Hi Gilles and Ralph, > > I was able to sort out my mess. In my last email I compared the > files from "SunOS_sparc/openmpi-2.0.0_64_gcc/lib64/openmpi" from > the attachment of my email to Ralph with the files from > "SunOS_sparc/openmpi-2.0.0_64_cc/lib64/openmpi" from my current > file system. That's the reason while I have had different > timestamps. The other problem was that Ralph didn't recognize > that "mca_pmix_pmix112.so" wasn't built on Solaris with the > Sun C compiler. I've removed most of the files from the attachment > of my email so that it is easier to see the relevant files. Below > I try to give you more information that may be relevant to track > down the problem. I still get an error running one of my small > test programs, when I use my gcc-version of Open MPI. > "mca_pmix_pmix112.so" is a 64 bits library. > > Linux_x86_64/openmpi-2.0.0_64_cc/lib64/openmpi: > ... > -rwxr-xr-x 1 root root 261327 Apr 19 16:46 mca_plm_slurm.so > -rwxr-xr-x 1 root root 1002 Apr 19 16:45 mca_pmix_pmix112.la > <http://mca_pmix_pmix112.la> > -rwxr-xr-x 1 root root 3906526 Apr 19 16:45 mca_pmix_pmix112.so > -rwxr-xr-x 1 root root 966 Apr 19 16:51 mca_pml_cm.la > <http://mca_pml_cm.la> > -rwxr-xr-x 1 root root 1574265 Apr 19 16:51 mca_pml_cm.so > ... > > Linux_x86_64/openmpi-2.0.0_64_gcc/lib64/openmpi: > ... > -rwxr-xr-x 1 root root 70371 Apr 19 16:43 mca_plm_slurm.so > -rwxr-xr-x 1 root root 1008 Apr 19 16:42 mca_pmix_pmix112.la > <http://mca_pmix_pmix112.la> > -rwxr-xr-x 1 root root 1029005 Apr 19 16:42 mca_pmix_pmix112.so > -rwxr-xr-x 1 root root 972 Apr 19 16:46 mca_pml_cm.la > <http://mca_pml_cm.la> > -rwxr-xr-x 1 root root 284858 Apr 19 16:46 mca_pml_cm.so > ... > > SunOS_sparc/openmpi-2.0.0_64_cc/lib64/openmpi: > ... > -rwxr-xr-x 1 root root 319816 Apr 19 19:58 mca_plm_rsh.so > -rwxr-xr-x 1 root root 970 Apr 19 20:00 mca_pml_cm.la > <http://mca_pml_cm.la> > -rwxr-xr-x 1 root root 1507440 Apr 19 20:00 mca_pml_cm.so > ... > > SunOS_sparc/openmpi-2.0.0_64_gcc/lib64/openmpi: > ... > -rwxr-xr-x 1 root root 153280 Apr 19 19:49 mca_plm_rsh.so > -rwxr-xr-x 1 root root 1007 Apr 19 19:47 mca_pmix_pmix112.la > <http://mca_pmix_pmix112.la> > -rwxr-xr-x 1 root root 1400512 Apr 19 19:47 mca_pmix_pmix112.so > -rwxr-xr-x 1 root root 971 Apr 19 19:52 mca_pml_cm.la > <http://mca_pml_cm.la> > -rwxr-xr-x 1 root root 342440 Apr 19 19:52 mca_pml_cm.so > ... > > SunOS_x86_64/openmpi-2.0.0_64_cc/lib64/openmpi: > ... > -rwxr-xr-x 1 root root 300096 Apr 19 17:18 mca_plm_rsh.so > -rwxr-xr-x 1 root root 970 Apr 19 17:23 mca_pml_cm.la > <http://mca_pml_cm.la> > -rwxr-xr-x 1 root root 1458816 Apr 19 17:23 mca_pml_cm.so > ... > > SunOS_x86_64/openmpi-2.0.0_64_gcc/lib64/openmpi: > ... > -rwxr-xr-x 1 root root 133096 Apr 19 17:42 mca_plm_rsh.so > -rwxr-xr-x 1 root root 1007 Apr 19 17:41 mca_pmix_pmix112.la > <http://mca_pmix_pmix112.la> > -rwxr-xr-x 1 root root 1320240 Apr 19 17:41 mca_pmix_pmix112.so > -rwxr-xr-x 1 root root 971 Apr 19 17:46 mca_pml_cm.la > <http://mca_pml_cm.la> > -rwxr-xr-x 1 root root 419848 Apr 19 17:46 mca_pml_cm.so > ... > > > Yesterday I've installed openmpi-v2.x-dev-1290-gbd0e4e1 so that we > have a current version for the investigation of the problem. Once > more mca_pmix_pmix112.so isn't available on Solaris if I use the > Sun C compiler. > > "config.log" for gcc-5.1.0 shows the following. > > ... > configure:127799: /bin/bash > > '../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/ > pmix/configure' succeeded for opal/mca/pmix/pmix112/pmix > configure:127916: checking if MCA component pmix:pmix112 can > compile > configure:127918: result: yes > configure:5637: --- MCA component pmix:external (m4 configuration > macro) > configure:128523: checking for MCA component pmix:external compile > mode > configure:128529: result: dso > configure:129054: checking if MCA component pmix:external can > compile > configure:129056: result: no > ... > config.status:3897: creating opal/mca/pmix/Makefile > config.status:3897: creating opal/mca/pmix/s1/Makefile > config.status:3897: creating opal/mca/pmix/cray/Makefile > config.status:3897: creating opal/mca/pmix/s2/Makefile > config.status:3897: creating opal/mca/pmix/pmix112/Makefile > config.status:3897: creating opal/mca/pmix/external/Makefile > ... > MCA_BUILD_opal_pmix_cray_DSO_FALSE='#' > MCA_BUILD_opal_pmix_cray_DSO_TRUE='' > MCA_BUILD_opal_pmix_external_DSO_FALSE='#' > MCA_BUILD_opal_pmix_external_DSO_TRUE='' > MCA_BUILD_opal_pmix_pmix112_DSO_FALSE='#' > MCA_BUILD_opal_pmix_pmix112_DSO_TRUE='' > MCA_BUILD_opal_pmix_s1_DSO_FALSE='#' > MCA_BUILD_opal_pmix_s1_DSO_TRUE='' > MCA_BUILD_opal_pmix_s2_DSO_FALSE='#' > MCA_BUILD_opal_pmix_s2_DSO_TRUE='' > ... > MCA_opal_FRAMEWORKS='common allocator backtrace btl dl event > hwloc if > installdirs memchecker memcpy memory mpool pmix pstat rcache sec > shmem timer' > MCA_opal_FRAMEWORKS_SUBDIRS='mca/common mca/allocator > mca/backtrace mca/btl > mca/dl mca/event mca/hwloc mca/if mca/installdirs mca/memchecker > mca/memcpy > mca/memory mca/mpool mca/pmix mca/pstat mca/rcache mca/sec > mca/shmem mca/timer' > > MCA_opal_FRAMEWORK_COMPONENT_ALL_SUBDIRS='$(MCA_opal_common_ALL_SUBDIRS) > $(MCA_opal_allocator_ALL_SUBDIRS) > $(MCA_opal_backtrace_ALL_SUBDIRS) > $(MCA_opal_btl_ALL_SUBDIRS) $(MCA_opal_dl_ALL_SUBDIRS) > $(MCA_opal_event_ALL_SUBDIRS) $(MCA_opal_hwloc_ALL_SUBDIRS) > $(MCA_opal_if_ALL_SUBDIRS) $(MCA_opal_installdirs_ALL_SUBDIRS) > $(MCA_opal_memchecker_ALL_SUBDIRS) $(MCA_opal_memcpy_ALL_SUBDIRS) > $(MCA_opal_memory_ALL_SUBDIRS) $(MCA_opal_mpool_ALL_SUBDIRS) > $(MCA_opal_pmix_ALL_SUBDIRS) $(MCA_opal_pstat_ALL_SUBDIRS) > $(MCA_opal_rcache_ALL_SUBDIRS) $(MCA_opal_sec_ALL_SUBDIRS) > $(MCA_opal_shmem_ALL_SUBDIRS) $(MCA_opal_timer_ALL_SUBDIRS)' > > MCA_opal_FRAMEWORK_COMPONENT_DSO_SUBDIRS='$(MCA_opal_common_DSO_SUBDIRS) > $(MCA_opal_allocator_DSO_SUBDIRS) > $(MCA_opal_backtrace_DSO_SUBDIRS) > $(MCA_opal_btl_DSO_SUBDIRS) $(MCA_opal_dl_DSO_SUBDIRS) > $(MCA_opal_event_DSO_SUBDIRS) $(MCA_opal_hwloc_DSO_SUBDIRS) > $(MCA_opal_if_DSO_SUBDIRS) $(MCA_opal_installdirs_DSO_SUBDIRS) > $(MCA_opal_memchecker_DSO_SUBDIRS) $(MCA_opal_memcpy_DSO_SUBDIRS) > $(MCA_opal_memory_DSO_SUBDIRS) $(MCA_opal_mpool_DSO_SUBDIRS) > $(MCA_opal_pmix_DSO_SUBDIRS) $(MCA_opal_pstat_DSO_SUBDIRS) > $(MCA_opal_rcache_DSO_SUBDIRS) $(MCA_opal_sec_DSO_SUBDIRS) > $(MCA_opal_shmem_DSO_SUBDIRS) $(MCA_opal_timer_DSO_SUBDIRS)' > > MCA_opal_FRAMEWORK_COMPONENT_STATIC_SUBDIRS='$(MCA_opal_common_STATIC_SUBDIRS) > $(MCA_opal_allocator_STATIC_SUBDIRS) > $(MCA_opal_backtrace_STATIC_SUBDIRS) > $(MCA_opal_btl_STATIC_SUBDIRS) $(MCA_opal_dl_STATIC_SUBDIRS) > $(MCA_opal_event_STATIC_SUBDIRS) $(MCA_opal_hwloc_STATIC_SUBDIRS) > $(MCA_opal_if_STATIC_SUBDIRS) > $(MCA_opal_installdirs_STATIC_SUBDIRS) > $(MCA_opal_memchecker_STATIC_SUBDIRS) > $(MCA_opal_memcpy_STATIC_SUBDIRS) > $(MCA_opal_memory_STATIC_SUBDIRS) $(MCA_opal_mpool_STATIC_SUBDIRS) > $(MCA_opal_pmix_STATIC_SUBDIRS) $(MCA_opal_pstat_STATIC_SUBDIRS) > $(MCA_opal_rcache_STATIC_SUBDIRS) $(MCA_opal_sec_STATIC_SUBDIRS) > $(MCA_opal_shmem_STATIC_SUBDIRS) $(MCA_opal_timer_STATIC_SUBDIRS)' > MCA_opal_FRAMEWORK_LIBS=' $(MCA_opal_common_STATIC_LTLIBS) > mca/allocator/libmca_allocator.la <http://libmca_allocator.la> > $(MCA_opal_allocator_STATIC_LTLIBS) > mca/backtrace/libmca_backtrace.la <http://libmca_backtrace.la> > $(MCA_opal_backtrace_STATIC_LTLIBS) > mca/btl/libmca_btl.la <http://libmca_btl.la> > $(MCA_opal_btl_STATIC_LTLIBS) mca/dl/libmca_dl.la > <http://libmca_dl.la> > $(MCA_opal_dl_STATIC_LTLIBS) mca/event/libmca_event.la > <http://libmca_event.la> > $(MCA_opal_event_STATIC_LTLIBS) mca/hwloc/libmca_hwloc.la > <http://libmca_hwloc.la> > $(MCA_opal_hwloc_STATIC_LTLIBS) mca/if/libmca_if.la > <http://libmca_if.la> > $(MCA_opal_if_STATIC_LTLIBS) mca/installdirs/libmca_installdirs.la > <http://libmca_installdirs.la> > $(MCA_opal_installdirs_STATIC_LTLIBS) > mca/memchecker/libmca_memchecker.la <http://libmca_memchecker.la> > $(MCA_opal_memchecker_STATIC_LTLIBS) mca/memcpy/libmca_memcpy.la > <http://libmca_memcpy.la> > $(MCA_opal_memcpy_STATIC_LTLIBS) mca/memory/libmca_memory.la > <http://libmca_memory.la> > $(MCA_opal_memory_STATIC_LTLIBS) mca/mpool/libmca_mpool.la > <http://libmca_mpool.la> > $(MCA_opal_mpool_STATIC_LTLIBS) mca/pmix/libmca_pmix.la > <http://libmca_pmix.la> > $(MCA_opal_pmix_STATIC_LTLIBS) mca/pstat/libmca_pstat.la > <http://libmca_pstat.la> > $(MCA_opal_pstat_STATIC_LTLIBS) mca/rcache/libmca_rcache.la > <http://libmca_rcache.la> > $(MCA_opal_rcache_STATIC_LTLIBS) mca/sec/libmca_sec.la > <http://libmca_sec.la> > $(MCA_opal_sec_STATIC_LTLIBS) mca/shmem/libmca_shmem.la > <http://libmca_shmem.la> > $(MCA_opal_shmem_STATIC_LTLIBS) mca/timer/libmca_timer.la > <http://libmca_timer.la> > $(MCA_opal_timer_STATIC_LTLIBS)' > ... > MCA_opal_pmix_ALL_COMPONENTS=' s1 cray s2 pmix112 external' > MCA_opal_pmix_ALL_SUBDIRS=' mca/pmix/s1 mca/pmix/cray mca/pmix/s2 > mca/pmix/pmix112 mca/pmix/external' > MCA_opal_pmix_DSO_COMPONENTS=' pmix112' > MCA_opal_pmix_DSO_SUBDIRS=' mca/pmix/pmix112' > MCA_opal_pmix_STATIC_COMPONENTS='' > MCA_opal_pmix_STATIC_LTLIBS='' > MCA_opal_pmix_STATIC_SUBDIRS='' > ... > opal_pmix_ext_CPPFLAGS='' > opal_pmix_ext_LDFLAGS='' > opal_pmix_ext_LIBS='' > > opal_pmix_pmix112_CPPFLAGS='-I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/include/pmix > -I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/include > -I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix > -I$(OPAL_TOP_SRCDIR)/opal/mca/pmix/pmix112/pmix' > > opal_pmix_pmix112_LIBS='$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/libpmix.la > <http://libpmix.la>' > > ... > > > > "config.log" for Sun C 5.13 shows the following. > > ... > configure:127803: /bin/bash > > '../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/ > pmix/configure' *failed* for opal/mca/pmix/pmix112/pmix > configure:128379: checking if MCA component pmix:pmix112 can > compile > configure:128381: result: no > configure:5637: --- MCA component pmix:external (m4 configuration > macro) > configure:128523: checking for MCA component pmix:external compile > mode > configure:128529: result: dso > configure:129054: checking if MCA component pmix:external can > compile > configure:129056: result: no > ... > config.status:3887: creating opal/mca/pmix/Makefile > config.status:3887: creating opal/mca/pmix/s1/Makefile > config.status:3887: creating opal/mca/pmix/cray/Makefile > config.status:3887: creating opal/mca/pmix/s2/Makefile > config.status:3887: creating opal/mca/pmix/pmix112/Makefile > config.status:3887: creating opal/mca/pmix/external/Makefile > ... > MCA_BUILD_opal_pmix_cray_DSO_FALSE='#' > MCA_BUILD_opal_pmix_cray_DSO_TRUE='' > MCA_BUILD_opal_pmix_external_DSO_FALSE='#' > MCA_BUILD_opal_pmix_external_DSO_TRUE='' > MCA_BUILD_opal_pmix_pmix112_DSO_FALSE='#' > MCA_BUILD_opal_pmix_pmix112_DSO_TRUE='' > MCA_BUILD_opal_pmix_s1_DSO_FALSE='#' > MCA_BUILD_opal_pmix_s1_DSO_TRUE='' > MCA_BUILD_opal_pmix_s2_DSO_FALSE='#' > MCA_BUILD_opal_pmix_s2_DSO_TRUE='' > ... > MCA_opal_FRAMEWORKS='common allocator backtrace btl dl event > hwloc if > installdirs memchecker memcpy memory mpool pmix pstat rcache sec > shmem timer' > MCA_opal_FRAMEWORKS_SUBDIRS='mca/common mca/allocator > mca/backtrace mca/btl > mca/dl mca/event mca/hwloc mca/if mca/installdirs mca/memchecker > mca/memcpy > mca/memory mca/mpool mca/pmix mca/pstat mca/rcache mca/sec > mca/shmem mca/timer' > > MCA_opal_FRAMEWORK_COMPONENT_ALL_SUBDIRS='$(MCA_opal_common_ALL_SUBDIRS) > $(MCA_opal_allocator_ALL_SUBDIRS) > $(MCA_opal_backtrace_ALL_SUBDIRS) > $(MCA_opal_btl_ALL_SUBDIRS) $(MCA_opal_dl_ALL_SUBDIRS) > $(MCA_opal_event_ALL_SUBDIRS) $(MCA_opal_hwloc_ALL_SUBDIRS) > $(MCA_opal_if_ALL_SUBDIRS) $(MCA_opal_installdirs_ALL_SUBDIRS) > $(MCA_opal_memchecker_ALL_SUBDIRS) $(MCA_opal_memcpy_ALL_SUBDIRS) > $(MCA_opal_memory_ALL_SUBDIRS) $(MCA_opal_mpool_ALL_SUBDIRS) > $(MCA_opal_pmix_ALL_SUBDIRS) $(MCA_opal_pstat_ALL_SUBDIRS) > $(MCA_opal_rcache_ALL_SUBDIRS) $(MCA_opal_sec_ALL_SUBDIRS) > $(MCA_opal_shmem_ALL_SUBDIRS) $(MCA_opal_timer_ALL_SUBDIRS)' > > MCA_opal_FRAMEWORK_COMPONENT_DSO_SUBDIRS='$(MCA_opal_common_DSO_SUBDIRS) > $(MCA_opal_allocator_DSO_SUBDIRS) > $(MCA_opal_backtrace_DSO_SUBDIRS) > $(MCA_opal_btl_DSO_SUBDIRS) $(MCA_opal_dl_DSO_SUBDIRS) > $(MCA_opal_event_DSO_SUBDIRS) $(MCA_opal_hwloc_DSO_SUBDIRS) > $(MCA_opal_if_DSO_SUBDIRS) $(MCA_opal_installdirs_DSO_SUBDIRS) > $(MCA_opal_memchecker_DSO_SUBDIRS) $(MCA_opal_memcpy_DSO_SUBDIRS) > $(MCA_opal_memory_DSO_SUBDIRS) $(MCA_opal_mpool_DSO_SUBDIRS) > $(MCA_opal_pmix_DSO_SUBDIRS) $(MCA_opal_pstat_DSO_SUBDIRS) > $(MCA_opal_rcache_DSO_SUBDIRS) $(MCA_opal_sec_DSO_SUBDIRS) > $(MCA_opal_shmem_DSO_SUBDIRS) $(MCA_opal_timer_DSO_SUBDIRS)' > > MCA_opal_FRAMEWORK_COMPONENT_STATIC_SUBDIRS='$(MCA_opal_common_STATIC_SUBDIRS) > $(MCA_opal_allocator_STATIC_SUBDIRS) > $(MCA_opal_backtrace_STATIC_SUBDIRS) > $(MCA_opal_btl_STATIC_SUBDIRS) $(MCA_opal_dl_STATIC_SUBDIRS) > $(MCA_opal_event_STATIC_SUBDIRS) $(MCA_opal_hwloc_STATIC_SUBDIRS) > $(MCA_opal_if_STATIC_SUBDIRS) > $(MCA_opal_installdirs_STATIC_SUBDIRS) > $(MCA_opal_memchecker_STATIC_SUBDIRS) > $(MCA_opal_memcpy_STATIC_SUBDIRS) > $(MCA_opal_memory_STATIC_SUBDIRS) $(MCA_opal_mpool_STATIC_SUBDIRS) > $(MCA_opal_pmix_STATIC_SUBDIRS) $(MCA_opal_pstat_STATIC_SUBDIRS) > $(MCA_opal_rcache_STATIC_SUBDIRS) $(MCA_opal_sec_STATIC_SUBDIRS) > $(MCA_opal_shmem_STATIC_SUBDIRS) $(MCA_opal_timer_STATIC_SUBDIRS)' > MCA_opal_FRAMEWORK_LIBS=' $(MCA_opal_common_STATIC_LTLIBS) > mca/allocator/libmca_allocator.la <http://libmca_allocator.la> > $(MCA_opal_allocator_STATIC_LTLIBS) > mca/backtrace/libmca_backtrace.la <http://libmca_backtrace.la> > $(MCA_opal_backtrace_STATIC_LTLIBS) > mca/btl/libmca_btl.la <http://libmca_btl.la> > $(MCA_opal_btl_STATIC_LTLIBS) mca/dl/libmca_dl.la > <http://libmca_dl.la> > $(MCA_opal_dl_STATIC_LTLIBS) mca/event/libmca_event.la > <http://libmca_event.la> > $(MCA_opal_event_STATIC_LTLIBS) mca/hwloc/libmca_hwloc.la > <http://libmca_hwloc.la> > $(MCA_opal_hwloc_STATIC_LTLIBS) mca/if/libmca_if.la > <http://libmca_if.la> > $(MCA_opal_if_STATIC_LTLIBS) mca/installdirs/libmca_installdirs.la > <http://libmca_installdirs.la> > $(MCA_opal_installdirs_STATIC_LTLIBS) > mca/memchecker/libmca_memchecker.la <http://libmca_memchecker.la> > $(MCA_opal_memchecker_STATIC_LTLIBS) mca/memcpy/libmca_memcpy.la > <http://libmca_memcpy.la> > $(MCA_opal_memcpy_STATIC_LTLIBS) mca/memory/libmca_memory.la > <http://libmca_memory.la> > $(MCA_opal_memory_STATIC_LTLIBS) mca/mpool/libmca_mpool.la > <http://libmca_mpool.la> > $(MCA_opal_mpool_STATIC_LTLIBS) mca/pmix/libmca_pmix.la > <http://libmca_pmix.la> > $(MCA_opal_pmix_STATIC_LTLIBS) mca/pstat/libmca_pstat.la > <http://libmca_pstat.la> > $(MCA_opal_pstat_STATIC_LTLIBS) mca/rcache/libmca_rcache.la > <http://libmca_rcache.la> > $(MCA_opal_rcache_STATIC_LTLIBS) mca/sec/libmca_sec.la > <http://libmca_sec.la> > $(MCA_opal_sec_STATIC_LTLIBS) mca/shmem/libmca_shmem.la > <http://libmca_shmem.la> > $(MCA_opal_shmem_STATIC_LTLIBS) mca/timer/libmca_timer.la > <http://libmca_timer.la> > $(MCA_opal_timer_STATIC_LTLIBS)' > ... > MCA_opal_pmix_ALL_COMPONENTS=' s1 cray s2 pmix112 external' > MCA_opal_pmix_ALL_SUBDIRS=' mca/pmix/s1 mca/pmix/cray mca/pmix/s2 > mca/pmix/pmix112 mca/pmix/external' > MCA_opal_pmix_DSO_COMPONENTS='' > MCA_opal_pmix_DSO_SUBDIRS='' > MCA_opal_pmix_STATIC_COMPONENTS='' > MCA_opal_pmix_STATIC_LTLIBS='' > MCA_opal_pmix_STATIC_SUBDIRS='' > ... > opal_pmix_ext_CPPFLAGS='' > opal_pmix_ext_LDFLAGS='' > opal_pmix_ext_LIBS='' > opal_pmix_pmix112_CPPFLAGS='' > opal_pmix_pmix112_LIBS='' > ... > > > > > I've attached the config.log files for pmix. > > tyr openmpi-2.0.0 142 tar zvft pmix_config.log.tar.gz > -rw-r--r-- root/root 136291 2016-04-25 08:05:34 > > openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/pmix/pmix112/pmix/config.log > -rw-r--r-- root/root 528808 2016-04-25 08:07:54 > > openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_gcc/opal/mca/pmix/pmix112/pmix/config.log > tyr openmpi-2.0.0 143 > > > > I've also attached the output for the broken execution of > "spawn_multiple_master" for my gcc-version of Open MPI. > "spawn_master" works as expected with my gcc-version of Open MPI. > > Hopefully you can fix the problem. > > > Kind regards and thank you very much for your help > > Siegmar > > > > Am 23.04.2016 um 21:34 schrieb Siegmar Gross: > > Hi Gilles, > > I don't know what happened, but the files are not available > now > and they were definitely available when I answered the email > from > Ralph. The files also have a different timestamp now. This is > an > extract from my email to Ralph for Solaris Sparc. > > -rwxr-xr-x 1 root root 977 Apr 19 19:49 mca_plm_rsh.la > <http://mca_plm_rsh.la> > -rwxr-xr-x 1 root root 153280 Apr 19 19:49 mca_plm_rsh.so > -rwxr-xr-x 1 root root 1007 Apr 19 19:47 > mca_pmix_pmix112.la <http://mca_pmix_pmix112.la> > -rwxr-xr-x 1 root root 1400512 Apr 19 19:47 > mca_pmix_pmix112.so > -rwxr-xr-x 1 root root 971 Apr 19 19:52 mca_pml_cm.la > <http://mca_pml_cm.la> > -rwxr-xr-x 1 root root 342440 Apr 19 19:52 mca_pml_cm.so > > Now I have the following output for these files. > > -rwxr-xr-x 1 root root 976 Apr 19 19:58 mca_plm_rsh.la > <http://mca_plm_rsh.la> > -rwxr-xr-x 1 root root 319816 Apr 19 19:58 mca_plm_rsh.so > -rwxr-xr-x 1 root root 970 Apr 19 20:00 mca_pml_cm.la > <http://mca_pml_cm.la> > -rwxr-xr-x 1 root root 1507440 Apr 19 20:00 mca_pml_cm.so > > I'll try to find out what happened next week when I'm back in > my office. > > > Kind regards > > Siegmar > > > > > > Am 23.04.16 um 02:12 schrieb Gilles Gouaillardet: > > Siegmar, > > I will try to reproduce this on my solaris11 x86_64 vm > > In the mean time, can you please double check > mca_pmix_pmix_pmix112.so > is a 64 bits library ? > (E.g, confirm "-m64" was correctly passed to pmix) > > Cheers, > > Gilles > > On Friday, April 22, 2016, Siegmar Gross > <siegmar.gr...@informatik.hs-fulda.de > <mailto:siegmar.gr...@informatik.hs-fulda.de>> wrote: > > Hi Ralph, > > I've already used "-enable-debug". "SYSTEM_ENV" is > "SunOS" or > "Linux" and "MACHINE_ENV" is "sparc" or "x86_84". > > mkdir > > openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_gcc > cd > > openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_gcc > > ../openmpi-v2.x-dev-1280-gc110ae8/configure \ > --prefix=/usr/local/openmpi-2.0.0_64_gcc \ > --libdir=/usr/local/openmpi-2.0.0_64_gcc/lib64 \ > --with-jdk-bindir=/usr/local/jdk1.8.0/bin \ > --with-jdk-headers=/usr/local/jdk1.8.0/include \ > JAVA_HOME=/usr/local/jdk1.8.0 \ > LDFLAGS="-m64" CC="gcc" CXX="g++" FC="gfortran" \ > CFLAGS="-m64" CXXFLAGS="-m64" FCFLAGS="-m64" \ > CPP="cpp" CXXCPP="cpp" \ > --enable-mpi-cxx \ > --enable-cxx-exceptions \ > --enable-mpi-java \ > --enable-heterogeneous \ > --enable-mpi-thread-multiple \ > --with-hwloc=internal \ > --without-verbs \ > --with-wrapper-cflags="-std=c11 -m64" \ > --with-wrapper-cxxflags="-m64" \ > --with-wrapper-fcflags="-m64" \ > --enable-debug \ > |& tee log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_gcc > > > mkdir > > openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_cc > cd > > openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_cc > > ../openmpi-v2.x-dev-1280-gc110ae8/configure \ > --prefix=/usr/local/openmpi-2.0.0_64_cc \ > --libdir=/usr/local/openmpi-2.0.0_64_cc/lib64 \ > --with-jdk-bindir=/usr/local/jdk1.8.0/bin \ > --with-jdk-headers=/usr/local/jdk1.8.0/include \ > JAVA_HOME=/usr/local/jdk1.8.0 \ > LDFLAGS="-m64" CC="cc" CXX="CC" FC="f95" \ > CFLAGS="-m64" CXXFLAGS="-m64 -library=stlport4" > FCFLAGS="-m64" \ > CPP="cpp" CXXCPP="cpp" \ > --enable-mpi-cxx \ > --enable-cxx-exceptions \ > --enable-mpi-java \ > --enable-heterogeneous \ > --enable-mpi-thread-multiple \ > --with-hwloc=internal \ > --without-verbs \ > --with-wrapper-cflags="-m64" \ > --with-wrapper-cxxflags="-m64 -library=stlport4" \ > --with-wrapper-fcflags="-m64" \ > --with-wrapper-ldflags="" \ > --enable-debug \ > |& tee log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_cc > > > Kind regards > > Siegmar > > Am 21.04.2016 um 18:18 schrieb Ralph Castain: > > Can you please rebuild OMPI with -enable-debug in > the configure > cmd? It will let us see more error output > > > On Apr 21, 2016, at 8:52 AM, Siegmar Gross > <siegmar.gr...@informatik.hs-fulda.de> wrote: > > Hi Ralph, > > I don't see any additional information. > > tyr hello_1 108 mpiexec -np 4 --host > tyr,sunpc1,linpc1,ruester -mca > mca_base_component_show_load_errors 1 > hello_1_mpi > [tyr.informatik.hs-fulda.de:06211 > <http://tyr.informatik.hs-fulda.de:06211> > <http://tyr.informatik.hs-fulda.de:06211>] > [[48741,0],0] > ORTE_ERROR_LOG: Not found in file > > > ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c > > at line 638 > > > -------------------------------------------------------------------------- > It looks like orte_init failed for some > reason; your > parallel process is > likely to abort. There are many reasons that > a parallel > process can > fail during orte_init; some of which are due > to configuration or > environment problems. This failure appears to > be an > internal failure; > here's some additional information (which may > only be > relevant to an > Open MPI developer): > > opal_pmix_base_select failed > --> Returned value Not found (-13) instead of > ORTE_SUCCESS > > > -------------------------------------------------------------------------- > > > tyr hello_1 109 mpiexec -np 4 --host > tyr,sunpc1,linpc1,ruester -mca > mca_base_component_show_load_errors 1 -mca > pmix_base_verbose > 10 -mca pmix_server_verbose 5 hello_1_mpi > [tyr.informatik.hs-fulda.de:06212 > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212>] > mca: base: > components_register: registering framework > pmix components > [tyr.informatik.hs-fulda.de:06212 > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212>] > mca: base: > components_open: opening pmix components > [tyr.informatik.hs-fulda.de:06212 > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212>] > mca:base:select: > Auto-selecting pmix components > [tyr.informatik.hs-fulda.de:06212 > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212>] > mca:base:select:( > pmix) No component selected! > [tyr.informatik.hs-fulda.de:06212 > <http://tyr.informatik.hs-fulda.de:06212> > <http://tyr.informatik.hs-fulda.de:06212>] > [[48738,0],0] > ORTE_ERROR_LOG: Not found in file > > > ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c > > at line 638 > > > -------------------------------------------------------------------------- > It looks like orte_init failed for some > reason; your > parallel process is > likely to abort. There are many reasons that > a parallel > process can > fail during orte_init; some of which are due > to configuration or > environment problems. This failure appears to > be an > internal failure; > here's some additional information (which may > only be > relevant to an > Open MPI developer): > > opal_pmix_base_select failed > --> Returned value Not found (-13) instead of > ORTE_SUCCESS > > > -------------------------------------------------------------------------- > tyr hello_1 110 > > > Kind regards > > Siegmar > > > Am 21.04.2016 um 17:24 schrieb Ralph Castain: > > Hmmm?it looks like you built the right > components, but > they are not being picked up. Can you run > your mpiexec > command again, adding ?-mca > mca_base_component_show_load_errors 1? to > the cmd line? > > > On Apr 21, 2016, at 8:16 AM, Siegmar > Gross > <siegmar.gr...@informatik.hs-fulda.de> > wrote: > > Hi Ralph, > > I have attached ompi_info output for > both compilers > from my > sparc machine and the listings for > both compilers > from the > <prefix>/lib/openmpi directories. > Hopefully that > helps to > find the problem. > > hermes tmp 3 tar zvft > openmpi-2.x_info.tar.gz > -rw-r--r-- root/root 10969 > 2016-04-21 17:06 > ompi_info_SunOS_sparc_cc.txt > -rw-r--r-- root/root 11044 > 2016-04-21 17:06 > ompi_info_SunOS_sparc_gcc.txt > -rw-r--r-- root/root 71252 > 2016-04-21 17:02 > lib64_openmpi.txt > hermes tmp 4 > > > Kind regards and thank you very much > once more for > your help > > Siegmar > > > Am 21.04.2016 um 15:54 schrieb Ralph > Castain: > > Odd - it would appear that none of > the pmix > components built? Can you send > along the output from ompi_info? > Or just send a > listing of the files in the > <prefix>/lib/openmpi directory? > > > On Apr 21, 2016, at 1:27 AM, > Siegmar Gross > > <siegmar.gr...@informatik.hs-fulda.de > > <mailto:siegmar.gr...@informatik.hs-fulda.de>> > wrote: > > Hi Ralph, > > Am 21.04.2016 um 00:18 schrieb > Ralph Castain: > > Could you please rerun > these test and > add ?-mca > pmix_base_verbose 10 > -mca pmix_server_verbose > 5? to your cmd > line? I need to see why > the > pmix components failed. > > > > tyr spawn 111 mpiexec -np 1 > --host > tyr,sunpc1,linpc1,ruester -mca > pmix_base_verbose 10 -mca > pmix_server_verbose 5 > spawn_multiple_master > [tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/>:26652] mca: > base: components_register: > registering > framework pmix components > [tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/>:26652] mca: > base: components_open: opening > pmix components > [tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/>:26652] > mca:base:select: > Auto-selecting pmix components > [tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/>:26652] > mca:base:select:( pmix) No > component selected! > [tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/>:26652] > [[52794,0],0] ORTE_ERROR_LOG: > Not found in file > > > ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c > > at line 638 > > > -------------------------------------------------------------------------- > It looks like orte_init failed > for some > reason; your parallel process > is > likely to abort. There are > many reasons > that a parallel process can > fail during orte_init; some of > which are due > to configuration or > environment problems. This > failure appears > to be an internal failure; > here's some additional > information (which > may only be relevant to an > Open MPI developer): > > opal_pmix_base_select failed > --> Returned value Not found > (-13) instead > of ORTE_SUCCESS > > > -------------------------------------------------------------------------- > tyr spawn 112 > > > > > tyr hello_1 116 mpiexec -np 1 > --host > tyr,sunpc1,linpc1,ruester -mca > pmix_base_verbose 10 -mca > pmix_server_verbose 5 > hello_1_mpi > [tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/>:27261] mca: > base: components_register: > registering > framework pmix components > [tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/>:27261] mca: > base: components_open: opening > pmix components > [tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/>:27261] > mca:base:select: > Auto-selecting pmix components > [tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/>:27261] > mca:base:select:( pmix) No > component selected! > [tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de/>:27261] > [[52315,0],0] ORTE_ERROR_LOG: > Not found in file > > > ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c > > at line 638 > > > -------------------------------------------------------------------------- > It looks like orte_init failed > for some > reason; your parallel process > is > likely to abort. There are > many reasons > that a parallel process can > fail during orte_init; some of > which are due > to configuration or > environment problems. This > failure appears > to be an internal failure; > here's some additional > information (which > may only be relevant to an > Open MPI developer): > > opal_pmix_base_select failed > --> Returned value Not found > (-13) instead > of ORTE_SUCCESS > > > -------------------------------------------------------------------------- > tyr hello_1 117 > > > > Thank you very much for your > help. > > > Kind regards > > Siegmar > > > > > Thanks > Ralph > > On Apr 20, 2016, at > 10:12 AM, > Siegmar Gross > > <siegmar.gr...@informatik.hs-fulda.de > > <mailto:siegmar.gr...@informatik.hs-fulda.de>> > wrote: > > Hi, > > I have built > > openmpi-v2.x-dev-1280-gc110ae8 on my > machines > (Solaris 10 Sparc, > Solaris 10 > x86_64, and openSUSE > Linux > 12.1 x86_64) with > gcc-5.1.0 and Sun > C 5.13. Unfortunately > I get > runtime errors for > some programs. > > > Sun C 5.13: > =========== > > For all my test > programs I get the > same error on Solaris > Sparc and > Solaris x86_64, while > the programs > work fine on Linux. > > tyr hello_1 115 > mpiexec -np 2 > hello_1_mpi > > [tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de>:22373] > [[61763,0],0] > ORTE_ERROR_LOG: Not > found in file > > > ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c > > at line 638 > > > -------------------------------------------------------------------------- > It looks like > orte_init failed for > some reason; your > parallel process is > likely to abort. > There are many > reasons that a > parallel process can > fail during orte_init; > some of which > are due to > configuration or > environment problems. > This failure > appears to be an > internal failure; > here's some additional > information > (which may only be > relevant to an > Open MPI developer): > > opal_pmix_base_select > failed > --> Returned value Not > found (-13) > instead of > ORTE_SUCCESS > > > -------------------------------------------------------------------------- > tyr hello_1 116 > > > > > GCC-5.1.0: > ========== > > tyr spawn 121 mpiexec > -np 1 --host > > tyr,sunpc1,linpc1,ruester > spawn_multiple_master > > Parent process 0 > running on > > tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > I create 3 slave > processes. > > > [tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de>:25366] > PMIX ERROR: > UNPACK-PAST-END in file > > > ../../../../../../openmpi-v2.x-dev-1280-gc110ae8/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c > > > at line 829 > > [tyr.informatik.hs-fulda.de > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de> > > <http://tyr.informatik.hs-fulda.de>:25366] > PMIX ERROR: > UNPACK-PAST-END in file > > > ../../../../../../openmpi-v2.x-dev-1280-gc110ae8/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c > > > at line 2176 > [tyr:25377] *** An > error occurred in > > MPI_Comm_spawn_multiple > [tyr:25377] *** > reported by process > [3308257281,0] > [tyr:25377] *** on > communicator > MPI_COMM_WORLD > [tyr:25377] *** > MPI_ERR_SPAWN: could > not spawn processes > [tyr:25377] *** > MPI_ERRORS_ARE_FATAL > (processes in this > communicator will > now abort, > [tyr:25377] *** and > potentially > your MPI job) > tyr spawn 122 > > > I would be grateful if > somebody can > fix the problems. > Thank you very > much for any help in > advance. > > > Kind regards > > Siegmar > > > <hello_1_mpi.c><spawn_multiple_master.c>_______________________________________________ > > > users mailing list > us...@open-mpi.org > <mailto:us...@open-mpi.org> > Subscription: > > http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > > > http://www.open-mpi.org/community/lists/users/2016/04/28983.php > > > > _______________________________________________ > users mailing list > us...@open-mpi.org > <mailto:us...@open-mpi.org> > Subscription: > > http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this > post: > > > http://www.open-mpi.org/community/lists/users/2016/04/28986.php > > > > _______________________________________________ > users mailing list > us...@open-mpi.org > <mailto:us...@open-mpi.org> > Subscription: > > http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this > post: > > > http://www.open-mpi.org/community/lists/users/2016/04/28987.php > > > > > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: > > http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > > > http://www.open-mpi.org/community/lists/users/2016/04/28988.php > > > > <openmpi-2.x_info.tar.gz>_______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: > > http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > > > http://www.open-mpi.org/community/lists/users/2016/04/28989.php > > > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: > > http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > > > http://www.open-mpi.org/community/lists/users/2016/04/28990.php > > > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: > http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > > > http://www.open-mpi.org/community/lists/users/2016/04/28991.php > > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: > http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > > > http://www.open-mpi.org/community/lists/users/2016/04/28992.php > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: > http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > > > http://www.open-mpi.org/community/lists/users/2016/04/28993.php > > > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: > http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > > http://www.open-mpi.org/community/lists/users/2016/04/28999.php > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: > http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > > http://www.open-mpi.org/community/lists/users/2016/04/29009.php > > > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > http://www.open-mpi.org/community/lists/users/2016/04/29033.php > > > > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > http://www.open-mpi.org/community/lists/users/2016/04/29038.php -------------- next part -------------- A non-text attachment was scrubbed... Name: spawn_multiple_master.c Type: text/x-csrc Size: 7819 bytes Desc: not available URL: <http://www.open-mpi.org/MailArchives/users/attachments/20160427/cf27b14b/attachment.bin> ------------------------------ Message: 2 List-Post: users@lists.open-mpi.org Date: Thu, 28 Apr 2016 10:34:09 +0900 From: Gilles Gouaillardet <gil...@rist.or.jp> To: Open MPI Users <us...@open-mpi.org> Subject: Re: [OMPI users] runtime errors for openmpi-v2.x-dev-1280-gc110ae8 Message-ID: <c173ccda-4979-d90c-a8ae-99f588e26...@rist.or.jp> Content-Type: text/plain; charset="windows-1252"; Format="flowed" Siegmar, can you please also post the source of spawn_slave ? Cheers, Gilles On 4/28/2016 1:17 AM, Siegmar Gross wrote: > Hi Gilles, > > it is not necessary to have a heterogeneous environment to reproduce > the error as you can see below. All machines are 64 bit. > > tyr spawn 119 ompi_info | grep -e "OPAL repo revision" -e "C compiler > absolute" > OPAL repo revision: v2.x-dev-1290-gbd0e4e1 > C compiler absolute: /usr/local/gcc-5.1.0/bin/gcc > tyr spawn 120 uname -a > SunOS tyr.informatik.hs-fulda.de 5.10 Generic_150400-11 sun4u sparc > SUNW,A70 Solaris > tyr spawn 121 mpiexec -np 1 --host tyr,tyr,tyr,tyr spawn_multiple_master > > Parent process 0 running on tyr.informatik.hs-fulda.de > I create 3 slave processes. > > [tyr.informatik.hs-fulda.de:27286] PMIX ERROR: UNPACK-PAST-END in file > ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c > > at line 829 > [tyr.informatik.hs-fulda.de:27286] PMIX ERROR: UNPACK-PAST-END in file > ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c > > at line 2176 > [tyr:27288] *** An error occurred in MPI_Comm_spawn_multiple > [tyr:27288] *** reported by process [3434086401,0] > [tyr:27288] *** on communicator MPI_COMM_WORLD > [tyr:27288] *** MPI_ERR_SPAWN: could not spawn processes > [tyr:27288] *** MPI_ERRORS_ARE_FATAL (processes in this communicator > will now abort, > [tyr:27288] *** and potentially your MPI job) > tyr spawn 122 > > > > > > > sunpc1 fd1026 105 ompi_info | grep -e "OPAL repo revision" -e "C > compiler absolute" > OPAL repo revision: v2.x-dev-1290-gbd0e4e1 > C compiler absolute: /usr/local/gcc-5.1.0/bin/gcc > sunpc1 fd1026 106 uname -a > SunOS sunpc1 5.10 Generic_147441-21 i86pc i386 i86pc Solaris > sunpc1 fd1026 107 mpiexec -np 1 --host sunpc1,sunpc1,sunpc1,sunpc1 > spawn_multiple_master > > Parent process 0 running on sunpc1 > I create 3 slave processes. > > [sunpc1:00368] PMIX ERROR: UNPACK-PAST-END in file > ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c > > at line 829 > [sunpc1:00368] PMIX ERROR: UNPACK-PAST-END in file > ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c > > at line 2176 > [sunpc1:370] *** An error occurred in MPI_Comm_spawn_multiple > [sunpc1:370] *** reported by process [43909121,0] > [sunpc1:370] *** on communicator MPI_COMM_WORLD > [sunpc1:370] *** MPI_ERR_SPAWN: could not spawn processes > [sunpc1:370] *** MPI_ERRORS_ARE_FATAL (processes in this communicator > will now abort, > [sunpc1:370] *** and potentially your MPI job) > sunpc1 fd1026 108 > > > > > > linpc1 fd1026 105 ompi_info | grep -e "OPAL repo revision" -e "C > compiler absolute" > OPAL repo revision: v2.x-dev-1290-gbd0e4e1 > C compiler absolute: /usr/local/gcc-5.1.0/bin/gcc > linpc1 fd1026 106 uname -a > Linux linpc1 3.1.10-1.29-desktop #1 SMP PREEMPT Fri May 31 20:10:04 > UTC 2013 (2529847) x86_64 x86_64 x86_64 GNU/Linux > linpc1 fd1026 107 mpiexec -np 1 --host linpc1,linpc1,linpc1,linpc1 > spawn_multiple_master > > Parent process 0 running on linpc1 > I create 3 slave processes. > > [linpc1:21502] PMIX ERROR: UNPACK-PAST-END in file > ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c > > at line 829 > [linpc1:21502] PMIX ERROR: UNPACK-PAST-END in file > ../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c > > at line 2176 > [linpc1:21507] *** An error occurred in MPI_Comm_spawn_multiple > [linpc1:21507] *** reported by process [1005518849,0] > [linpc1:21507] *** on communicator MPI_COMM_WORLD > [linpc1:21507] *** MPI_ERR_SPAWN: could not spawn processes > [linpc1:21507] *** MPI_ERRORS_ARE_FATAL (processes in this > communicator will now abort, > [linpc1:21507] *** and potentially your MPI job) > linpc1 fd1026 108 > > > I used the following configure command. > > ../openmpi-v2.x-dev-1290-gbd0e4e1/configure \ > --prefix=/usr/local/openmpi-2.0.0_64_gcc \ > --libdir=/usr/local/openmpi-2.0.0_64_gcc/lib64 \ > --with-jdk-bindir=/usr/local/jdk1.8.0/bin \ > --with-jdk-headers=/usr/local/jdk1.8.0/include \ > JAVA_HOME=/usr/local/jdk1.8.0 \ > LDFLAGS="-m64" CC="gcc" CXX="g++" FC="gfortran" \ > CFLAGS="-m64" CXXFLAGS="-m64" FCFLAGS="-m64" \ > CPP="cpp" CXXCPP="cpp" \ > --enable-mpi-cxx \ > --enable-cxx-exceptions \ > --enable-mpi-java \ > --enable-heterogeneous \ > --enable-mpi-thread-multiple \ > --with-hwloc=internal \ > --without-verbs \ > --with-wrapper-cflags="-std=c11 -m64" \ > --with-wrapper-cxxflags="-m64" \ > --with-wrapper-fcflags="-m64" \ > --enable-debug \ > |& tee log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_gcc > > > Kind regards > > Siegmar > > > > Am 27.04.2016 um 13:21 schrieb Gilles Gouaillardet: >> Siegmar, >> >> please add this to your CFLAGS for the time being. >> >> configure tries to detect which flags must be added for C99 support, >> and it seems >> the test is not working for Solaris 10 and Oracle compilers. >> this is no more a widely used environment, and I am not sure I can >> find the >> time to fix this >> in a near future. >> >> >> regarding the runtime issue, can you please describe your 4 hosts (os, >> endianness and bitness) >> >> Cheers, >> >> Gilles >> >> On Wednesday, April 27, 2016, Siegmar Gross >> <siegmar.gr...@informatik.hs-fulda.de >> <javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de');>> >> wrote: >> >> Hi Gilles, >> >> adding "-std=c99" to CFLAGS solves the problem with the missing >> library. >> Shall I add it permanently to my configure command or will you >> add it, >> so that I will not run into problems if you need the C11 standard >> later? >> >> "spawn_multiple_master" breaks with the same error that I reported >> yesterday for my gcc-version of Open MPI. Hopefully you can solve >> the >> problem as well. >> >> >> Kind regards and thank you very much for your help >> >> Siegmar >> >> >> Am 27.04.2016 um 08:05 schrieb Gilles Gouaillardet: >> >> Siegmar, >> >> >> here is the error : >> >> configure:17969: cc -o conftest -m64 -D_REENTRANT -g -g >> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1 >> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc >> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/include >> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/include >> -D_REENTRANT >> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/hwloc/hwloc1112/hwloc/include >> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/hwloc/hwloc1112/hwloc/include >> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/event/libevent2022/libevent >> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/event/libevent2022/libevent/include >> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/event/libevent2022/libevent/include >> -m64 conftest.c >&5 >> "/usr/include/stdbool.h", line 42: #error: "Use of >> <stdbool.h> is >> valid only >> in a c99 compilation environment." >> >> >> i cannot reproduce this on solaris 11 with oracle studio 5.3 >> compiler, >> and i >> do not have solaris 10 yet. >> >> could you please re-configure with '-std=c99' appended to >> your CFLAGS >> and see >> if it helps ? >> >> >> Cheers, >> >> >> Gilles >> >> >> On 4/26/2016 7:57 PM, Siegmar Gross wrote: >> >> Hi Gilles and Ralph, >> >> I was able to sort out my mess. In my last email I >> compared the >> files from >> "SunOS_sparc/openmpi-2.0.0_64_gcc/lib64/openmpi" from >> the attachment of my email to Ralph with the files from >> "SunOS_sparc/openmpi-2.0.0_64_cc/lib64/openmpi" from my >> current >> file system. That's the reason while I have had different >> timestamps. The other problem was that Ralph didn't >> recognize >> that "mca_pmix_pmix112.so" wasn't built on Solaris with the >> Sun C compiler. I've removed most of the files from the >> attachment >> of my email so that it is easier to see the relevant >> files. Below >> I try to give you more information that may be relevant >> to track >> down the problem. I still get an error running one of my >> small >> test programs, when I use my gcc-version of Open MPI. >> "mca_pmix_pmix112.so" is a 64 bits library. >> >> Linux_x86_64/openmpi-2.0.0_64_cc/lib64/openmpi: >> ... >> -rwxr-xr-x 1 root root 261327 Apr 19 16:46 mca_plm_slurm.so >> -rwxr-xr-x 1 root root 1002 Apr 19 16:45 >> mca_pmix_pmix112.la >> <http://mca_pmix_pmix112.la> >> -rwxr-xr-x 1 root root 3906526 Apr 19 16:45 >> mca_pmix_pmix112.so >> -rwxr-xr-x 1 root root 966 Apr 19 16:51 mca_pml_cm.la >> <http://mca_pml_cm.la> >> -rwxr-xr-x 1 root root 1574265 Apr 19 16:51 mca_pml_cm.so >> ... >> >> Linux_x86_64/openmpi-2.0.0_64_gcc/lib64/openmpi: >> ... >> -rwxr-xr-x 1 root root 70371 Apr 19 16:43 mca_plm_slurm.so >> -rwxr-xr-x 1 root root 1008 Apr 19 16:42 >> mca_pmix_pmix112.la >> <http://mca_pmix_pmix112.la> >> -rwxr-xr-x 1 root root 1029005 Apr 19 16:42 >> mca_pmix_pmix112.so >> -rwxr-xr-x 1 root root 972 Apr 19 16:46 mca_pml_cm.la >> <http://mca_pml_cm.la> >> -rwxr-xr-x 1 root root 284858 Apr 19 16:46 mca_pml_cm.so >> ... >> >> SunOS_sparc/openmpi-2.0.0_64_cc/lib64/openmpi: >> ... >> -rwxr-xr-x 1 root root 319816 Apr 19 19:58 mca_plm_rsh.so >> -rwxr-xr-x 1 root root 970 Apr 19 20:00 mca_pml_cm.la >> <http://mca_pml_cm.la> >> -rwxr-xr-x 1 root root 1507440 Apr 19 20:00 mca_pml_cm.so >> ... >> >> SunOS_sparc/openmpi-2.0.0_64_gcc/lib64/openmpi: >> ... >> -rwxr-xr-x 1 root root 153280 Apr 19 19:49 mca_plm_rsh.so >> -rwxr-xr-x 1 root root 1007 Apr 19 19:47 >> mca_pmix_pmix112.la >> <http://mca_pmix_pmix112.la> >> -rwxr-xr-x 1 root root 1400512 Apr 19 19:47 >> mca_pmix_pmix112.so >> -rwxr-xr-x 1 root root 971 Apr 19 19:52 mca_pml_cm.la >> <http://mca_pml_cm.la> >> -rwxr-xr-x 1 root root 342440 Apr 19 19:52 mca_pml_cm.so >> ... >> >> SunOS_x86_64/openmpi-2.0.0_64_cc/lib64/openmpi: >> ... >> -rwxr-xr-x 1 root root 300096 Apr 19 17:18 mca_plm_rsh.so >> -rwxr-xr-x 1 root root 970 Apr 19 17:23 mca_pml_cm.la >> <http://mca_pml_cm.la> >> -rwxr-xr-x 1 root root 1458816 Apr 19 17:23 mca_pml_cm.so >> ... >> >> SunOS_x86_64/openmpi-2.0.0_64_gcc/lib64/openmpi: >> ... >> -rwxr-xr-x 1 root root 133096 Apr 19 17:42 mca_plm_rsh.so >> -rwxr-xr-x 1 root root 1007 Apr 19 17:41 >> mca_pmix_pmix112.la >> <http://mca_pmix_pmix112.la> >> -rwxr-xr-x 1 root root 1320240 Apr 19 17:41 >> mca_pmix_pmix112.so >> -rwxr-xr-x 1 root root 971 Apr 19 17:46 mca_pml_cm.la >> <http://mca_pml_cm.la> >> -rwxr-xr-x 1 root root 419848 Apr 19 17:46 mca_pml_cm.so >> ... >> >> >> Yesterday I've installed openmpi-v2.x-dev-1290-gbd0e4e1 >> so that we >> have a current version for the investigation of the >> problem. Once >> more mca_pmix_pmix112.so isn't available on Solaris if I >> use the >> Sun C compiler. >> >> "config.log" for gcc-5.1.0 shows the following. >> >> ... >> configure:127799: /bin/bash >> '../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/ >> pmix/configure' succeeded for opal/mca/pmix/pmix112/pmix >> configure:127916: checking if MCA component pmix:pmix112 >> can compile >> configure:127918: result: yes >> configure:5637: --- MCA component pmix:external (m4 >> configuration >> macro) >> configure:128523: checking for MCA component >> pmix:external compile >> mode >> configure:128529: result: dso >> configure:129054: checking if MCA component pmix:external >> can compile >> configure:129056: result: no >> ... >> config.status:3897: creating opal/mca/pmix/Makefile >> config.status:3897: creating opal/mca/pmix/s1/Makefile >> config.status:3897: creating opal/mca/pmix/cray/Makefile >> config.status:3897: creating opal/mca/pmix/s2/Makefile >> config.status:3897: creating opal/mca/pmix/pmix112/Makefile >> config.status:3897: creating opal/mca/pmix/external/Makefile >> ... >> MCA_BUILD_opal_pmix_cray_DSO_FALSE='#' >> MCA_BUILD_opal_pmix_cray_DSO_TRUE='' >> MCA_BUILD_opal_pmix_external_DSO_FALSE='#' >> MCA_BUILD_opal_pmix_external_DSO_TRUE='' >> MCA_BUILD_opal_pmix_pmix112_DSO_FALSE='#' >> MCA_BUILD_opal_pmix_pmix112_DSO_TRUE='' >> MCA_BUILD_opal_pmix_s1_DSO_FALSE='#' >> MCA_BUILD_opal_pmix_s1_DSO_TRUE='' >> MCA_BUILD_opal_pmix_s2_DSO_FALSE='#' >> MCA_BUILD_opal_pmix_s2_DSO_TRUE='' >> ... >> MCA_opal_FRAMEWORKS='common allocator backtrace btl dl >> event hwloc if >> installdirs memchecker memcpy memory mpool pmix pstat >> rcache sec >> shmem timer' >> MCA_opal_FRAMEWORKS_SUBDIRS='mca/common mca/allocator >> mca/backtrace mca/btl >> mca/dl mca/event mca/hwloc mca/if mca/installdirs >> mca/memchecker >> mca/memcpy >> mca/memory mca/mpool mca/pmix mca/pstat mca/rcache mca/sec >> mca/shmem mca/timer' >> MCA_opal_FRAMEWORK_COMPONENT_ALL_SUBDIRS='$(MCA_opal_common_ALL_SUBDIRS) >> $(MCA_opal_allocator_ALL_SUBDIRS) >> $(MCA_opal_backtrace_ALL_SUBDIRS) >> $(MCA_opal_btl_ALL_SUBDIRS) $(MCA_opal_dl_ALL_SUBDIRS) >> $(MCA_opal_event_ALL_SUBDIRS) $(MCA_opal_hwloc_ALL_SUBDIRS) >> $(MCA_opal_if_ALL_SUBDIRS) >> $(MCA_opal_installdirs_ALL_SUBDIRS) >> $(MCA_opal_memchecker_ALL_SUBDIRS) >> $(MCA_opal_memcpy_ALL_SUBDIRS) >> $(MCA_opal_memory_ALL_SUBDIRS) $(MCA_opal_mpool_ALL_SUBDIRS) >> $(MCA_opal_pmix_ALL_SUBDIRS) $(MCA_opal_pstat_ALL_SUBDIRS) >> $(MCA_opal_rcache_ALL_SUBDIRS) $(MCA_opal_sec_ALL_SUBDIRS) >> $(MCA_opal_shmem_ALL_SUBDIRS) $(MCA_opal_timer_ALL_SUBDIRS)' >> MCA_opal_FRAMEWORK_COMPONENT_DSO_SUBDIRS='$(MCA_opal_common_DSO_SUBDIRS) >> $(MCA_opal_allocator_DSO_SUBDIRS) >> $(MCA_opal_backtrace_DSO_SUBDIRS) >> $(MCA_opal_btl_DSO_SUBDIRS) $(MCA_opal_dl_DSO_SUBDIRS) >> $(MCA_opal_event_DSO_SUBDIRS) $(MCA_opal_hwloc_DSO_SUBDIRS) >> $(MCA_opal_if_DSO_SUBDIRS) >> $(MCA_opal_installdirs_DSO_SUBDIRS) >> $(MCA_opal_memchecker_DSO_SUBDIRS) >> $(MCA_opal_memcpy_DSO_SUBDIRS) >> $(MCA_opal_memory_DSO_SUBDIRS) $(MCA_opal_mpool_DSO_SUBDIRS) >> $(MCA_opal_pmix_DSO_SUBDIRS) $(MCA_opal_pstat_DSO_SUBDIRS) >> $(MCA_opal_rcache_DSO_SUBDIRS) $(MCA_opal_sec_DSO_SUBDIRS) >> $(MCA_opal_shmem_DSO_SUBDIRS) $(MCA_opal_timer_DSO_SUBDIRS)' >> MCA_opal_FRAMEWORK_COMPONENT_STATIC_SUBDIRS='$(MCA_opal_common_STATIC_SUBDIRS) >> $(MCA_opal_allocator_STATIC_SUBDIRS) >> $(MCA_opal_backtrace_STATIC_SUBDIRS) >> $(MCA_opal_btl_STATIC_SUBDIRS) $(MCA_opal_dl_STATIC_SUBDIRS) >> $(MCA_opal_event_STATIC_SUBDIRS) >> $(MCA_opal_hwloc_STATIC_SUBDIRS) >> $(MCA_opal_if_STATIC_SUBDIRS) >> $(MCA_opal_installdirs_STATIC_SUBDIRS) >> $(MCA_opal_memchecker_STATIC_SUBDIRS) >> $(MCA_opal_memcpy_STATIC_SUBDIRS) >> $(MCA_opal_memory_STATIC_SUBDIRS) >> $(MCA_opal_mpool_STATIC_SUBDIRS) >> $(MCA_opal_pmix_STATIC_SUBDIRS) >> $(MCA_opal_pstat_STATIC_SUBDIRS) >> $(MCA_opal_rcache_STATIC_SUBDIRS) >> $(MCA_opal_sec_STATIC_SUBDIRS) >> $(MCA_opal_shmem_STATIC_SUBDIRS) >> $(MCA_opal_timer_STATIC_SUBDIRS)' >> MCA_opal_FRAMEWORK_LIBS=' $(MCA_opal_common_STATIC_LTLIBS) >> mca/allocator/libmca_allocator.la >> <http://libmca_allocator.la> >> $(MCA_opal_allocator_STATIC_LTLIBS) >> mca/backtrace/libmca_backtrace.la >> <http://libmca_backtrace.la> >> $(MCA_opal_backtrace_STATIC_LTLIBS) >> mca/btl/libmca_btl.la <http://libmca_btl.la> >> $(MCA_opal_btl_STATIC_LTLIBS) mca/dl/libmca_dl.la >> <http://libmca_dl.la> >> $(MCA_opal_dl_STATIC_LTLIBS) mca/event/libmca_event.la >> <http://libmca_event.la> >> $(MCA_opal_event_STATIC_LTLIBS) mca/hwloc/libmca_hwloc.la >> <http://libmca_hwloc.la> >> $(MCA_opal_hwloc_STATIC_LTLIBS) mca/if/libmca_if.la >> <http://libmca_if.la> >> $(MCA_opal_if_STATIC_LTLIBS) >> mca/installdirs/libmca_installdirs.la >> <http://libmca_installdirs.la> >> $(MCA_opal_installdirs_STATIC_LTLIBS) >> mca/memchecker/libmca_memchecker.la >> <http://libmca_memchecker.la> >> $(MCA_opal_memchecker_STATIC_LTLIBS) >> mca/memcpy/libmca_memcpy.la >> <http://libmca_memcpy.la> >> $(MCA_opal_memcpy_STATIC_LTLIBS) mca/memory/libmca_memory.la >> <http://libmca_memory.la> >> $(MCA_opal_memory_STATIC_LTLIBS) mca/mpool/libmca_mpool.la >> <http://libmca_mpool.la> >> $(MCA_opal_mpool_STATIC_LTLIBS) mca/pmix/libmca_pmix.la >> <http://libmca_pmix.la> >> $(MCA_opal_pmix_STATIC_LTLIBS) mca/pstat/libmca_pstat.la >> <http://libmca_pstat.la> >> $(MCA_opal_pstat_STATIC_LTLIBS) mca/rcache/libmca_rcache.la >> <http://libmca_rcache.la> >> $(MCA_opal_rcache_STATIC_LTLIBS) mca/sec/libmca_sec.la >> <http://libmca_sec.la> >> $(MCA_opal_sec_STATIC_LTLIBS) mca/shmem/libmca_shmem.la >> <http://libmca_shmem.la> >> $(MCA_opal_shmem_STATIC_LTLIBS) mca/timer/libmca_timer.la >> <http://libmca_timer.la> >> $(MCA_opal_timer_STATIC_LTLIBS)' >> ... >> MCA_opal_pmix_ALL_COMPONENTS=' s1 cray s2 pmix112 external' >> MCA_opal_pmix_ALL_SUBDIRS=' mca/pmix/s1 mca/pmix/cray >> mca/pmix/s2 >> mca/pmix/pmix112 mca/pmix/external' >> MCA_opal_pmix_DSO_COMPONENTS=' pmix112' >> MCA_opal_pmix_DSO_SUBDIRS=' mca/pmix/pmix112' >> MCA_opal_pmix_STATIC_COMPONENTS='' >> MCA_opal_pmix_STATIC_LTLIBS='' >> MCA_opal_pmix_STATIC_SUBDIRS='' >> ... >> opal_pmix_ext_CPPFLAGS='' >> opal_pmix_ext_LDFLAGS='' >> opal_pmix_ext_LIBS='' >> opal_pmix_pmix112_CPPFLAGS='-I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/include/pmix >> -I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/include >> -I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix >> -I$(OPAL_TOP_SRCDIR)/opal/mca/pmix/pmix112/pmix' >> opal_pmix_pmix112_LIBS='$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/libpmix.la >> <http://libpmix.la>' >> >> ... >> >> >> >> "config.log" for Sun C 5.13 shows the following. >> >> ... >> configure:127803: /bin/bash >> '../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/ >> pmix/configure' *failed* for opal/mca/pmix/pmix112/pmix >> configure:128379: checking if MCA component pmix:pmix112 >> can compile >> configure:128381: result: no >> configure:5637: --- MCA component pmix:external (m4 >> configuration >> macro) >> configure:128523: checking for MCA component >> pmix:external compile >> mode >> configure:128529: result: dso >> configure:129054: checking if MCA component pmix:external >> can compile >> configure:129056: result: no >> ... >> config.status:3887: creating opal/mca/pmix/Makefile >> config.status:3887: creating opal/mca/pmix/s1/Makefile >> config.status:3887: creating opal/mca/pmix/cray/Makefile >> config.status:3887: creating opal/mca/pmix/s2/Makefile >> config.status:3887: creating opal/mca/pmix/pmix112/Makefile >> config.status:3887: creating opal/mca/pmix/external/Makefile >> ... >> MCA_BUILD_opal_pmix_cray_DSO_FALSE='#' >> MCA_BUILD_opal_pmix_cray_DSO_TRUE='' >> MCA_BUILD_opal_pmix_external_DSO_FALSE='#' >> MCA_BUILD_opal_pmix_external_DSO_TRUE='' >> MCA_BUILD_opal_pmix_pmix112_DSO_FALSE='#' >> MCA_BUILD_opal_pmix_pmix112_DSO_TRUE='' >> MCA_BUILD_opal_pmix_s1_DSO_FALSE='#' >> MCA_BUILD_opal_pmix_s1_DSO_TRUE='' >> MCA_BUILD_opal_pmix_s2_DSO_FALSE='#' >> MCA_BUILD_opal_pmix_s2_DSO_TRUE='' >> ... >> MCA_opal_FRAMEWORKS='common allocator backtrace btl dl >> event hwloc if >> installdirs memchecker memcpy memory mpool pmix pstat >> rcache sec >> shmem timer' >> MCA_opal_FRAMEWORKS_SUBDIRS='mca/common mca/allocator >> mca/backtrace mca/btl >> mca/dl mca/event mca/hwloc mca/if mca/installdirs >> mca/memchecker >> mca/memcpy >> mca/memory mca/mpool mca/pmix mca/pstat mca/rcache mca/sec >> mca/shmem mca/timer' >> MCA_opal_FRAMEWORK_COMPONENT_ALL_SUBDIRS='$(MCA_opal_common_ALL_SUBDIRS) >> $(MCA_opal_allocator_ALL_SUBDIRS) >> $(MCA_opal_backtrace_ALL_SUBDIRS) >> $(MCA_opal_btl_ALL_SUBDIRS) $(MCA_opal_dl_ALL_SUBDIRS) >> $(MCA_opal_event_ALL_SUBDIRS) $(MCA_opal_hwloc_ALL_SUBDIRS) >> $(MCA_opal_if_ALL_SUBDIRS) >> $(MCA_opal_installdirs_ALL_SUBDIRS) >> $(MCA_opal_memchecker_ALL_SUBDIRS) >> $(MCA_opal_memcpy_ALL_SUBDIRS) >> $(MCA_opal_memory_ALL_SUBDIRS) $(MCA_opal_mpool_ALL_SUBDIRS) >> $(MCA_opal_pmix_ALL_SUBDIRS) $(MCA_opal_pstat_ALL_SUBDIRS) >> $(MCA_opal_rcache_ALL_SUBDIRS) $(MCA_opal_sec_ALL_SUBDIRS) >> $(MCA_opal_shmem_ALL_SUBDIRS) $(MCA_opal_timer_ALL_SUBDIRS)' >> MCA_opal_FRAMEWORK_COMPONENT_DSO_SUBDIRS='$(MCA_opal_common_DSO_SUBDIRS) >> $(MCA_opal_allocator_DSO_SUBDIRS) >> $(MCA_opal_backtrace_DSO_SUBDIRS) >> $(MCA_opal_btl_DSO_SUBDIRS) $(MCA_opal_dl_DSO_SUBDIRS) >> $(MCA_opal_event_DSO_SUBDIRS) $(MCA_opal_hwloc_DSO_SUBDIRS) >> $(MCA_opal_if_DSO_SUBDIRS) >> $(MCA_opal_installdirs_DSO_SUBDIRS) >> $(MCA_opal_memchecker_DSO_SUBDIRS) >> $(MCA_opal_memcpy_DSO_SUBDIRS) >> $(MCA_opal_memory_DSO_SUBDIRS) $(MCA_opal_mpool_DSO_SUBDIRS) >> $(MCA_opal_pmix_DSO_SUBDIRS) $(MCA_opal_pstat_DSO_SUBDIRS) >> $(MCA_opal_rcache_DSO_SUBDIRS) $(MCA_opal_sec_DSO_SUBDIRS) >> $(MCA_opal_shmem_DSO_SUBDIRS) $(MCA_opal_timer_DSO_SUBDIRS)' >> MCA_opal_FRAMEWORK_COMPONENT_STATIC_SUBDIRS='$(MCA_opal_common_STATIC_SUBDIRS) >> $(MCA_opal_allocator_STATIC_SUBDIRS) >> $(MCA_opal_backtrace_STATIC_SUBDIRS) >> $(MCA_opal_btl_STATIC_SUBDIRS) $(MCA_opal_dl_STATIC_SUBDIRS) >> $(MCA_opal_event_STATIC_SUBDIRS) >> $(MCA_opal_hwloc_STATIC_SUBDIRS) >> $(MCA_opal_if_STATIC_SUBDIRS) >> $(MCA_opal_installdirs_STATIC_SUBDIRS) >> $(MCA_opal_memchecker_STATIC_SUBDIRS) >> $(MCA_opal_memcpy_STATIC_SUBDIRS) >> $(MCA_opal_memory_STATIC_SUBDIRS) >> $(MCA_opal_mpool_STATIC_SUBDIRS) >> $(MCA_opal_pmix_STATIC_SUBDIRS) >> $(MCA_opal_pstat_STATIC_SUBDIRS) >> $(MCA_opal_rcache_STATIC_SUBDIRS) >> $(MCA_opal_sec_STATIC_SUBDIRS) >> $(MCA_opal_shmem_STATIC_SUBDIRS) >> $(MCA_opal_timer_STATIC_SUBDIRS)' >> MCA_opal_FRAMEWORK_LIBS=' $(MCA_opal_common_STATIC_LTLIBS) >> mca/allocator/libmca_allocator.la >> <http://libmca_allocator.la> >> $(MCA_opal_allocator_STATIC_LTLIBS) >> mca/backtrace/libmca_backtrace.la >> <http://libmca_backtrace.la> >> $(MCA_opal_backtrace_STATIC_LTLIBS) >> mca/btl/libmca_btl.la <http://libmca_btl.la> >> $(MCA_opal_btl_STATIC_LTLIBS) mca/dl/libmca_dl.la >> <http://libmca_dl.la> >> $(MCA_opal_dl_STATIC_LTLIBS) mca/event/libmca_event.la >> <http://libmca_event.la> >> $(MCA_opal_event_STATIC_LTLIBS) mca/hwloc/libmca_hwloc.la >> <http://libmca_hwloc.la> >> $(MCA_opal_hwloc_STATIC_LTLIBS) mca/if/libmca_if.la >> <http://libmca_if.la> >> $(MCA_opal_if_STATIC_LTLIBS) >> mca/installdirs/libmca_installdirs.la >> <http://libmca_installdirs.la> >> $(MCA_opal_installdirs_STATIC_LTLIBS) >> mca/memchecker/libmca_memchecker.la >> <http://libmca_memchecker.la> >> $(MCA_opal_memchecker_STATIC_LTLIBS) >> mca/memcpy/libmca_memcpy.la >> <http://libmca_memcpy.la> >> $(MCA_opal_memcpy_STATIC_LTLIBS) mca/memory/libmca_memory.la >> <http://libmca_memory.la> >> $(MCA_opal_memory_STATIC_LTLIBS) mca/mpool/libmca_mpool.la >> <http://libmca_mpool.la> >> $(MCA_opal_mpool_STATIC_LTLIBS) mca/pmix/libmca_pmix.la >> <http://libmca_pmix.la> >> $(MCA_opal_pmix_STATIC_LTLIBS) mca/pstat/libmca_pstat.la >> <http://libmca_pstat.la> >> $(MCA_opal_pstat_STATIC_LTLIBS) mca/rcache/libmca_rcache.la >> <http://libmca_rcache.la> >> $(MCA_opal_rcache_STATIC_LTLIBS) mca/sec/libmca_sec.la >> <http://libmca_sec.la> >> $(MCA_opal_sec_STATIC_LTLIBS) mca/shmem/libmca_shmem.la >> <http://libmca_shmem.la> >> $(MCA_opal_shmem_STATIC_LTLIBS) mca/timer/libmca_timer.la >> <http://libmca_timer.la> >> $(MCA_opal_timer_STATIC_LTLIBS)' >> ... >> MCA_opal_pmix_ALL_COMPONENTS=' s1 cray s2 pmix112 external' >> MCA_opal_pmix_ALL_SUBDIRS=' mca/pmix/s1 mca/pmix/cray >> mca/pmix/s2 >> mca/pmix/pmix112 mca/pmix/external' >> MCA_opal_pmix_DSO_COMPONENTS='' >> MCA_opal_pmix_DSO_SUBDIRS='' >> MCA_opal_pmix_STATIC_COMPONENTS='' >> MCA_opal_pmix_STATIC_LTLIBS='' >> MCA_opal_pmix_STATIC_SUBDIRS='' >> ... >> opal_pmix_ext_CPPFLAGS='' >> opal_pmix_ext_LDFLAGS='' >> opal_pmix_ext_LIBS='' >> opal_pmix_pmix112_CPPFLAGS='' >> opal_pmix_pmix112_LIBS='' >> ... >> >> >> >> >> I've attached the config.log files for pmix. >> >> tyr openmpi-2.0.0 142 tar zvft pmix_config.log.tar.gz >> -rw-r--r-- root/root 136291 2016-04-25 08:05:34 >> openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/pmix/pmix112/pmix/config.log >> -rw-r--r-- root/root 528808 2016-04-25 08:07:54 >> openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_gcc/opal/mca/pmix/pmix112/pmix/config.log >> tyr openmpi-2.0.0 143 >> >> >> >> I've also attached the output for the broken execution of >> "spawn_multiple_master" for my gcc-version of Open MPI. >> "spawn_master" works as expected with my gcc-version of >> Open MPI. >> >> Hopefully you can fix the problem. >> >> >> Kind regards and thank you very much for your help >> >> Siegmar >> >> >> >> Am 23.04.2016 um 21:34 schrieb Siegmar Gross: >> >> Hi Gilles, >> >> I don't know what happened, but the files are not >> available now >> and they were definitely available when I answered >> the email from >> Ralph. The files also have a different timestamp now. >> This is an >> extract from my email to Ralph for Solaris Sparc. >> >> -rwxr-xr-x 1 root root 977 Apr 19 19:49 >> mca_plm_rsh.la >> <http://mca_plm_rsh.la> >> -rwxr-xr-x 1 root root 153280 Apr 19 19:49 >> mca_plm_rsh.so >> -rwxr-xr-x 1 root root 1007 Apr 19 19:47 >> mca_pmix_pmix112.la <http://mca_pmix_pmix112.la> >> -rwxr-xr-x 1 root root 1400512 Apr 19 19:47 >> mca_pmix_pmix112.so >> -rwxr-xr-x 1 root root 971 Apr 19 19:52 >> mca_pml_cm.la >> <http://mca_pml_cm.la> >> -rwxr-xr-x 1 root root 342440 Apr 19 19:52 >> mca_pml_cm.so >> >> Now I have the following output for these files. >> >> -rwxr-xr-x 1 root root 976 Apr 19 19:58 >> mca_plm_rsh.la >> <http://mca_plm_rsh.la> >> -rwxr-xr-x 1 root root 319816 Apr 19 19:58 >> mca_plm_rsh.so >> -rwxr-xr-x 1 root root 970 Apr 19 20:00 >> mca_pml_cm.la >> <http://mca_pml_cm.la> >> -rwxr-xr-x 1 root root 1507440 Apr 19 20:00 >> mca_pml_cm.so >> >> I'll try to find out what happened next week when I'm >> back in >> my office. >> >> >> Kind regards >> >> Siegmar >> >> >> >> >> >> Am 23.04.16 um 02:12 schrieb Gilles Gouaillardet: >> >> Siegmar, >> >> I will try to reproduce this on my solaris11 >> x86_64 vm >> >> In the mean time, can you please double check >> mca_pmix_pmix_pmix112.so >> is a 64 bits library ? >> (E.g, confirm "-m64" was correctly passed to pmix) >> >> Cheers, >> >> Gilles >> >> On Friday, April 22, 2016, Siegmar Gross >> <siegmar.gr...@informatik.hs-fulda.de >> <mailto:siegmar.gr...@informatik.hs-fulda.de>> wrote: >> >> Hi Ralph, >> >> I've already used "-enable-debug". >> "SYSTEM_ENV" is >> "SunOS" or >> "Linux" and "MACHINE_ENV" is "sparc" or >> "x86_84". >> >> mkdir >> openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_gcc >> cd >> openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_gcc >> >> ../openmpi-v2.x-dev-1280-gc110ae8/configure \ >> --prefix=/usr/local/openmpi-2.0.0_64_gcc \ >> --libdir=/usr/local/openmpi-2.0.0_64_gcc/lib64 \ >> --with-jdk-bindir=/usr/local/jdk1.8.0/bin \ >> --with-jdk-headers=/usr/local/jdk1.8.0/include \ >> JAVA_HOME=/usr/local/jdk1.8.0 \ >> LDFLAGS="-m64" CC="gcc" CXX="g++" >> FC="gfortran" \ >> CFLAGS="-m64" CXXFLAGS="-m64" FCFLAGS="-m64" \ >> CPP="cpp" CXXCPP="cpp" \ >> --enable-mpi-cxx \ >> --enable-cxx-exceptions \ >> --enable-mpi-java \ >> --enable-heterogeneous \ >> --enable-mpi-thread-multiple \ >> --with-hwloc=internal \ >> --without-verbs \ >> --with-wrapper-cflags="-std=c11 -m64" \ >> --with-wrapper-cxxflags="-m64" \ >> --with-wrapper-fcflags="-m64" \ >> --enable-debug \ >> |& tee >> log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_gcc >> >> >> mkdir >> openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_cc >> cd >> openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_cc >> >> ../openmpi-v2.x-dev-1280-gc110ae8/configure \ >> --prefix=/usr/local/openmpi-2.0.0_64_cc \ >> --libdir=/usr/local/openmpi-2.0.0_64_cc/lib64 \ >> --with-jdk-bindir=/usr/local/jdk1.8.0/bin \ >> --with-jdk-headers=/usr/local/jdk1.8.0/include \ >> JAVA_HOME=/usr/local/jdk1.8.0 \ >> LDFLAGS="-m64" CC="cc" CXX="CC" FC="f95" \ >> CFLAGS="-m64" CXXFLAGS="-m64 >> -library=stlport4" >> FCFLAGS="-m64" \ >> CPP="cpp" CXXCPP="cpp" \ >> --enable-mpi-cxx \ >> --enable-cxx-exceptions \ >> --enable-mpi-java \ >> --enable-heterogeneous \ >> --enable-mpi-thread-multiple \ >> --with-hwloc=internal \ >> --without-verbs \ >> --with-wrapper-cflags="-m64" \ >> --with-wrapper-cxxflags="-m64 >> -library=stlport4" \ >> --with-wrapper-fcflags="-m64" \ >> --with-wrapper-ldflags="" \ >> --enable-debug \ >> |& tee >> log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_cc >> >> >> Kind regards >> >> Siegmar >> >> Am 21.04.2016 um 18:18 schrieb Ralph Castain: >> >> Can you please rebuild OMPI with >> -enable-debug in >> the configure >> cmd? It will let us see more error output >> >> >> On Apr 21, 2016, at 8:52 AM, Siegmar >> Gross >> <siegmar.gr...@informatik.hs-fulda.de> wrote: >> >> Hi Ralph, >> >> I don't see any additional information. >> >> tyr hello_1 108 mpiexec -np 4 --host >> tyr,sunpc1,linpc1,ruester -mca >> mca_base_component_show_load_errors 1 hello_1_mpi >> [tyr.informatik.hs-fulda.de:06211 >> <http://tyr.informatik.hs-fulda.de:06211> >> <http://tyr.informatik.hs-fulda.de:06211>] >> [[48741,0],0] >> ORTE_ERROR_LOG: Not found in file >> >> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c >> >> at line 638 >> >> -------------------------------------------------------------------------- >> It looks like orte_init failed for some >> reason; your >> parallel process is >> likely to abort. There are many >> reasons that >> a parallel >> process can >> fail during orte_init; some of which >> are due >> to configuration or >> environment problems. This failure >> appears to >> be an >> internal failure; >> here's some additional information >> (which may >> only be >> relevant to an >> Open MPI developer): >> >> opal_pmix_base_select failed >> --> Returned value Not found (-13) >> instead of >> ORTE_SUCCESS >> >> -------------------------------------------------------------------------- >> >> >> tyr hello_1 109 mpiexec -np 4 --host >> tyr,sunpc1,linpc1,ruester -mca >> mca_base_component_show_load_errors 1 -mca >> pmix_base_verbose >> 10 -mca pmix_server_verbose 5 >> hello_1_mpi >> [tyr.informatik.hs-fulda.de:06212 >> <http://tyr.informatik.hs-fulda.de:06212> >> <http://tyr.informatik.hs-fulda.de:06212>] >> mca: base: >> components_register: registering >> framework >> pmix components >> [tyr.informatik.hs-fulda.de:06212 >> <http://tyr.informatik.hs-fulda.de:06212> >> <http://tyr.informatik.hs-fulda.de:06212>] >> mca: base: >> components_open: opening pmix components >> [tyr.informatik.hs-fulda.de:06212 >> <http://tyr.informatik.hs-fulda.de:06212> >> <http://tyr.informatik.hs-fulda.de:06212>] >> mca:base:select: >> Auto-selecting pmix components >> [tyr.informatik.hs-fulda.de:06212 >> <http://tyr.informatik.hs-fulda.de:06212> >> <http://tyr.informatik.hs-fulda.de:06212>] >> mca:base:select:( >> pmix) No component selected! >> [tyr.informatik.hs-fulda.de:06212 >> <http://tyr.informatik.hs-fulda.de:06212> >> <http://tyr.informatik.hs-fulda.de:06212>] >> [[48738,0],0] >> ORTE_ERROR_LOG: Not found in file >> >> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c >> >> at line 638 >> >> -------------------------------------------------------------------------- >> It looks like orte_init failed for some >> reason; your >> parallel process is >> likely to abort. There are many >> reasons that >> a parallel >> process can >> fail during orte_init; some of which >> are due >> to configuration or >> environment problems. This failure >> appears to >> be an >> internal failure; >> here's some additional information >> (which may >> only be >> relevant to an >> Open MPI developer): >> >> opal_pmix_base_select failed >> --> Returned value Not found (-13) >> instead of >> ORTE_SUCCESS >> >> -------------------------------------------------------------------------- >> tyr hello_1 110 >> >> >> Kind regards >> >> Siegmar >> >> >> Am 21.04.2016 um 17:24 schrieb Ralph >> Castain: >> >> Hmmm?it looks like you built the >> right >> components, but >> they are not being picked up. Can >> you run >> your mpiexec >> command again, adding ?-mca >> mca_base_component_show_load_errors 1? to >> the cmd line? >> >> >> On Apr 21, 2016, at 8:16 AM, >> Siegmar Gross >> <siegmar.gr...@informatik.hs-fulda.de> >> wrote: >> >> Hi Ralph, >> >> I have attached ompi_info >> output for >> both compilers >> from my >> sparc machine and the >> listings for >> both compilers >> from the >> <prefix>/lib/openmpi directories. >> Hopefully that >> helps to >> find the problem. >> >> hermes tmp 3 tar zvft >> openmpi-2.x_info.tar.gz >> -rw-r--r-- root/root 10969 >> 2016-04-21 17:06 >> ompi_info_SunOS_sparc_cc.txt >> -rw-r--r-- root/root 11044 >> 2016-04-21 17:06 >> ompi_info_SunOS_sparc_gcc.txt >> -rw-r--r-- root/root 71252 >> 2016-04-21 17:02 >> lib64_openmpi.txt >> hermes tmp 4 >> >> >> Kind regards and thank you >> very much >> once more for >> your help >> >> Siegmar >> >> >> Am 21.04.2016 um 15:54 >> schrieb Ralph >> Castain: >> >> Odd - it would appear >> that none of >> the pmix >> components built? Can you >> send >> along the output from >> ompi_info? >> Or just send a >> listing of the files in the >> <prefix>/lib/openmpi directory? >> >> >> On Apr 21, 2016, at >> 1:27 AM, >> Siegmar Gross >> >> <siegmar.gr...@informatik.hs-fulda.de >> >> <mailto:siegmar.gr...@informatik.hs-fulda.de>> >> wrote: >> >> Hi Ralph, >> >> Am 21.04.2016 um >> 00:18 schrieb >> Ralph Castain: >> >> Could you please >> rerun >> these test and >> add ?-mca >> pmix_base_verbose 10 >> -mca >> pmix_server_verbose >> 5? to your cmd >> line? I need to >> see why the >> pmix components >> failed. >> >> >> >> tyr spawn 111 mpiexec >> -np 1 --host >> tyr,sunpc1,linpc1,ruester -mca >> pmix_base_verbose 10 -mca >> pmix_server_verbose 5 >> spawn_multiple_master >> [tyr.informatik.hs-fulda.de >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de/>:26652] mca: >> base: >> components_register: >> registering >> framework pmix >> components >> [tyr.informatik.hs-fulda.de >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de/>:26652] mca: >> base: >> components_open: opening >> pmix components >> [tyr.informatik.hs-fulda.de >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de/>:26652] >> mca:base:select: >> Auto-selecting pmix components >> [tyr.informatik.hs-fulda.de >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de/>:26652] >> mca:base:select:( pmix) No >> component selected! >> [tyr.informatik.hs-fulda.de >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de/>:26652] >> [[52794,0],0] >> ORTE_ERROR_LOG: >> Not found in file >> >> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c >> >> at line 638 >> >> -------------------------------------------------------------------------- >> It looks like >> orte_init failed >> for some >> reason; your parallel >> process is >> likely to abort. >> There are >> many reasons >> that a parallel >> process can >> fail during >> orte_init; some of >> which are due >> to configuration or >> environment >> problems. This >> failure appears >> to be an internal >> failure; >> here's some additional >> information (which >> may only be relevant >> to an >> Open MPI developer): >> >> opal_pmix_base_select failed >> --> Returned value >> Not found >> (-13) instead >> of ORTE_SUCCESS >> >> -------------------------------------------------------------------------- >> tyr spawn 112 >> >> >> >> >> tyr hello_1 116 >> mpiexec -np 1 >> --host >> tyr,sunpc1,linpc1,ruester -mca >> pmix_base_verbose 10 -mca >> pmix_server_verbose 5 hello_1_mpi >> [tyr.informatik.hs-fulda.de >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de/>:27261] mca: >> base: >> components_register: >> registering >> framework pmix >> components >> [tyr.informatik.hs-fulda.de >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de/>:27261] mca: >> base: >> components_open: opening >> pmix components >> [tyr.informatik.hs-fulda.de >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de/>:27261] >> mca:base:select: >> Auto-selecting pmix components >> [tyr.informatik.hs-fulda.de >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de/>:27261] >> mca:base:select:( pmix) No >> component selected! >> [tyr.informatik.hs-fulda.de >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de/>:27261] >> [[52315,0],0] >> ORTE_ERROR_LOG: >> Not found in file >> >> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c >> >> at line 638 >> >> -------------------------------------------------------------------------- >> It looks like >> orte_init failed >> for some >> reason; your parallel >> process is >> likely to abort. >> There are >> many reasons >> that a parallel >> process can >> fail during >> orte_init; some of >> which are due >> to configuration or >> environment >> problems. This >> failure appears >> to be an internal >> failure; >> here's some additional >> information (which >> may only be relevant >> to an >> Open MPI developer): >> >> opal_pmix_base_select failed >> --> Returned value >> Not found >> (-13) instead >> of ORTE_SUCCESS >> >> -------------------------------------------------------------------------- >> tyr hello_1 117 >> >> >> >> Thank you very much >> for your help. >> >> >> Kind regards >> >> Siegmar >> >> >> >> >> Thanks >> Ralph >> >> On Apr 20, >> 2016, at >> 10:12 AM, >> Siegmar Gross >> >> <siegmar.gr...@informatik.hs-fulda.de >> >> <mailto:siegmar.gr...@informatik.hs-fulda.de>> >> wrote: >> >> Hi, >> >> I have built >> >> openmpi-v2.x-dev-1280-gc110ae8 on my >> machines >> (Solaris 10 >> Sparc, >> Solaris 10 >> x86_64, and >> openSUSE Linux >> 12.1 x86_64) >> with >> gcc-5.1.0 and Sun >> C 5.13. >> Unfortunately >> I get >> runtime >> errors for >> some programs. >> >> >> Sun C 5.13: >> =========== >> >> For all my test >> programs I get the >> same error on >> Solaris >> Sparc and >> Solaris >> x86_64, while >> the programs >> work fine on >> Linux. >> >> tyr hello_1 115 >> mpiexec -np 2 >> hello_1_mpi >> >> [tyr.informatik.hs-fulda.de >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de>:22373] >> [[61763,0],0] >> ORTE_ERROR_LOG: Not >> found in file >> >> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c >> >> at line 638 >> >> -------------------------------------------------------------------------- >> It looks like >> orte_init failed for >> some reason; >> your >> parallel process is >> likely to abort. >> There are many >> reasons that a >> parallel process can >> fail during >> orte_init; >> some of which >> are due to >> configuration or >> environment problems. >> This failure >> appears to be an >> internal failure; >> here's some >> additional >> information >> (which may >> only be >> relevant to an >> Open MPI >> developer): >> >> opal_pmix_base_select >> failed >> --> Returned >> value Not >> found (-13) >> instead of >> ORTE_SUCCESS >> >> -------------------------------------------------------------------------- >> tyr hello_1 116 >> >> >> >> >> GCC-5.1.0: >> ========== >> >> tyr spawn 121 >> mpiexec >> -np 1 --host >> tyr,sunpc1,linpc1,ruester >> spawn_multiple_master >> >> Parent process 0 >> running on >> >> tyr.informatik.hs-fulda.de >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de> >> I create 3 slave >> processes. >> >> >> [tyr.informatik.hs-fulda.de >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de>:25366] >> PMIX ERROR: >> UNPACK-PAST-END in file >> >> ../../../../../../openmpi-v2.x-dev-1280-gc110ae8/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c >> >> >> at line 829 >> >> [tyr.informatik.hs-fulda.de >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de> >> >> <http://tyr.informatik.hs-fulda.de>:25366] >> PMIX ERROR: >> UNPACK-PAST-END in file >> >> ../../../../../../openmpi-v2.x-dev-1280-gc110ae8/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c >> >> >> at line 2176 >> [tyr:25377] *** An >> error occurred in >> MPI_Comm_spawn_multiple >> [tyr:25377] *** >> reported by process >> [3308257281,0] >> [tyr:25377] *** on >> communicator >> MPI_COMM_WORLD >> [tyr:25377] *** >> MPI_ERR_SPAWN: could >> not spawn >> processes >> [tyr:25377] *** >> MPI_ERRORS_ARE_FATAL >> (processes in this >> communicator will >> now abort, >> [tyr:25377] *** and >> potentially >> your MPI job) >> tyr spawn 122 >> >> >> I would be >> grateful if >> somebody can >> fix the >> problems. >> Thank you very >> much for any >> help in >> advance. >> >> >> Kind regards >> >> Siegmar >> >> <hello_1_mpi.c><spawn_multiple_master.c>_______________________________________________ >> >> >> users mailing >> list >> us...@open-mpi.org >> <mailto:us...@open-mpi.org> >> Subscription: >> >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> Link to this >> post: >> >> http://www.open-mpi.org/community/lists/users/2016/04/28983.php >> >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> <mailto:us...@open-mpi.org> >> Subscription: >> >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> Link to this >> post: >> >> http://www.open-mpi.org/community/lists/users/2016/04/28986.php >> >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> <mailto:us...@open-mpi.org> >> Subscription: >> >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> Link to this >> post: >> >> http://www.open-mpi.org/community/lists/users/2016/04/28987.php >> >> >> >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> Subscription: >> >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> Link to this post: >> >> http://www.open-mpi.org/community/lists/users/2016/04/28988.php >> >> >> <openmpi-2.x_info.tar.gz>_______________________________________________ >> users mailing list >> us...@open-mpi.org >> Subscription: >> >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> Link to this post: >> >> http://www.open-mpi.org/community/lists/users/2016/04/28989.php >> >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> Subscription: >> >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> Link to this post: >> >> http://www.open-mpi.org/community/lists/users/2016/04/28990.php >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> Subscription: >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> Link to this post: >> >> http://www.open-mpi.org/community/lists/users/2016/04/28991.php >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> Subscription: >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> Link to this post: >> >> http://www.open-mpi.org/community/lists/users/2016/04/28992.php >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> Subscription: >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> Link to this post: >> >> http://www.open-mpi.org/community/lists/users/2016/04/28993.php >> >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> Subscription: >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> Link to this post: >> http://www.open-mpi.org/community/lists/users/2016/04/28999.php >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> Subscription: >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> Link to this post: >> http://www.open-mpi.org/community/lists/users/2016/04/29009.php >> >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> Subscription: >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> Link to this post: >> http://www.open-mpi.org/community/lists/users/2016/04/29033.php >> >> >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> Subscription: >> https://www.open-mpi.org/mailman/listinfo.cgi/users >> Link to this post: >> http://www.open-mpi.org/community/lists/users/2016/04/29038.php > > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > http://www.open-mpi.org/community/lists/users/2016/04/29041.php -------------- next part -------------- HTML attachment scrubbed and removed ------------------------------ Subject: Digest Footer _______________________________________________ users mailing list us...@open-mpi.org https://www.open-mpi.org/mailman/listinfo.cgi/users ------------------------------ End of users Digest, Vol 3486, Issue 3 **************************************