Hi Gilles, adding "-std=c99" to CFLAGS solves the problem with the missing library. Shall I add it permanently to my configure command or will you add it, so that I will not run into problems if you need the C11 standard later?
"spawn_multiple_master" breaks with the same error that I reported yesterday for my gcc-version of Open MPI. Hopefully you can solve the problem as well. Kind regards and thank you very much for your help Siegmar Am 27.04.2016 um 08:05 schrieb Gilles Gouaillardet:
Siegmar, here is the error : configure:17969: cc -o conftest -m64 -D_REENTRANT -g -g -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1 -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/include -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/include -D_REENTRANT -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/hwloc/hwloc1112/hwloc/include -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/hwloc/hwloc1112/hwloc/include -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/event/libevent2022/libevent -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/event/libevent2022/libevent/include -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/event/libevent2022/libevent/include -m64 conftest.c >&5 "/usr/include/stdbool.h", line 42: #error: "Use of <stdbool.h> is valid only in a c99 compilation environment." i cannot reproduce this on solaris 11 with oracle studio 5.3 compiler, and i do not have solaris 10 yet. could you please re-configure with '-std=c99' appended to your CFLAGS and see if it helps ? Cheers, Gilles On 4/26/2016 7:57 PM, Siegmar Gross wrote:Hi Gilles and Ralph, I was able to sort out my mess. In my last email I compared the files from "SunOS_sparc/openmpi-2.0.0_64_gcc/lib64/openmpi" from the attachment of my email to Ralph with the files from "SunOS_sparc/openmpi-2.0.0_64_cc/lib64/openmpi" from my current file system. That's the reason while I have had different timestamps. The other problem was that Ralph didn't recognize that "mca_pmix_pmix112.so" wasn't built on Solaris with the Sun C compiler. I've removed most of the files from the attachment of my email so that it is easier to see the relevant files. Below I try to give you more information that may be relevant to track down the problem. I still get an error running one of my small test programs, when I use my gcc-version of Open MPI. "mca_pmix_pmix112.so" is a 64 bits library. Linux_x86_64/openmpi-2.0.0_64_cc/lib64/openmpi: ... -rwxr-xr-x 1 root root 261327 Apr 19 16:46 mca_plm_slurm.so -rwxr-xr-x 1 root root 1002 Apr 19 16:45 mca_pmix_pmix112.la -rwxr-xr-x 1 root root 3906526 Apr 19 16:45 mca_pmix_pmix112.so -rwxr-xr-x 1 root root 966 Apr 19 16:51 mca_pml_cm.la -rwxr-xr-x 1 root root 1574265 Apr 19 16:51 mca_pml_cm.so ... Linux_x86_64/openmpi-2.0.0_64_gcc/lib64/openmpi: ... -rwxr-xr-x 1 root root 70371 Apr 19 16:43 mca_plm_slurm.so -rwxr-xr-x 1 root root 1008 Apr 19 16:42 mca_pmix_pmix112.la -rwxr-xr-x 1 root root 1029005 Apr 19 16:42 mca_pmix_pmix112.so -rwxr-xr-x 1 root root 972 Apr 19 16:46 mca_pml_cm.la -rwxr-xr-x 1 root root 284858 Apr 19 16:46 mca_pml_cm.so ... SunOS_sparc/openmpi-2.0.0_64_cc/lib64/openmpi: ... -rwxr-xr-x 1 root root 319816 Apr 19 19:58 mca_plm_rsh.so -rwxr-xr-x 1 root root 970 Apr 19 20:00 mca_pml_cm.la -rwxr-xr-x 1 root root 1507440 Apr 19 20:00 mca_pml_cm.so ... SunOS_sparc/openmpi-2.0.0_64_gcc/lib64/openmpi: ... -rwxr-xr-x 1 root root 153280 Apr 19 19:49 mca_plm_rsh.so -rwxr-xr-x 1 root root 1007 Apr 19 19:47 mca_pmix_pmix112.la -rwxr-xr-x 1 root root 1400512 Apr 19 19:47 mca_pmix_pmix112.so -rwxr-xr-x 1 root root 971 Apr 19 19:52 mca_pml_cm.la -rwxr-xr-x 1 root root 342440 Apr 19 19:52 mca_pml_cm.so ... SunOS_x86_64/openmpi-2.0.0_64_cc/lib64/openmpi: ... -rwxr-xr-x 1 root root 300096 Apr 19 17:18 mca_plm_rsh.so -rwxr-xr-x 1 root root 970 Apr 19 17:23 mca_pml_cm.la -rwxr-xr-x 1 root root 1458816 Apr 19 17:23 mca_pml_cm.so ... SunOS_x86_64/openmpi-2.0.0_64_gcc/lib64/openmpi: ... -rwxr-xr-x 1 root root 133096 Apr 19 17:42 mca_plm_rsh.so -rwxr-xr-x 1 root root 1007 Apr 19 17:41 mca_pmix_pmix112.la -rwxr-xr-x 1 root root 1320240 Apr 19 17:41 mca_pmix_pmix112.so -rwxr-xr-x 1 root root 971 Apr 19 17:46 mca_pml_cm.la -rwxr-xr-x 1 root root 419848 Apr 19 17:46 mca_pml_cm.so ... Yesterday I've installed openmpi-v2.x-dev-1290-gbd0e4e1 so that we have a current version for the investigation of the problem. Once more mca_pmix_pmix112.so isn't available on Solaris if I use the Sun C compiler. "config.log" for gcc-5.1.0 shows the following. ... configure:127799: /bin/bash '../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/ pmix/configure' succeeded for opal/mca/pmix/pmix112/pmix configure:127916: checking if MCA component pmix:pmix112 can compile configure:127918: result: yes configure:5637: --- MCA component pmix:external (m4 configuration macro) configure:128523: checking for MCA component pmix:external compile mode configure:128529: result: dso configure:129054: checking if MCA component pmix:external can compile configure:129056: result: no ... config.status:3897: creating opal/mca/pmix/Makefile config.status:3897: creating opal/mca/pmix/s1/Makefile config.status:3897: creating opal/mca/pmix/cray/Makefile config.status:3897: creating opal/mca/pmix/s2/Makefile config.status:3897: creating opal/mca/pmix/pmix112/Makefile config.status:3897: creating opal/mca/pmix/external/Makefile ... MCA_BUILD_opal_pmix_cray_DSO_FALSE='#' MCA_BUILD_opal_pmix_cray_DSO_TRUE='' MCA_BUILD_opal_pmix_external_DSO_FALSE='#' MCA_BUILD_opal_pmix_external_DSO_TRUE='' MCA_BUILD_opal_pmix_pmix112_DSO_FALSE='#' MCA_BUILD_opal_pmix_pmix112_DSO_TRUE='' MCA_BUILD_opal_pmix_s1_DSO_FALSE='#' MCA_BUILD_opal_pmix_s1_DSO_TRUE='' MCA_BUILD_opal_pmix_s2_DSO_FALSE='#' MCA_BUILD_opal_pmix_s2_DSO_TRUE='' ... MCA_opal_FRAMEWORKS='common allocator backtrace btl dl event hwloc if installdirs memchecker memcpy memory mpool pmix pstat rcache sec shmem timer' MCA_opal_FRAMEWORKS_SUBDIRS='mca/common mca/allocator mca/backtrace mca/btl mca/dl mca/event mca/hwloc mca/if mca/installdirs mca/memchecker mca/memcpy mca/memory mca/mpool mca/pmix mca/pstat mca/rcache mca/sec mca/shmem mca/timer' MCA_opal_FRAMEWORK_COMPONENT_ALL_SUBDIRS='$(MCA_opal_common_ALL_SUBDIRS) $(MCA_opal_allocator_ALL_SUBDIRS) $(MCA_opal_backtrace_ALL_SUBDIRS) $(MCA_opal_btl_ALL_SUBDIRS) $(MCA_opal_dl_ALL_SUBDIRS) $(MCA_opal_event_ALL_SUBDIRS) $(MCA_opal_hwloc_ALL_SUBDIRS) $(MCA_opal_if_ALL_SUBDIRS) $(MCA_opal_installdirs_ALL_SUBDIRS) $(MCA_opal_memchecker_ALL_SUBDIRS) $(MCA_opal_memcpy_ALL_SUBDIRS) $(MCA_opal_memory_ALL_SUBDIRS) $(MCA_opal_mpool_ALL_SUBDIRS) $(MCA_opal_pmix_ALL_SUBDIRS) $(MCA_opal_pstat_ALL_SUBDIRS) $(MCA_opal_rcache_ALL_SUBDIRS) $(MCA_opal_sec_ALL_SUBDIRS) $(MCA_opal_shmem_ALL_SUBDIRS) $(MCA_opal_timer_ALL_SUBDIRS)' MCA_opal_FRAMEWORK_COMPONENT_DSO_SUBDIRS='$(MCA_opal_common_DSO_SUBDIRS) $(MCA_opal_allocator_DSO_SUBDIRS) $(MCA_opal_backtrace_DSO_SUBDIRS) $(MCA_opal_btl_DSO_SUBDIRS) $(MCA_opal_dl_DSO_SUBDIRS) $(MCA_opal_event_DSO_SUBDIRS) $(MCA_opal_hwloc_DSO_SUBDIRS) $(MCA_opal_if_DSO_SUBDIRS) $(MCA_opal_installdirs_DSO_SUBDIRS) $(MCA_opal_memchecker_DSO_SUBDIRS) $(MCA_opal_memcpy_DSO_SUBDIRS) $(MCA_opal_memory_DSO_SUBDIRS) $(MCA_opal_mpool_DSO_SUBDIRS) $(MCA_opal_pmix_DSO_SUBDIRS) $(MCA_opal_pstat_DSO_SUBDIRS) $(MCA_opal_rcache_DSO_SUBDIRS) $(MCA_opal_sec_DSO_SUBDIRS) $(MCA_opal_shmem_DSO_SUBDIRS) $(MCA_opal_timer_DSO_SUBDIRS)' MCA_opal_FRAMEWORK_COMPONENT_STATIC_SUBDIRS='$(MCA_opal_common_STATIC_SUBDIRS) $(MCA_opal_allocator_STATIC_SUBDIRS) $(MCA_opal_backtrace_STATIC_SUBDIRS) $(MCA_opal_btl_STATIC_SUBDIRS) $(MCA_opal_dl_STATIC_SUBDIRS) $(MCA_opal_event_STATIC_SUBDIRS) $(MCA_opal_hwloc_STATIC_SUBDIRS) $(MCA_opal_if_STATIC_SUBDIRS) $(MCA_opal_installdirs_STATIC_SUBDIRS) $(MCA_opal_memchecker_STATIC_SUBDIRS) $(MCA_opal_memcpy_STATIC_SUBDIRS) $(MCA_opal_memory_STATIC_SUBDIRS) $(MCA_opal_mpool_STATIC_SUBDIRS) $(MCA_opal_pmix_STATIC_SUBDIRS) $(MCA_opal_pstat_STATIC_SUBDIRS) $(MCA_opal_rcache_STATIC_SUBDIRS) $(MCA_opal_sec_STATIC_SUBDIRS) $(MCA_opal_shmem_STATIC_SUBDIRS) $(MCA_opal_timer_STATIC_SUBDIRS)' MCA_opal_FRAMEWORK_LIBS=' $(MCA_opal_common_STATIC_LTLIBS) mca/allocator/libmca_allocator.la $(MCA_opal_allocator_STATIC_LTLIBS) mca/backtrace/libmca_backtrace.la $(MCA_opal_backtrace_STATIC_LTLIBS) mca/btl/libmca_btl.la $(MCA_opal_btl_STATIC_LTLIBS) mca/dl/libmca_dl.la $(MCA_opal_dl_STATIC_LTLIBS) mca/event/libmca_event.la $(MCA_opal_event_STATIC_LTLIBS) mca/hwloc/libmca_hwloc.la $(MCA_opal_hwloc_STATIC_LTLIBS) mca/if/libmca_if.la $(MCA_opal_if_STATIC_LTLIBS) mca/installdirs/libmca_installdirs.la $(MCA_opal_installdirs_STATIC_LTLIBS) mca/memchecker/libmca_memchecker.la $(MCA_opal_memchecker_STATIC_LTLIBS) mca/memcpy/libmca_memcpy.la $(MCA_opal_memcpy_STATIC_LTLIBS) mca/memory/libmca_memory.la $(MCA_opal_memory_STATIC_LTLIBS) mca/mpool/libmca_mpool.la $(MCA_opal_mpool_STATIC_LTLIBS) mca/pmix/libmca_pmix.la $(MCA_opal_pmix_STATIC_LTLIBS) mca/pstat/libmca_pstat.la $(MCA_opal_pstat_STATIC_LTLIBS) mca/rcache/libmca_rcache.la $(MCA_opal_rcache_STATIC_LTLIBS) mca/sec/libmca_sec.la $(MCA_opal_sec_STATIC_LTLIBS) mca/shmem/libmca_shmem.la $(MCA_opal_shmem_STATIC_LTLIBS) mca/timer/libmca_timer.la $(MCA_opal_timer_STATIC_LTLIBS)' ... MCA_opal_pmix_ALL_COMPONENTS=' s1 cray s2 pmix112 external' MCA_opal_pmix_ALL_SUBDIRS=' mca/pmix/s1 mca/pmix/cray mca/pmix/s2 mca/pmix/pmix112 mca/pmix/external' MCA_opal_pmix_DSO_COMPONENTS=' pmix112' MCA_opal_pmix_DSO_SUBDIRS=' mca/pmix/pmix112' MCA_opal_pmix_STATIC_COMPONENTS='' MCA_opal_pmix_STATIC_LTLIBS='' MCA_opal_pmix_STATIC_SUBDIRS='' ... opal_pmix_ext_CPPFLAGS='' opal_pmix_ext_LDFLAGS='' opal_pmix_ext_LIBS='' opal_pmix_pmix112_CPPFLAGS='-I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/include/pmix -I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/include -I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix -I$(OPAL_TOP_SRCDIR)/opal/mca/pmix/pmix112/pmix' opal_pmix_pmix112_LIBS='$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/libpmix.la' ... "config.log" for Sun C 5.13 shows the following. ... configure:127803: /bin/bash '../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/ pmix/configure' *failed* for opal/mca/pmix/pmix112/pmix configure:128379: checking if MCA component pmix:pmix112 can compile configure:128381: result: no configure:5637: --- MCA component pmix:external (m4 configuration macro) configure:128523: checking for MCA component pmix:external compile mode configure:128529: result: dso configure:129054: checking if MCA component pmix:external can compile configure:129056: result: no ... config.status:3887: creating opal/mca/pmix/Makefile config.status:3887: creating opal/mca/pmix/s1/Makefile config.status:3887: creating opal/mca/pmix/cray/Makefile config.status:3887: creating opal/mca/pmix/s2/Makefile config.status:3887: creating opal/mca/pmix/pmix112/Makefile config.status:3887: creating opal/mca/pmix/external/Makefile ... MCA_BUILD_opal_pmix_cray_DSO_FALSE='#' MCA_BUILD_opal_pmix_cray_DSO_TRUE='' MCA_BUILD_opal_pmix_external_DSO_FALSE='#' MCA_BUILD_opal_pmix_external_DSO_TRUE='' MCA_BUILD_opal_pmix_pmix112_DSO_FALSE='#' MCA_BUILD_opal_pmix_pmix112_DSO_TRUE='' MCA_BUILD_opal_pmix_s1_DSO_FALSE='#' MCA_BUILD_opal_pmix_s1_DSO_TRUE='' MCA_BUILD_opal_pmix_s2_DSO_FALSE='#' MCA_BUILD_opal_pmix_s2_DSO_TRUE='' ... MCA_opal_FRAMEWORKS='common allocator backtrace btl dl event hwloc if installdirs memchecker memcpy memory mpool pmix pstat rcache sec shmem timer' MCA_opal_FRAMEWORKS_SUBDIRS='mca/common mca/allocator mca/backtrace mca/btl mca/dl mca/event mca/hwloc mca/if mca/installdirs mca/memchecker mca/memcpy mca/memory mca/mpool mca/pmix mca/pstat mca/rcache mca/sec mca/shmem mca/timer' MCA_opal_FRAMEWORK_COMPONENT_ALL_SUBDIRS='$(MCA_opal_common_ALL_SUBDIRS) $(MCA_opal_allocator_ALL_SUBDIRS) $(MCA_opal_backtrace_ALL_SUBDIRS) $(MCA_opal_btl_ALL_SUBDIRS) $(MCA_opal_dl_ALL_SUBDIRS) $(MCA_opal_event_ALL_SUBDIRS) $(MCA_opal_hwloc_ALL_SUBDIRS) $(MCA_opal_if_ALL_SUBDIRS) $(MCA_opal_installdirs_ALL_SUBDIRS) $(MCA_opal_memchecker_ALL_SUBDIRS) $(MCA_opal_memcpy_ALL_SUBDIRS) $(MCA_opal_memory_ALL_SUBDIRS) $(MCA_opal_mpool_ALL_SUBDIRS) $(MCA_opal_pmix_ALL_SUBDIRS) $(MCA_opal_pstat_ALL_SUBDIRS) $(MCA_opal_rcache_ALL_SUBDIRS) $(MCA_opal_sec_ALL_SUBDIRS) $(MCA_opal_shmem_ALL_SUBDIRS) $(MCA_opal_timer_ALL_SUBDIRS)' MCA_opal_FRAMEWORK_COMPONENT_DSO_SUBDIRS='$(MCA_opal_common_DSO_SUBDIRS) $(MCA_opal_allocator_DSO_SUBDIRS) $(MCA_opal_backtrace_DSO_SUBDIRS) $(MCA_opal_btl_DSO_SUBDIRS) $(MCA_opal_dl_DSO_SUBDIRS) $(MCA_opal_event_DSO_SUBDIRS) $(MCA_opal_hwloc_DSO_SUBDIRS) $(MCA_opal_if_DSO_SUBDIRS) $(MCA_opal_installdirs_DSO_SUBDIRS) $(MCA_opal_memchecker_DSO_SUBDIRS) $(MCA_opal_memcpy_DSO_SUBDIRS) $(MCA_opal_memory_DSO_SUBDIRS) $(MCA_opal_mpool_DSO_SUBDIRS) $(MCA_opal_pmix_DSO_SUBDIRS) $(MCA_opal_pstat_DSO_SUBDIRS) $(MCA_opal_rcache_DSO_SUBDIRS) $(MCA_opal_sec_DSO_SUBDIRS) $(MCA_opal_shmem_DSO_SUBDIRS) $(MCA_opal_timer_DSO_SUBDIRS)' MCA_opal_FRAMEWORK_COMPONENT_STATIC_SUBDIRS='$(MCA_opal_common_STATIC_SUBDIRS) $(MCA_opal_allocator_STATIC_SUBDIRS) $(MCA_opal_backtrace_STATIC_SUBDIRS) $(MCA_opal_btl_STATIC_SUBDIRS) $(MCA_opal_dl_STATIC_SUBDIRS) $(MCA_opal_event_STATIC_SUBDIRS) $(MCA_opal_hwloc_STATIC_SUBDIRS) $(MCA_opal_if_STATIC_SUBDIRS) $(MCA_opal_installdirs_STATIC_SUBDIRS) $(MCA_opal_memchecker_STATIC_SUBDIRS) $(MCA_opal_memcpy_STATIC_SUBDIRS) $(MCA_opal_memory_STATIC_SUBDIRS) $(MCA_opal_mpool_STATIC_SUBDIRS) $(MCA_opal_pmix_STATIC_SUBDIRS) $(MCA_opal_pstat_STATIC_SUBDIRS) $(MCA_opal_rcache_STATIC_SUBDIRS) $(MCA_opal_sec_STATIC_SUBDIRS) $(MCA_opal_shmem_STATIC_SUBDIRS) $(MCA_opal_timer_STATIC_SUBDIRS)' MCA_opal_FRAMEWORK_LIBS=' $(MCA_opal_common_STATIC_LTLIBS) mca/allocator/libmca_allocator.la $(MCA_opal_allocator_STATIC_LTLIBS) mca/backtrace/libmca_backtrace.la $(MCA_opal_backtrace_STATIC_LTLIBS) mca/btl/libmca_btl.la $(MCA_opal_btl_STATIC_LTLIBS) mca/dl/libmca_dl.la $(MCA_opal_dl_STATIC_LTLIBS) mca/event/libmca_event.la $(MCA_opal_event_STATIC_LTLIBS) mca/hwloc/libmca_hwloc.la $(MCA_opal_hwloc_STATIC_LTLIBS) mca/if/libmca_if.la $(MCA_opal_if_STATIC_LTLIBS) mca/installdirs/libmca_installdirs.la $(MCA_opal_installdirs_STATIC_LTLIBS) mca/memchecker/libmca_memchecker.la $(MCA_opal_memchecker_STATIC_LTLIBS) mca/memcpy/libmca_memcpy.la $(MCA_opal_memcpy_STATIC_LTLIBS) mca/memory/libmca_memory.la $(MCA_opal_memory_STATIC_LTLIBS) mca/mpool/libmca_mpool.la $(MCA_opal_mpool_STATIC_LTLIBS) mca/pmix/libmca_pmix.la $(MCA_opal_pmix_STATIC_LTLIBS) mca/pstat/libmca_pstat.la $(MCA_opal_pstat_STATIC_LTLIBS) mca/rcache/libmca_rcache.la $(MCA_opal_rcache_STATIC_LTLIBS) mca/sec/libmca_sec.la $(MCA_opal_sec_STATIC_LTLIBS) mca/shmem/libmca_shmem.la $(MCA_opal_shmem_STATIC_LTLIBS) mca/timer/libmca_timer.la $(MCA_opal_timer_STATIC_LTLIBS)' ... MCA_opal_pmix_ALL_COMPONENTS=' s1 cray s2 pmix112 external' MCA_opal_pmix_ALL_SUBDIRS=' mca/pmix/s1 mca/pmix/cray mca/pmix/s2 mca/pmix/pmix112 mca/pmix/external' MCA_opal_pmix_DSO_COMPONENTS='' MCA_opal_pmix_DSO_SUBDIRS='' MCA_opal_pmix_STATIC_COMPONENTS='' MCA_opal_pmix_STATIC_LTLIBS='' MCA_opal_pmix_STATIC_SUBDIRS='' ... opal_pmix_ext_CPPFLAGS='' opal_pmix_ext_LDFLAGS='' opal_pmix_ext_LIBS='' opal_pmix_pmix112_CPPFLAGS='' opal_pmix_pmix112_LIBS='' ... I've attached the config.log files for pmix. tyr openmpi-2.0.0 142 tar zvft pmix_config.log.tar.gz -rw-r--r-- root/root 136291 2016-04-25 08:05:34 openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/pmix/pmix112/pmix/config.log -rw-r--r-- root/root 528808 2016-04-25 08:07:54 openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_gcc/opal/mca/pmix/pmix112/pmix/config.log tyr openmpi-2.0.0 143 I've also attached the output for the broken execution of "spawn_multiple_master" for my gcc-version of Open MPI. "spawn_master" works as expected with my gcc-version of Open MPI. Hopefully you can fix the problem. Kind regards and thank you very much for your help Siegmar Am 23.04.2016 um 21:34 schrieb Siegmar Gross:Hi Gilles, I don't know what happened, but the files are not available now and they were definitely available when I answered the email from Ralph. The files also have a different timestamp now. This is an extract from my email to Ralph for Solaris Sparc. -rwxr-xr-x 1 root root 977 Apr 19 19:49 mca_plm_rsh.la -rwxr-xr-x 1 root root 153280 Apr 19 19:49 mca_plm_rsh.so -rwxr-xr-x 1 root root 1007 Apr 19 19:47 mca_pmix_pmix112.la -rwxr-xr-x 1 root root 1400512 Apr 19 19:47 mca_pmix_pmix112.so -rwxr-xr-x 1 root root 971 Apr 19 19:52 mca_pml_cm.la -rwxr-xr-x 1 root root 342440 Apr 19 19:52 mca_pml_cm.so Now I have the following output for these files. -rwxr-xr-x 1 root root 976 Apr 19 19:58 mca_plm_rsh.la -rwxr-xr-x 1 root root 319816 Apr 19 19:58 mca_plm_rsh.so -rwxr-xr-x 1 root root 970 Apr 19 20:00 mca_pml_cm.la -rwxr-xr-x 1 root root 1507440 Apr 19 20:00 mca_pml_cm.so I'll try to find out what happened next week when I'm back in my office. Kind regards Siegmar Am 23.04.16 um 02:12 schrieb Gilles Gouaillardet:Siegmar, I will try to reproduce this on my solaris11 x86_64 vm In the mean time, can you please double check mca_pmix_pmix_pmix112.so is a 64 bits library ? (E.g, confirm "-m64" was correctly passed to pmix) Cheers, Gilles On Friday, April 22, 2016, Siegmar Gross <siegmar.gr...@informatik.hs-fulda.de <mailto:siegmar.gr...@informatik.hs-fulda.de>> wrote: Hi Ralph, I've already used "-enable-debug". "SYSTEM_ENV" is "SunOS" or "Linux" and "MACHINE_ENV" is "sparc" or "x86_84". mkdir openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_gcc cd openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_gcc ../openmpi-v2.x-dev-1280-gc110ae8/configure \ --prefix=/usr/local/openmpi-2.0.0_64_gcc \ --libdir=/usr/local/openmpi-2.0.0_64_gcc/lib64 \ --with-jdk-bindir=/usr/local/jdk1.8.0/bin \ --with-jdk-headers=/usr/local/jdk1.8.0/include \ JAVA_HOME=/usr/local/jdk1.8.0 \ LDFLAGS="-m64" CC="gcc" CXX="g++" FC="gfortran" \ CFLAGS="-m64" CXXFLAGS="-m64" FCFLAGS="-m64" \ CPP="cpp" CXXCPP="cpp" \ --enable-mpi-cxx \ --enable-cxx-exceptions \ --enable-mpi-java \ --enable-heterogeneous \ --enable-mpi-thread-multiple \ --with-hwloc=internal \ --without-verbs \ --with-wrapper-cflags="-std=c11 -m64" \ --with-wrapper-cxxflags="-m64" \ --with-wrapper-fcflags="-m64" \ --enable-debug \ |& tee log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_gcc mkdir openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_cc cd openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_cc ../openmpi-v2.x-dev-1280-gc110ae8/configure \ --prefix=/usr/local/openmpi-2.0.0_64_cc \ --libdir=/usr/local/openmpi-2.0.0_64_cc/lib64 \ --with-jdk-bindir=/usr/local/jdk1.8.0/bin \ --with-jdk-headers=/usr/local/jdk1.8.0/include \ JAVA_HOME=/usr/local/jdk1.8.0 \ LDFLAGS="-m64" CC="cc" CXX="CC" FC="f95" \ CFLAGS="-m64" CXXFLAGS="-m64 -library=stlport4" FCFLAGS="-m64" \ CPP="cpp" CXXCPP="cpp" \ --enable-mpi-cxx \ --enable-cxx-exceptions \ --enable-mpi-java \ --enable-heterogeneous \ --enable-mpi-thread-multiple \ --with-hwloc=internal \ --without-verbs \ --with-wrapper-cflags="-m64" \ --with-wrapper-cxxflags="-m64 -library=stlport4" \ --with-wrapper-fcflags="-m64" \ --with-wrapper-ldflags="" \ --enable-debug \ |& tee log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_cc Kind regards Siegmar Am 21.04.2016 um 18:18 schrieb Ralph Castain: Can you please rebuild OMPI with -enable-debug in the configure cmd? It will let us see more error output On Apr 21, 2016, at 8:52 AM, Siegmar Gross <siegmar.gr...@informatik.hs-fulda.de> wrote: Hi Ralph, I don't see any additional information. tyr hello_1 108 mpiexec -np 4 --host tyr,sunpc1,linpc1,ruester -mca mca_base_component_show_load_errors 1 hello_1_mpi [tyr.informatik.hs-fulda.de:06211 <http://tyr.informatik.hs-fulda.de:06211>] [[48741,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c at line 638 -------------------------------------------------------------------------- It looks like orte_init failed for some reason; your parallel process is likely to abort. There are many reasons that a parallel process can fail during orte_init; some of which are due to configuration or environment problems. This failure appears to be an internal failure; here's some additional information (which may only be relevant to an Open MPI developer): opal_pmix_base_select failed --> Returned value Not found (-13) instead of ORTE_SUCCESS -------------------------------------------------------------------------- tyr hello_1 109 mpiexec -np 4 --host tyr,sunpc1,linpc1,ruester -mca mca_base_component_show_load_errors 1 -mca pmix_base_verbose 10 -mca pmix_server_verbose 5 hello_1_mpi [tyr.informatik.hs-fulda.de:06212 <http://tyr.informatik.hs-fulda.de:06212>] mca: base: components_register: registering framework pmix components [tyr.informatik.hs-fulda.de:06212 <http://tyr.informatik.hs-fulda.de:06212>] mca: base: components_open: opening pmix components [tyr.informatik.hs-fulda.de:06212 <http://tyr.informatik.hs-fulda.de:06212>] mca:base:select: Auto-selecting pmix components [tyr.informatik.hs-fulda.de:06212 <http://tyr.informatik.hs-fulda.de:06212>] mca:base:select:( pmix) No component selected! [tyr.informatik.hs-fulda.de:06212 <http://tyr.informatik.hs-fulda.de:06212>] [[48738,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c at line 638 -------------------------------------------------------------------------- It looks like orte_init failed for some reason; your parallel process is likely to abort. There are many reasons that a parallel process can fail during orte_init; some of which are due to configuration or environment problems. This failure appears to be an internal failure; here's some additional information (which may only be relevant to an Open MPI developer): opal_pmix_base_select failed --> Returned value Not found (-13) instead of ORTE_SUCCESS -------------------------------------------------------------------------- tyr hello_1 110 Kind regards Siegmar Am 21.04.2016 um 17:24 schrieb Ralph Castain: Hmmm…it looks like you built the right components, but they are not being picked up. Can you run your mpiexec command again, adding “-mca mca_base_component_show_load_errors 1” to the cmd line? On Apr 21, 2016, at 8:16 AM, Siegmar Gross <siegmar.gr...@informatik.hs-fulda.de> wrote: Hi Ralph, I have attached ompi_info output for both compilers from my sparc machine and the listings for both compilers from the <prefix>/lib/openmpi directories. Hopefully that helps to find the problem. hermes tmp 3 tar zvft openmpi-2.x_info.tar.gz -rw-r--r-- root/root 10969 2016-04-21 17:06 ompi_info_SunOS_sparc_cc.txt -rw-r--r-- root/root 11044 2016-04-21 17:06 ompi_info_SunOS_sparc_gcc.txt -rw-r--r-- root/root 71252 2016-04-21 17:02 lib64_openmpi.txt hermes tmp 4 Kind regards and thank you very much once more for your help Siegmar Am 21.04.2016 um 15:54 schrieb Ralph Castain: Odd - it would appear that none of the pmix components built? Can you send along the output from ompi_info? Or just send a listing of the files in the <prefix>/lib/openmpi directory? On Apr 21, 2016, at 1:27 AM, Siegmar Gross <siegmar.gr...@informatik.hs-fulda.de <mailto:siegmar.gr...@informatik.hs-fulda.de>> wrote: Hi Ralph, Am 21.04.2016 um 00:18 schrieb Ralph Castain: Could you please rerun these test and add “-mca pmix_base_verbose 10 -mca pmix_server_verbose 5” to your cmd line? I need to see why the pmix components failed. tyr spawn 111 mpiexec -np 1 --host tyr,sunpc1,linpc1,ruester -mca pmix_base_verbose 10 -mca pmix_server_verbose 5 spawn_multiple_master [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de> <http://tyr.informatik.hs-fulda.de/>:26652] mca: base: components_register: registering framework pmix components [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de> <http://tyr.informatik.hs-fulda.de/>:26652] mca: base: components_open: opening pmix components [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de> <http://tyr.informatik.hs-fulda.de/>:26652] mca:base:select: Auto-selecting pmix components [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de> <http://tyr.informatik.hs-fulda.de/>:26652] mca:base:select:( pmix) No component selected! [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de> <http://tyr.informatik.hs-fulda.de/>:26652] [[52794,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c at line 638 -------------------------------------------------------------------------- It looks like orte_init failed for some reason; your parallel process is likely to abort. There are many reasons that a parallel process can fail during orte_init; some of which are due to configuration or environment problems. This failure appears to be an internal failure; here's some additional information (which may only be relevant to an Open MPI developer): opal_pmix_base_select failed --> Returned value Not found (-13) instead of ORTE_SUCCESS -------------------------------------------------------------------------- tyr spawn 112 tyr hello_1 116 mpiexec -np 1 --host tyr,sunpc1,linpc1,ruester -mca pmix_base_verbose 10 -mca pmix_server_verbose 5 hello_1_mpi [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de> <http://tyr.informatik.hs-fulda.de/>:27261] mca: base: components_register: registering framework pmix components [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de> <http://tyr.informatik.hs-fulda.de/>:27261] mca: base: components_open: opening pmix components [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de> <http://tyr.informatik.hs-fulda.de/>:27261] mca:base:select: Auto-selecting pmix components [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de> <http://tyr.informatik.hs-fulda.de/>:27261] mca:base:select:( pmix) No component selected! [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de> <http://tyr.informatik.hs-fulda.de/>:27261] [[52315,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c at line 638 -------------------------------------------------------------------------- It looks like orte_init failed for some reason; your parallel process is likely to abort. There are many reasons that a parallel process can fail during orte_init; some of which are due to configuration or environment problems. This failure appears to be an internal failure; here's some additional information (which may only be relevant to an Open MPI developer): opal_pmix_base_select failed --> Returned value Not found (-13) instead of ORTE_SUCCESS -------------------------------------------------------------------------- tyr hello_1 117 Thank you very much for your help. Kind regards Siegmar Thanks Ralph On Apr 20, 2016, at 10:12 AM, Siegmar Gross <siegmar.gr...@informatik.hs-fulda.de <mailto:siegmar.gr...@informatik.hs-fulda.de>> wrote: Hi, I have built openmpi-v2.x-dev-1280-gc110ae8 on my machines (Solaris 10 Sparc, Solaris 10 x86_64, and openSUSE Linux 12.1 x86_64) with gcc-5.1.0 and Sun C 5.13. Unfortunately I get runtime errors for some programs. Sun C 5.13: =========== For all my test programs I get the same error on Solaris Sparc and Solaris x86_64, while the programs work fine on Linux. tyr hello_1 115 mpiexec -np 2 hello_1_mpi [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de> <http://tyr.informatik.hs-fulda.de>:22373] [[61763,0],0] ORTE_ERROR_LOG: Not found in file ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c at line 638 -------------------------------------------------------------------------- It looks like orte_init failed for some reason; your parallel process is likely to abort. There are many reasons that a parallel process can fail during orte_init; some of which are due to configuration or environment problems. This failure appears to be an internal failure; here's some additional information (which may only be relevant to an Open MPI developer): opal_pmix_base_select failed --> Returned value Not found (-13) instead of ORTE_SUCCESS -------------------------------------------------------------------------- tyr hello_1 116 GCC-5.1.0: ========== tyr spawn 121 mpiexec -np 1 --host tyr,sunpc1,linpc1,ruester spawn_multiple_master Parent process 0 running on tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de> <http://tyr.informatik.hs-fulda.de> I create 3 slave processes. [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de> <http://tyr.informatik.hs-fulda.de>:25366] PMIX ERROR: UNPACK-PAST-END in file ../../../../../../openmpi-v2.x-dev-1280-gc110ae8/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c at line 829 [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de> <http://tyr.informatik.hs-fulda.de>:25366] PMIX ERROR: UNPACK-PAST-END in file ../../../../../../openmpi-v2.x-dev-1280-gc110ae8/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c at line 2176 [tyr:25377] *** An error occurred in MPI_Comm_spawn_multiple [tyr:25377] *** reported by process [3308257281,0] [tyr:25377] *** on communicator MPI_COMM_WORLD [tyr:25377] *** MPI_ERR_SPAWN: could not spawn processes [tyr:25377] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort, [tyr:25377] *** and potentially your MPI job) tyr spawn 122 I would be grateful if somebody can fix the problems. Thank you very much for any help in advance. Kind regards Siegmar <hello_1_mpi.c><spawn_multiple_master.c>_______________________________________________ users mailing list us...@open-mpi.org <mailto:us...@open-mpi.org> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users Link to this post: http://www.open-mpi.org/community/lists/users/2016/04/28983.php _______________________________________________ users mailing list us...@open-mpi.org <mailto:us...@open-mpi.org> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users Link to this post: http://www.open-mpi.org/community/lists/users/2016/04/28986.php _______________________________________________ users mailing list us...@open-mpi.org <mailto:us...@open-mpi.org> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users Link to this post: http://www.open-mpi.org/community/lists/users/2016/04/28987.php _______________________________________________ users mailing list us...@open-mpi.org Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users Link to this post: http://www.open-mpi.org/community/lists/users/2016/04/28988.php <openmpi-2.x_info.tar.gz>_______________________________________________ users mailing list us...@open-mpi.org Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users Link to this post: http://www.open-mpi.org/community/lists/users/2016/04/28989.php _______________________________________________ users mailing list us...@open-mpi.org Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users Link to this post: http://www.open-mpi.org/community/lists/users/2016/04/28990.php _______________________________________________ users mailing list us...@open-mpi.org Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users Link to this post: http://www.open-mpi.org/community/lists/users/2016/04/28991.php _______________________________________________ users mailing list us...@open-mpi.org Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users Link to this post: http://www.open-mpi.org/community/lists/users/2016/04/28992.php _______________________________________________ users mailing list us...@open-mpi.org Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users Link to this post: http://www.open-mpi.org/community/lists/users/2016/04/28993.php _______________________________________________ users mailing list us...@open-mpi.org Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users Link to this post: http://www.open-mpi.org/community/lists/users/2016/04/28999.php_______________________________________________ users mailing list us...@open-mpi.org Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users Link to this post: http://www.open-mpi.org/community/lists/users/2016/04/29009.php_______________________________________________ users mailing list us...@open-mpi.org Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users Link to this post: http://www.open-mpi.org/community/lists/users/2016/04/29033.php_______________________________________________ users mailing list us...@open-mpi.org Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users Link to this post: http://www.open-mpi.org/community/lists/users/2016/04/29038.php
-- ######################################################################## # # # Hochschule Fulda University of Applied Sciences # # FB Angewandte Informatik Department of Applied Computer Science # # # # Prof. Dr. Siegmar Gross Tel.: +49 (0)661 9640 - 333 # # Fax: +49 (0)661 9640 - 349 # # Leipziger Str. 123 WWW: http://www.hs-fulda.de/~gross # # D-36037 Fulda Mail: siegmar.gr...@informatik.hs-fulda.de # # # # # # IT-Sicherheit: http://www.hs-fulda.de/it-sicherheit # # # ########################################################################
smime.p7s
Description: S/MIME Cryptographic Signature