Siegmar,

here is the error :

configure:17969: cc -o conftest -m64 -D_REENTRANT -g -g -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1 -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/include -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/include -D_REENTRANT -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/hwloc/hwloc1112/hwloc/include -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/hwloc/hwloc1112/hwloc/include -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/event/libevent2022/libevent -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/event/libevent2022/libevent/include -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/event/libevent2022/libevent/include -m64 conftest.c >&5 "/usr/include/stdbool.h", line 42: #error: "Use of <stdbool.h> is valid only in a c99 compilation environment."


i cannot reproduce this on solaris 11 with oracle studio 5.3 compiler, and i do not have solaris 10 yet.

could you please re-configure with '-std=c99' appended to your CFLAGS and see if it helps ?


Cheers,


Gilles


On 4/26/2016 7:57 PM, Siegmar Gross wrote:
Hi Gilles and Ralph,

I was able to sort out my mess. In my last email I compared the
files from "SunOS_sparc/openmpi-2.0.0_64_gcc/lib64/openmpi" from
the attachment of my email to Ralph with the files from
"SunOS_sparc/openmpi-2.0.0_64_cc/lib64/openmpi" from my current
file system. That's the reason while I have had different
timestamps. The other problem was that Ralph didn't recognize
that "mca_pmix_pmix112.so" wasn't built on Solaris with the
Sun C compiler. I've removed most of the files from the attachment
of my email so that it is easier to see the relevant files. Below
I try to give you more information that may be relevant to track
down the problem. I still get an error running one of my small
test programs, when I use my gcc-version of Open MPI.
"mca_pmix_pmix112.so" is a 64 bits library.

Linux_x86_64/openmpi-2.0.0_64_cc/lib64/openmpi:
...
-rwxr-xr-x 1 root root  261327 Apr 19 16:46 mca_plm_slurm.so
-rwxr-xr-x 1 root root    1002 Apr 19 16:45 mca_pmix_pmix112.la
-rwxr-xr-x 1 root root 3906526 Apr 19 16:45 mca_pmix_pmix112.so
-rwxr-xr-x 1 root root     966 Apr 19 16:51 mca_pml_cm.la
-rwxr-xr-x 1 root root 1574265 Apr 19 16:51 mca_pml_cm.so
...

Linux_x86_64/openmpi-2.0.0_64_gcc/lib64/openmpi:
...
-rwxr-xr-x 1 root root   70371 Apr 19 16:43 mca_plm_slurm.so
-rwxr-xr-x 1 root root    1008 Apr 19 16:42 mca_pmix_pmix112.la
-rwxr-xr-x 1 root root 1029005 Apr 19 16:42 mca_pmix_pmix112.so
-rwxr-xr-x 1 root root     972 Apr 19 16:46 mca_pml_cm.la
-rwxr-xr-x 1 root root  284858 Apr 19 16:46 mca_pml_cm.so
...

SunOS_sparc/openmpi-2.0.0_64_cc/lib64/openmpi:
...
-rwxr-xr-x 1 root root  319816 Apr 19 19:58 mca_plm_rsh.so
-rwxr-xr-x 1 root root     970 Apr 19 20:00 mca_pml_cm.la
-rwxr-xr-x 1 root root 1507440 Apr 19 20:00 mca_pml_cm.so
...

SunOS_sparc/openmpi-2.0.0_64_gcc/lib64/openmpi:
...
-rwxr-xr-x 1 root root  153280 Apr 19 19:49 mca_plm_rsh.so
-rwxr-xr-x 1 root root    1007 Apr 19 19:47 mca_pmix_pmix112.la
-rwxr-xr-x 1 root root 1400512 Apr 19 19:47 mca_pmix_pmix112.so
-rwxr-xr-x 1 root root     971 Apr 19 19:52 mca_pml_cm.la
-rwxr-xr-x 1 root root  342440 Apr 19 19:52 mca_pml_cm.so
...

SunOS_x86_64/openmpi-2.0.0_64_cc/lib64/openmpi:
...
-rwxr-xr-x 1 root root  300096 Apr 19 17:18 mca_plm_rsh.so
-rwxr-xr-x 1 root root     970 Apr 19 17:23 mca_pml_cm.la
-rwxr-xr-x 1 root root 1458816 Apr 19 17:23 mca_pml_cm.so
...

SunOS_x86_64/openmpi-2.0.0_64_gcc/lib64/openmpi:
...
-rwxr-xr-x 1 root root  133096 Apr 19 17:42 mca_plm_rsh.so
-rwxr-xr-x 1 root root    1007 Apr 19 17:41 mca_pmix_pmix112.la
-rwxr-xr-x 1 root root 1320240 Apr 19 17:41 mca_pmix_pmix112.so
-rwxr-xr-x 1 root root     971 Apr 19 17:46 mca_pml_cm.la
-rwxr-xr-x 1 root root  419848 Apr 19 17:46 mca_pml_cm.so
...


Yesterday I've installed openmpi-v2.x-dev-1290-gbd0e4e1 so that we
have a current version for the investigation of the problem. Once
more mca_pmix_pmix112.so isn't available on Solaris if I use the
Sun C compiler.

"config.log" for gcc-5.1.0 shows the following.

...
configure:127799: /bin/bash '../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/
pmix/configure' succeeded for opal/mca/pmix/pmix112/pmix
configure:127916: checking if MCA component pmix:pmix112 can compile
configure:127918: result: yes
configure:5637: --- MCA component pmix:external (m4 configuration macro)
configure:128523: checking for MCA component pmix:external compile mode
configure:128529: result: dso
configure:129054: checking if MCA component pmix:external can compile
configure:129056: result: no
...
config.status:3897: creating opal/mca/pmix/Makefile
config.status:3897: creating opal/mca/pmix/s1/Makefile
config.status:3897: creating opal/mca/pmix/cray/Makefile
config.status:3897: creating opal/mca/pmix/s2/Makefile
config.status:3897: creating opal/mca/pmix/pmix112/Makefile
config.status:3897: creating opal/mca/pmix/external/Makefile
...
MCA_BUILD_opal_pmix_cray_DSO_FALSE='#'
MCA_BUILD_opal_pmix_cray_DSO_TRUE=''
MCA_BUILD_opal_pmix_external_DSO_FALSE='#'
MCA_BUILD_opal_pmix_external_DSO_TRUE=''
MCA_BUILD_opal_pmix_pmix112_DSO_FALSE='#'
MCA_BUILD_opal_pmix_pmix112_DSO_TRUE=''
MCA_BUILD_opal_pmix_s1_DSO_FALSE='#'
MCA_BUILD_opal_pmix_s1_DSO_TRUE=''
MCA_BUILD_opal_pmix_s2_DSO_FALSE='#'
MCA_BUILD_opal_pmix_s2_DSO_TRUE=''
...
MCA_opal_FRAMEWORKS='common allocator backtrace btl dl event hwloc if installdirs memchecker memcpy memory mpool pmix pstat rcache sec shmem timer' MCA_opal_FRAMEWORKS_SUBDIRS='mca/common mca/allocator mca/backtrace mca/btl mca/dl mca/event mca/hwloc mca/if mca/installdirs mca/memchecker mca/memcpy mca/memory mca/mpool mca/pmix mca/pstat mca/rcache mca/sec mca/shmem mca/timer' MCA_opal_FRAMEWORK_COMPONENT_ALL_SUBDIRS='$(MCA_opal_common_ALL_SUBDIRS) $(MCA_opal_allocator_ALL_SUBDIRS) $(MCA_opal_backtrace_ALL_SUBDIRS) $(MCA_opal_btl_ALL_SUBDIRS) $(MCA_opal_dl_ALL_SUBDIRS) $(MCA_opal_event_ALL_SUBDIRS) $(MCA_opal_hwloc_ALL_SUBDIRS) $(MCA_opal_if_ALL_SUBDIRS) $(MCA_opal_installdirs_ALL_SUBDIRS) $(MCA_opal_memchecker_ALL_SUBDIRS) $(MCA_opal_memcpy_ALL_SUBDIRS) $(MCA_opal_memory_ALL_SUBDIRS) $(MCA_opal_mpool_ALL_SUBDIRS) $(MCA_opal_pmix_ALL_SUBDIRS) $(MCA_opal_pstat_ALL_SUBDIRS) $(MCA_opal_rcache_ALL_SUBDIRS) $(MCA_opal_sec_ALL_SUBDIRS) $(MCA_opal_shmem_ALL_SUBDIRS) $(MCA_opal_timer_ALL_SUBDIRS)' MCA_opal_FRAMEWORK_COMPONENT_DSO_SUBDIRS='$(MCA_opal_common_DSO_SUBDIRS) $(MCA_opal_allocator_DSO_SUBDIRS) $(MCA_opal_backtrace_DSO_SUBDIRS) $(MCA_opal_btl_DSO_SUBDIRS) $(MCA_opal_dl_DSO_SUBDIRS) $(MCA_opal_event_DSO_SUBDIRS) $(MCA_opal_hwloc_DSO_SUBDIRS) $(MCA_opal_if_DSO_SUBDIRS) $(MCA_opal_installdirs_DSO_SUBDIRS) $(MCA_opal_memchecker_DSO_SUBDIRS) $(MCA_opal_memcpy_DSO_SUBDIRS) $(MCA_opal_memory_DSO_SUBDIRS) $(MCA_opal_mpool_DSO_SUBDIRS) $(MCA_opal_pmix_DSO_SUBDIRS) $(MCA_opal_pstat_DSO_SUBDIRS) $(MCA_opal_rcache_DSO_SUBDIRS) $(MCA_opal_sec_DSO_SUBDIRS) $(MCA_opal_shmem_DSO_SUBDIRS) $(MCA_opal_timer_DSO_SUBDIRS)' MCA_opal_FRAMEWORK_COMPONENT_STATIC_SUBDIRS='$(MCA_opal_common_STATIC_SUBDIRS) $(MCA_opal_allocator_STATIC_SUBDIRS) $(MCA_opal_backtrace_STATIC_SUBDIRS) $(MCA_opal_btl_STATIC_SUBDIRS) $(MCA_opal_dl_STATIC_SUBDIRS) $(MCA_opal_event_STATIC_SUBDIRS) $(MCA_opal_hwloc_STATIC_SUBDIRS) $(MCA_opal_if_STATIC_SUBDIRS) $(MCA_opal_installdirs_STATIC_SUBDIRS) $(MCA_opal_memchecker_STATIC_SUBDIRS) $(MCA_opal_memcpy_STATIC_SUBDIRS) $(MCA_opal_memory_STATIC_SUBDIRS) $(MCA_opal_mpool_STATIC_SUBDIRS) $(MCA_opal_pmix_STATIC_SUBDIRS) $(MCA_opal_pstat_STATIC_SUBDIRS) $(MCA_opal_rcache_STATIC_SUBDIRS) $(MCA_opal_sec_STATIC_SUBDIRS) $(MCA_opal_shmem_STATIC_SUBDIRS) $(MCA_opal_timer_STATIC_SUBDIRS)' MCA_opal_FRAMEWORK_LIBS=' $(MCA_opal_common_STATIC_LTLIBS) mca/allocator/libmca_allocator.la $(MCA_opal_allocator_STATIC_LTLIBS) mca/backtrace/libmca_backtrace.la $(MCA_opal_backtrace_STATIC_LTLIBS) mca/btl/libmca_btl.la $(MCA_opal_btl_STATIC_LTLIBS) mca/dl/libmca_dl.la $(MCA_opal_dl_STATIC_LTLIBS) mca/event/libmca_event.la $(MCA_opal_event_STATIC_LTLIBS) mca/hwloc/libmca_hwloc.la $(MCA_opal_hwloc_STATIC_LTLIBS) mca/if/libmca_if.la $(MCA_opal_if_STATIC_LTLIBS) mca/installdirs/libmca_installdirs.la $(MCA_opal_installdirs_STATIC_LTLIBS) mca/memchecker/libmca_memchecker.la $(MCA_opal_memchecker_STATIC_LTLIBS) mca/memcpy/libmca_memcpy.la $(MCA_opal_memcpy_STATIC_LTLIBS) mca/memory/libmca_memory.la $(MCA_opal_memory_STATIC_LTLIBS) mca/mpool/libmca_mpool.la $(MCA_opal_mpool_STATIC_LTLIBS) mca/pmix/libmca_pmix.la $(MCA_opal_pmix_STATIC_LTLIBS) mca/pstat/libmca_pstat.la $(MCA_opal_pstat_STATIC_LTLIBS) mca/rcache/libmca_rcache.la $(MCA_opal_rcache_STATIC_LTLIBS) mca/sec/libmca_sec.la $(MCA_opal_sec_STATIC_LTLIBS) mca/shmem/libmca_shmem.la $(MCA_opal_shmem_STATIC_LTLIBS) mca/timer/libmca_timer.la $(MCA_opal_timer_STATIC_LTLIBS)'
...
MCA_opal_pmix_ALL_COMPONENTS=' s1 cray s2 pmix112 external'
MCA_opal_pmix_ALL_SUBDIRS=' mca/pmix/s1 mca/pmix/cray mca/pmix/s2 mca/pmix/pmix112 mca/pmix/external'
MCA_opal_pmix_DSO_COMPONENTS=' pmix112'
MCA_opal_pmix_DSO_SUBDIRS=' mca/pmix/pmix112'
MCA_opal_pmix_STATIC_COMPONENTS=''
MCA_opal_pmix_STATIC_LTLIBS=''
MCA_opal_pmix_STATIC_SUBDIRS=''
...
opal_pmix_ext_CPPFLAGS=''
opal_pmix_ext_LDFLAGS=''
opal_pmix_ext_LIBS=''
opal_pmix_pmix112_CPPFLAGS='-I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/include/pmix -I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/include -I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix -I$(OPAL_TOP_SRCDIR)/opal/mca/pmix/pmix112/pmix' opal_pmix_pmix112_LIBS='$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/libpmix.la'
...



"config.log" for Sun C 5.13 shows the following.

...
configure:127803: /bin/bash '../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/
pmix/configure' *failed* for opal/mca/pmix/pmix112/pmix
configure:128379: checking if MCA component pmix:pmix112 can compile
configure:128381: result: no
configure:5637: --- MCA component pmix:external (m4 configuration macro)
configure:128523: checking for MCA component pmix:external compile mode
configure:128529: result: dso
configure:129054: checking if MCA component pmix:external can compile
configure:129056: result: no
...
config.status:3887: creating opal/mca/pmix/Makefile
config.status:3887: creating opal/mca/pmix/s1/Makefile
config.status:3887: creating opal/mca/pmix/cray/Makefile
config.status:3887: creating opal/mca/pmix/s2/Makefile
config.status:3887: creating opal/mca/pmix/pmix112/Makefile
config.status:3887: creating opal/mca/pmix/external/Makefile
...
MCA_BUILD_opal_pmix_cray_DSO_FALSE='#'
MCA_BUILD_opal_pmix_cray_DSO_TRUE=''
MCA_BUILD_opal_pmix_external_DSO_FALSE='#'
MCA_BUILD_opal_pmix_external_DSO_TRUE=''
MCA_BUILD_opal_pmix_pmix112_DSO_FALSE='#'
MCA_BUILD_opal_pmix_pmix112_DSO_TRUE=''
MCA_BUILD_opal_pmix_s1_DSO_FALSE='#'
MCA_BUILD_opal_pmix_s1_DSO_TRUE=''
MCA_BUILD_opal_pmix_s2_DSO_FALSE='#'
MCA_BUILD_opal_pmix_s2_DSO_TRUE=''
...
MCA_opal_FRAMEWORKS='common allocator backtrace btl dl event hwloc if installdirs memchecker memcpy memory mpool pmix pstat rcache sec shmem timer' MCA_opal_FRAMEWORKS_SUBDIRS='mca/common mca/allocator mca/backtrace mca/btl mca/dl mca/event mca/hwloc mca/if mca/installdirs mca/memchecker mca/memcpy mca/memory mca/mpool mca/pmix mca/pstat mca/rcache mca/sec mca/shmem mca/timer' MCA_opal_FRAMEWORK_COMPONENT_ALL_SUBDIRS='$(MCA_opal_common_ALL_SUBDIRS) $(MCA_opal_allocator_ALL_SUBDIRS) $(MCA_opal_backtrace_ALL_SUBDIRS) $(MCA_opal_btl_ALL_SUBDIRS) $(MCA_opal_dl_ALL_SUBDIRS) $(MCA_opal_event_ALL_SUBDIRS) $(MCA_opal_hwloc_ALL_SUBDIRS) $(MCA_opal_if_ALL_SUBDIRS) $(MCA_opal_installdirs_ALL_SUBDIRS) $(MCA_opal_memchecker_ALL_SUBDIRS) $(MCA_opal_memcpy_ALL_SUBDIRS) $(MCA_opal_memory_ALL_SUBDIRS) $(MCA_opal_mpool_ALL_SUBDIRS) $(MCA_opal_pmix_ALL_SUBDIRS) $(MCA_opal_pstat_ALL_SUBDIRS) $(MCA_opal_rcache_ALL_SUBDIRS) $(MCA_opal_sec_ALL_SUBDIRS) $(MCA_opal_shmem_ALL_SUBDIRS) $(MCA_opal_timer_ALL_SUBDIRS)' MCA_opal_FRAMEWORK_COMPONENT_DSO_SUBDIRS='$(MCA_opal_common_DSO_SUBDIRS) $(MCA_opal_allocator_DSO_SUBDIRS) $(MCA_opal_backtrace_DSO_SUBDIRS) $(MCA_opal_btl_DSO_SUBDIRS) $(MCA_opal_dl_DSO_SUBDIRS) $(MCA_opal_event_DSO_SUBDIRS) $(MCA_opal_hwloc_DSO_SUBDIRS) $(MCA_opal_if_DSO_SUBDIRS) $(MCA_opal_installdirs_DSO_SUBDIRS) $(MCA_opal_memchecker_DSO_SUBDIRS) $(MCA_opal_memcpy_DSO_SUBDIRS) $(MCA_opal_memory_DSO_SUBDIRS) $(MCA_opal_mpool_DSO_SUBDIRS) $(MCA_opal_pmix_DSO_SUBDIRS) $(MCA_opal_pstat_DSO_SUBDIRS) $(MCA_opal_rcache_DSO_SUBDIRS) $(MCA_opal_sec_DSO_SUBDIRS) $(MCA_opal_shmem_DSO_SUBDIRS) $(MCA_opal_timer_DSO_SUBDIRS)' MCA_opal_FRAMEWORK_COMPONENT_STATIC_SUBDIRS='$(MCA_opal_common_STATIC_SUBDIRS) $(MCA_opal_allocator_STATIC_SUBDIRS) $(MCA_opal_backtrace_STATIC_SUBDIRS) $(MCA_opal_btl_STATIC_SUBDIRS) $(MCA_opal_dl_STATIC_SUBDIRS) $(MCA_opal_event_STATIC_SUBDIRS) $(MCA_opal_hwloc_STATIC_SUBDIRS) $(MCA_opal_if_STATIC_SUBDIRS) $(MCA_opal_installdirs_STATIC_SUBDIRS) $(MCA_opal_memchecker_STATIC_SUBDIRS) $(MCA_opal_memcpy_STATIC_SUBDIRS) $(MCA_opal_memory_STATIC_SUBDIRS) $(MCA_opal_mpool_STATIC_SUBDIRS) $(MCA_opal_pmix_STATIC_SUBDIRS) $(MCA_opal_pstat_STATIC_SUBDIRS) $(MCA_opal_rcache_STATIC_SUBDIRS) $(MCA_opal_sec_STATIC_SUBDIRS) $(MCA_opal_shmem_STATIC_SUBDIRS) $(MCA_opal_timer_STATIC_SUBDIRS)' MCA_opal_FRAMEWORK_LIBS=' $(MCA_opal_common_STATIC_LTLIBS) mca/allocator/libmca_allocator.la $(MCA_opal_allocator_STATIC_LTLIBS) mca/backtrace/libmca_backtrace.la $(MCA_opal_backtrace_STATIC_LTLIBS) mca/btl/libmca_btl.la $(MCA_opal_btl_STATIC_LTLIBS) mca/dl/libmca_dl.la $(MCA_opal_dl_STATIC_LTLIBS) mca/event/libmca_event.la $(MCA_opal_event_STATIC_LTLIBS) mca/hwloc/libmca_hwloc.la $(MCA_opal_hwloc_STATIC_LTLIBS) mca/if/libmca_if.la $(MCA_opal_if_STATIC_LTLIBS) mca/installdirs/libmca_installdirs.la $(MCA_opal_installdirs_STATIC_LTLIBS) mca/memchecker/libmca_memchecker.la $(MCA_opal_memchecker_STATIC_LTLIBS) mca/memcpy/libmca_memcpy.la $(MCA_opal_memcpy_STATIC_LTLIBS) mca/memory/libmca_memory.la $(MCA_opal_memory_STATIC_LTLIBS) mca/mpool/libmca_mpool.la $(MCA_opal_mpool_STATIC_LTLIBS) mca/pmix/libmca_pmix.la $(MCA_opal_pmix_STATIC_LTLIBS) mca/pstat/libmca_pstat.la $(MCA_opal_pstat_STATIC_LTLIBS) mca/rcache/libmca_rcache.la $(MCA_opal_rcache_STATIC_LTLIBS) mca/sec/libmca_sec.la $(MCA_opal_sec_STATIC_LTLIBS) mca/shmem/libmca_shmem.la $(MCA_opal_shmem_STATIC_LTLIBS) mca/timer/libmca_timer.la $(MCA_opal_timer_STATIC_LTLIBS)'
...
MCA_opal_pmix_ALL_COMPONENTS=' s1 cray s2 pmix112 external'
MCA_opal_pmix_ALL_SUBDIRS=' mca/pmix/s1 mca/pmix/cray mca/pmix/s2 mca/pmix/pmix112 mca/pmix/external'
MCA_opal_pmix_DSO_COMPONENTS=''
MCA_opal_pmix_DSO_SUBDIRS=''
MCA_opal_pmix_STATIC_COMPONENTS=''
MCA_opal_pmix_STATIC_LTLIBS=''
MCA_opal_pmix_STATIC_SUBDIRS=''
...
opal_pmix_ext_CPPFLAGS=''
opal_pmix_ext_LDFLAGS=''
opal_pmix_ext_LIBS=''
opal_pmix_pmix112_CPPFLAGS=''
opal_pmix_pmix112_LIBS=''
...




I've attached the config.log files for pmix.

tyr openmpi-2.0.0 142 tar zvft pmix_config.log.tar.gz
-rw-r--r-- root/root 136291 2016-04-25 08:05:34 openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/pmix/pmix112/pmix/config.log -rw-r--r-- root/root 528808 2016-04-25 08:07:54 openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_gcc/opal/mca/pmix/pmix112/pmix/config.log
tyr openmpi-2.0.0 143



I've also attached the output for the broken execution of
"spawn_multiple_master" for my gcc-version of Open MPI.
"spawn_master" works as expected with my gcc-version of Open MPI.

Hopefully you can fix the problem.


Kind regards and thank you very much for your help

Siegmar



Am 23.04.2016 um 21:34 schrieb Siegmar Gross:
Hi Gilles,

I don't know what happened, but the files are not available now
and they were definitely available when I answered the email from
Ralph. The files also have a different timestamp now. This is an
extract from my email to Ralph for Solaris Sparc.

-rwxr-xr-x 1 root root     977 Apr 19 19:49 mca_plm_rsh.la
-rwxr-xr-x 1 root root  153280 Apr 19 19:49 mca_plm_rsh.so
-rwxr-xr-x 1 root root    1007 Apr 19 19:47 mca_pmix_pmix112.la
-rwxr-xr-x 1 root root 1400512 Apr 19 19:47 mca_pmix_pmix112.so
-rwxr-xr-x 1 root root     971 Apr 19 19:52 mca_pml_cm.la
-rwxr-xr-x 1 root root  342440 Apr 19 19:52 mca_pml_cm.so

Now I have the following output for these files.

-rwxr-xr-x 1 root root     976 Apr 19 19:58 mca_plm_rsh.la
-rwxr-xr-x 1 root root  319816 Apr 19 19:58 mca_plm_rsh.so
-rwxr-xr-x 1 root root     970 Apr 19 20:00 mca_pml_cm.la
-rwxr-xr-x 1 root root 1507440 Apr 19 20:00 mca_pml_cm.so

I'll try to find out what happened next week when I'm back in
my office.


Kind regards

Siegmar





Am 23.04.16 um 02:12 schrieb Gilles Gouaillardet:
Siegmar,

I will try to reproduce this on my solaris11 x86_64 vm

In the mean time, can you please double check mca_pmix_pmix_pmix112.so
is a 64 bits library ?
(E.g, confirm "-m64" was correctly passed to pmix)

Cheers,

Gilles

On Friday, April 22, 2016, Siegmar Gross
<siegmar.gr...@informatik.hs-fulda.de
<mailto:siegmar.gr...@informatik.hs-fulda.de>> wrote:

    Hi Ralph,

    I've already used "-enable-debug". "SYSTEM_ENV" is "SunOS" or
    "Linux" and "MACHINE_ENV" is "sparc" or "x86_84".

mkdir openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_gcc cd openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_gcc

    ../openmpi-v2.x-dev-1280-gc110ae8/configure \
      --prefix=/usr/local/openmpi-2.0.0_64_gcc \
      --libdir=/usr/local/openmpi-2.0.0_64_gcc/lib64 \
      --with-jdk-bindir=/usr/local/jdk1.8.0/bin \
      --with-jdk-headers=/usr/local/jdk1.8.0/include \
      JAVA_HOME=/usr/local/jdk1.8.0 \
      LDFLAGS="-m64" CC="gcc" CXX="g++" FC="gfortran" \
      CFLAGS="-m64" CXXFLAGS="-m64" FCFLAGS="-m64" \
      CPP="cpp" CXXCPP="cpp" \
      --enable-mpi-cxx \
      --enable-cxx-exceptions \
      --enable-mpi-java \
      --enable-heterogeneous \
      --enable-mpi-thread-multiple \
      --with-hwloc=internal \
      --without-verbs \
      --with-wrapper-cflags="-std=c11 -m64" \
      --with-wrapper-cxxflags="-m64" \
      --with-wrapper-fcflags="-m64" \
      --enable-debug \
      |& tee log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_gcc


mkdir openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_cc cd openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_cc

    ../openmpi-v2.x-dev-1280-gc110ae8/configure \
      --prefix=/usr/local/openmpi-2.0.0_64_cc \
      --libdir=/usr/local/openmpi-2.0.0_64_cc/lib64 \
      --with-jdk-bindir=/usr/local/jdk1.8.0/bin \
      --with-jdk-headers=/usr/local/jdk1.8.0/include \
      JAVA_HOME=/usr/local/jdk1.8.0 \
      LDFLAGS="-m64" CC="cc" CXX="CC" FC="f95" \
      CFLAGS="-m64" CXXFLAGS="-m64 -library=stlport4" FCFLAGS="-m64" \
      CPP="cpp" CXXCPP="cpp" \
      --enable-mpi-cxx \
      --enable-cxx-exceptions \
      --enable-mpi-java \
      --enable-heterogeneous \
      --enable-mpi-thread-multiple \
      --with-hwloc=internal \
      --without-verbs \
      --with-wrapper-cflags="-m64" \
      --with-wrapper-cxxflags="-m64 -library=stlport4" \
      --with-wrapper-fcflags="-m64" \
      --with-wrapper-ldflags="" \
      --enable-debug \
      |& tee log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_cc


    Kind regards

    Siegmar

    Am 21.04.2016 um 18:18 schrieb Ralph Castain:

        Can you please rebuild OMPI with -enable-debug in the configure
        cmd? It will let us see more error output


            On Apr 21, 2016, at 8:52 AM, Siegmar Gross
            <siegmar.gr...@informatik.hs-fulda.de> wrote:

            Hi Ralph,

            I don't see any additional information.

            tyr hello_1 108 mpiexec -np 4 --host
            tyr,sunpc1,linpc1,ruester -mca
            mca_base_component_show_load_errors 1 hello_1_mpi
            [tyr.informatik.hs-fulda.de:06211
            <http://tyr.informatik.hs-fulda.de:06211>] [[48741,0],0]
            ORTE_ERROR_LOG: Not found in file

../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c
            at line 638

--------------------------------------------------------------------------
            It looks like orte_init failed for some reason; your
            parallel process is
            likely to abort.  There are many reasons that a parallel
            process can
fail during orte_init; some of which are due to configuration or
            environment problems.  This failure appears to be an
            internal failure;
            here's some additional information (which may only be
            relevant to an
            Open MPI developer):

             opal_pmix_base_select failed
             --> Returned value Not found (-13) instead of ORTE_SUCCESS

--------------------------------------------------------------------------


            tyr hello_1 109 mpiexec -np 4 --host
            tyr,sunpc1,linpc1,ruester -mca
mca_base_component_show_load_errors 1 -mca pmix_base_verbose
            10 -mca pmix_server_verbose 5 hello_1_mpi
            [tyr.informatik.hs-fulda.de:06212
            <http://tyr.informatik.hs-fulda.de:06212>] mca: base:
            components_register: registering framework pmix components
            [tyr.informatik.hs-fulda.de:06212
            <http://tyr.informatik.hs-fulda.de:06212>] mca: base:
            components_open: opening pmix components
            [tyr.informatik.hs-fulda.de:06212
            <http://tyr.informatik.hs-fulda.de:06212>] mca:base:select:
            Auto-selecting pmix components
            [tyr.informatik.hs-fulda.de:06212
<http://tyr.informatik.hs-fulda.de:06212>] mca:base:select:(
            pmix) No component selected!
            [tyr.informatik.hs-fulda.de:06212
            <http://tyr.informatik.hs-fulda.de:06212>] [[48738,0],0]
            ORTE_ERROR_LOG: Not found in file

../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c
            at line 638

--------------------------------------------------------------------------
            It looks like orte_init failed for some reason; your
            parallel process is
            likely to abort.  There are many reasons that a parallel
            process can
fail during orte_init; some of which are due to configuration or
            environment problems.  This failure appears to be an
            internal failure;
            here's some additional information (which may only be
            relevant to an
            Open MPI developer):

             opal_pmix_base_select failed
             --> Returned value Not found (-13) instead of ORTE_SUCCESS

--------------------------------------------------------------------------
            tyr hello_1 110


            Kind regards

            Siegmar


            Am 21.04.2016 um 17:24 schrieb Ralph Castain:

                Hmmm…it looks like you built the right components, but
                they are not being picked up. Can you run your mpiexec
                command again, adding “-mca
                mca_base_component_show_load_errors 1” to the cmd line?


                    On Apr 21, 2016, at 8:16 AM, Siegmar Gross
<siegmar.gr...@informatik.hs-fulda.de> wrote:

                    Hi Ralph,

                    I have attached ompi_info output for both compilers
                    from my
                    sparc machine and the listings for both compilers
                    from the
                    <prefix>/lib/openmpi directories. Hopefully that
                    helps to
                    find the problem.

                    hermes tmp 3 tar zvft openmpi-2.x_info.tar.gz
                    -rw-r--r-- root/root     10969 2016-04-21 17:06
                    ompi_info_SunOS_sparc_cc.txt
                    -rw-r--r-- root/root     11044 2016-04-21 17:06
                    ompi_info_SunOS_sparc_gcc.txt
                    -rw-r--r-- root/root     71252 2016-04-21 17:02
                    lib64_openmpi.txt
                    hermes tmp 4


                    Kind regards and thank you very much once more for
                    your help

                    Siegmar


                    Am 21.04.2016 um 15:54 schrieb Ralph Castain:

                        Odd - it would appear that none of the pmix
                        components built? Can you send
                        along the output from ompi_info? Or just send a
                        listing of the files in the
                        <prefix>/lib/openmpi directory?


                            On Apr 21, 2016, at 1:27 AM, Siegmar Gross
<siegmar.gr...@informatik.hs-fulda.de
<mailto:siegmar.gr...@informatik.hs-fulda.de>>
                            wrote:

                            Hi Ralph,

Am 21.04.2016 um 00:18 schrieb Ralph Castain:

                                Could you please rerun these test and
                                add “-mca pmix_base_verbose 10
                                -mca pmix_server_verbose 5” to your cmd
                                line? I need to see why the
                                pmix components failed.



                            tyr spawn 111 mpiexec -np 1 --host
                            tyr,sunpc1,linpc1,ruester -mca
                            pmix_base_verbose 10 -mca
                            pmix_server_verbose 5 spawn_multiple_master
                            [tyr.informatik.hs-fulda.de
<http://tyr.informatik.hs-fulda.de>
<http://tyr.informatik.hs-fulda.de/>:26652] mca:
                            base: components_register: registering
                            framework pmix components
                            [tyr.informatik.hs-fulda.de
<http://tyr.informatik.hs-fulda.de>
<http://tyr.informatik.hs-fulda.de/>:26652] mca:
base: components_open: opening pmix components
                            [tyr.informatik.hs-fulda.de
<http://tyr.informatik.hs-fulda.de>
<http://tyr.informatik.hs-fulda.de/>:26652]
mca:base:select: Auto-selecting pmix components
                            [tyr.informatik.hs-fulda.de
<http://tyr.informatik.hs-fulda.de>
<http://tyr.informatik.hs-fulda.de/>:26652]
mca:base:select:( pmix) No component selected!
                            [tyr.informatik.hs-fulda.de
<http://tyr.informatik.hs-fulda.de>
<http://tyr.informatik.hs-fulda.de/>:26652]
[[52794,0],0] ORTE_ERROR_LOG: Not found in file

../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c
                            at line 638

--------------------------------------------------------------------------
                            It looks like orte_init failed for some
                            reason; your parallel process is
                            likely to abort.  There are many reasons
                            that a parallel process can
fail during orte_init; some of which are due
                            to configuration or
                            environment problems.  This failure appears
                            to be an internal failure;
                            here's some additional information (which
                            may only be relevant to an
                            Open MPI developer):

                            opal_pmix_base_select failed
                            --> Returned value Not found (-13) instead
                            of ORTE_SUCCESS

--------------------------------------------------------------------------
                            tyr spawn 112




                            tyr hello_1 116 mpiexec -np 1 --host
                            tyr,sunpc1,linpc1,ruester -mca
                            pmix_base_verbose 10 -mca
                            pmix_server_verbose 5 hello_1_mpi
                            [tyr.informatik.hs-fulda.de
<http://tyr.informatik.hs-fulda.de>
<http://tyr.informatik.hs-fulda.de/>:27261] mca:
                            base: components_register: registering
                            framework pmix components
                            [tyr.informatik.hs-fulda.de
<http://tyr.informatik.hs-fulda.de>
<http://tyr.informatik.hs-fulda.de/>:27261] mca:
base: components_open: opening pmix components
                            [tyr.informatik.hs-fulda.de
<http://tyr.informatik.hs-fulda.de>
<http://tyr.informatik.hs-fulda.de/>:27261]
mca:base:select: Auto-selecting pmix components
                            [tyr.informatik.hs-fulda.de
<http://tyr.informatik.hs-fulda.de>
<http://tyr.informatik.hs-fulda.de/>:27261]
mca:base:select:( pmix) No component selected!
                            [tyr.informatik.hs-fulda.de
<http://tyr.informatik.hs-fulda.de>
<http://tyr.informatik.hs-fulda.de/>:27261]
[[52315,0],0] ORTE_ERROR_LOG: Not found in file

../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c
                            at line 638

--------------------------------------------------------------------------
                            It looks like orte_init failed for some
                            reason; your parallel process is
                            likely to abort.  There are many reasons
                            that a parallel process can
fail during orte_init; some of which are due
                            to configuration or
                            environment problems.  This failure appears
                            to be an internal failure;
                            here's some additional information (which
                            may only be relevant to an
                            Open MPI developer):

                            opal_pmix_base_select failed
                            --> Returned value Not found (-13) instead
                            of ORTE_SUCCESS

--------------------------------------------------------------------------
                            tyr hello_1 117



                            Thank you very much for your help.


                            Kind regards

                            Siegmar




                                Thanks
                                Ralph

                                    On Apr 20, 2016, at 10:12 AM,
                                    Siegmar Gross
<siegmar.gr...@informatik.hs-fulda.de

<mailto:siegmar.gr...@informatik.hs-fulda.de>>
                                    wrote:

                                    Hi,

                                    I have built
openmpi-v2.x-dev-1280-gc110ae8 on my
                                    machines
                                    (Solaris 10 Sparc, Solaris 10
                                    x86_64, and openSUSE Linux
                                    12.1 x86_64) with gcc-5.1.0 and Sun
                                    C 5.13. Unfortunately I get
                                    runtime errors for some programs.


                                    Sun C 5.13:
                                    ===========

                                    For all my test programs I get the
                                    same error on Solaris Sparc and
                                    Solaris x86_64, while the programs
                                    work fine on Linux.

                                    tyr hello_1 115 mpiexec -np 2
                                    hello_1_mpi
[tyr.informatik.hs-fulda.de
<http://tyr.informatik.hs-fulda.de>
<http://tyr.informatik.hs-fulda.de>:22373]
                                    [[61763,0],0] ORTE_ERROR_LOG: Not
                                    found in file

../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c
                                    at line 638

--------------------------------------------------------------------------
                                    It looks like orte_init failed for
some reason; your parallel process is
                                    likely to abort.  There are many
                                    reasons that a parallel process can
fail during orte_init; some of which
                                    are due to configuration or
                                    environment problems. This failure
                                    appears to be an internal failure;
                                    here's some additional information
                                    (which may only be relevant to an
                                    Open MPI developer):

                                    opal_pmix_base_select failed
                                    --> Returned value Not found (-13)
                                    instead of ORTE_SUCCESS

--------------------------------------------------------------------------
                                    tyr hello_1 116




                                    GCC-5.1.0:
                                    ==========

                                    tyr spawn 121 mpiexec -np 1 --host
                                    tyr,sunpc1,linpc1,ruester
                                    spawn_multiple_master

                                    Parent process 0 running on
                                    tyr.informatik.hs-fulda.de
<http://tyr.informatik.hs-fulda.de>
<http://tyr.informatik.hs-fulda.de>
                                    I create 3 slave processes.

[tyr.informatik.hs-fulda.de
<http://tyr.informatik.hs-fulda.de>
<http://tyr.informatik.hs-fulda.de>:25366]
                                    PMIX ERROR: UNPACK-PAST-END in file

../../../../../../openmpi-v2.x-dev-1280-gc110ae8/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c

                                    at line 829
[tyr.informatik.hs-fulda.de
<http://tyr.informatik.hs-fulda.de>
<http://tyr.informatik.hs-fulda.de>:25366]
                                    PMIX ERROR: UNPACK-PAST-END in file

../../../../../../openmpi-v2.x-dev-1280-gc110ae8/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c

                                    at line 2176
[tyr:25377] *** An error occurred in
                                    MPI_Comm_spawn_multiple
                                    [tyr:25377] *** reported by process
                                    [3308257281,0]
                                    [tyr:25377] *** on communicator
                                    MPI_COMM_WORLD
[tyr:25377] *** MPI_ERR_SPAWN: could
                                    not spawn processes
[tyr:25377] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will
                                    now abort,
                                    [tyr:25377] ***    and potentially
                                    your MPI job)
                                    tyr spawn 122


                                    I would be grateful if somebody can
                                    fix the problems. Thank you very
                                    much for any help in advance.


                                    Kind regards

                                    Siegmar

<hello_1_mpi.c><spawn_multiple_master.c>_______________________________________________

                                    users mailing list
                                    us...@open-mpi.org
<mailto:us...@open-mpi.org>
                                    Subscription:

http://www.open-mpi.org/mailman/listinfo.cgi/users
                                    Link to this post:

http://www.open-mpi.org/community/lists/users/2016/04/28983.php


_______________________________________________
                                users mailing list
us...@open-mpi.org <mailto:us...@open-mpi.org>
                                Subscription:

http://www.open-mpi.org/mailman/listinfo.cgi/users
                                Link to this
                                post:

http://www.open-mpi.org/community/lists/users/2016/04/28986.php


_______________________________________________
                            users mailing list
us...@open-mpi.org <mailto:us...@open-mpi.org>
                            Subscription:
http://www.open-mpi.org/mailman/listinfo.cgi/users
                            Link to this
                            post:

http://www.open-mpi.org/community/lists/users/2016/04/28987.php




_______________________________________________
                        users mailing list
                        us...@open-mpi.org
                        Subscription:
http://www.open-mpi.org/mailman/listinfo.cgi/users
                        Link to this post:

http://www.open-mpi.org/community/lists/users/2016/04/28988.php


<openmpi-2.x_info.tar.gz>_______________________________________________
                    users mailing list
                    us...@open-mpi.org
                    Subscription:
http://www.open-mpi.org/mailman/listinfo.cgi/users
                    Link to this post:

http://www.open-mpi.org/community/lists/users/2016/04/28989.php


_______________________________________________
                users mailing list
                us...@open-mpi.org
                Subscription:
http://www.open-mpi.org/mailman/listinfo.cgi/users
                Link to this post:
http://www.open-mpi.org/community/lists/users/2016/04/28990.php


            _______________________________________________
            users mailing list
            us...@open-mpi.org
Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
            Link to this post:
http://www.open-mpi.org/community/lists/users/2016/04/28991.php


        _______________________________________________
        users mailing list
        us...@open-mpi.org
Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
        Link to this post:
http://www.open-mpi.org/community/lists/users/2016/04/28992.php

    _______________________________________________
    users mailing list
    us...@open-mpi.org
    Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
    Link to this post:
http://www.open-mpi.org/community/lists/users/2016/04/28993.php



_______________________________________________
users mailing list
us...@open-mpi.org
Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
Link to this post:
http://www.open-mpi.org/community/lists/users/2016/04/28999.php

_______________________________________________
users mailing list
us...@open-mpi.org
Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
Link to this post:
http://www.open-mpi.org/community/lists/users/2016/04/29009.php


_______________________________________________
users mailing list
us...@open-mpi.org
Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
Link to this post: 
http://www.open-mpi.org/community/lists/users/2016/04/29033.php

Reply via email to