Siegmar,

please add this to your CFLAGS for the time being.

configure tries to detect which flags must be added for C99 support, and it
seems
the test is not working for Solaris 10 and Oracle compilers.
this is no more a widely used environment, and I am not sure I can find the
time to fix this
in a near future.


regarding the runtime issue, can you please describe your 4 hosts (os,
endianness and bitness)

Cheers,

Gilles

On Wednesday, April 27, 2016, Siegmar Gross <
siegmar.gr...@informatik.hs-fulda.de
<javascript:_e(%7B%7D,'cvml','siegmar.gr...@informatik.hs-fulda.de');>>
wrote:

> Hi Gilles,
>
> adding "-std=c99" to CFLAGS solves the problem with the missing library.
> Shall I add it permanently to my configure command or will you add it,
> so that I will not run into problems if you need the C11 standard later?
>
> "spawn_multiple_master" breaks with the same error that I reported
> yesterday for my gcc-version of Open MPI. Hopefully you can solve the
> problem as well.
>
>
> Kind regards and thank you very much for your help
>
> Siegmar
>
>
> Am 27.04.2016 um 08:05 schrieb Gilles Gouaillardet:
>
>> Siegmar,
>>
>>
>> here is the error :
>>
>> configure:17969: cc -o conftest -m64 -D_REENTRANT -g  -g
>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1
>>
>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc
>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/include
>>
>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/include
>> -D_REENTRANT
>>
>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/hwloc/hwloc1112/hwloc/include
>>
>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/hwloc/hwloc1112/hwloc/include
>>
>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/event/libevent2022/libevent
>>
>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/event/libevent2022/libevent/include
>>
>> -I/export2/src/openmpi-2.0.0/openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/event/libevent2022/libevent/include
>> -m64 conftest.c  >&5
>> "/usr/include/stdbool.h", line 42: #error: "Use of <stdbool.h> is valid
>> only
>> in a c99 compilation environment."
>>
>>
>> i cannot reproduce this on solaris 11 with oracle studio 5.3 compiler,
>> and i
>> do not have solaris 10 yet.
>>
>> could you please re-configure with '-std=c99' appended to your CFLAGS and
>> see
>> if it helps ?
>>
>>
>> Cheers,
>>
>>
>> Gilles
>>
>>
>> On 4/26/2016 7:57 PM, Siegmar Gross wrote:
>>
>>> Hi Gilles and Ralph,
>>>
>>> I was able to sort out my mess. In my last email I compared the
>>> files from "SunOS_sparc/openmpi-2.0.0_64_gcc/lib64/openmpi" from
>>> the attachment of my email to Ralph with the files from
>>> "SunOS_sparc/openmpi-2.0.0_64_cc/lib64/openmpi" from my current
>>> file system. That's the reason while I have had different
>>> timestamps. The other problem was that Ralph didn't recognize
>>> that "mca_pmix_pmix112.so" wasn't built on Solaris with the
>>> Sun C compiler. I've removed most of the files from the attachment
>>> of my email so that it is easier to see the relevant files. Below
>>> I try to give you more information that may be relevant to track
>>> down the problem. I still get an error running one of my small
>>> test programs, when I use my gcc-version of Open MPI.
>>> "mca_pmix_pmix112.so" is a 64 bits library.
>>>
>>> Linux_x86_64/openmpi-2.0.0_64_cc/lib64/openmpi:
>>> ...
>>> -rwxr-xr-x 1 root root  261327 Apr 19 16:46 mca_plm_slurm.so
>>> -rwxr-xr-x 1 root root    1002 Apr 19 16:45 mca_pmix_pmix112.la
>>> -rwxr-xr-x 1 root root 3906526 Apr 19 16:45 mca_pmix_pmix112.so
>>> -rwxr-xr-x 1 root root     966 Apr 19 16:51 mca_pml_cm.la
>>> -rwxr-xr-x 1 root root 1574265 Apr 19 16:51 mca_pml_cm.so
>>> ...
>>>
>>> Linux_x86_64/openmpi-2.0.0_64_gcc/lib64/openmpi:
>>> ...
>>> -rwxr-xr-x 1 root root   70371 Apr 19 16:43 mca_plm_slurm.so
>>> -rwxr-xr-x 1 root root    1008 Apr 19 16:42 mca_pmix_pmix112.la
>>> -rwxr-xr-x 1 root root 1029005 Apr 19 16:42 mca_pmix_pmix112.so
>>> -rwxr-xr-x 1 root root     972 Apr 19 16:46 mca_pml_cm.la
>>> -rwxr-xr-x 1 root root  284858 Apr 19 16:46 mca_pml_cm.so
>>> ...
>>>
>>> SunOS_sparc/openmpi-2.0.0_64_cc/lib64/openmpi:
>>> ...
>>> -rwxr-xr-x 1 root root  319816 Apr 19 19:58 mca_plm_rsh.so
>>> -rwxr-xr-x 1 root root     970 Apr 19 20:00 mca_pml_cm.la
>>> -rwxr-xr-x 1 root root 1507440 Apr 19 20:00 mca_pml_cm.so
>>> ...
>>>
>>> SunOS_sparc/openmpi-2.0.0_64_gcc/lib64/openmpi:
>>> ...
>>> -rwxr-xr-x 1 root root  153280 Apr 19 19:49 mca_plm_rsh.so
>>> -rwxr-xr-x 1 root root    1007 Apr 19 19:47 mca_pmix_pmix112.la
>>> -rwxr-xr-x 1 root root 1400512 Apr 19 19:47 mca_pmix_pmix112.so
>>> -rwxr-xr-x 1 root root     971 Apr 19 19:52 mca_pml_cm.la
>>> -rwxr-xr-x 1 root root  342440 Apr 19 19:52 mca_pml_cm.so
>>> ...
>>>
>>> SunOS_x86_64/openmpi-2.0.0_64_cc/lib64/openmpi:
>>> ...
>>> -rwxr-xr-x 1 root root  300096 Apr 19 17:18 mca_plm_rsh.so
>>> -rwxr-xr-x 1 root root     970 Apr 19 17:23 mca_pml_cm.la
>>> -rwxr-xr-x 1 root root 1458816 Apr 19 17:23 mca_pml_cm.so
>>> ...
>>>
>>> SunOS_x86_64/openmpi-2.0.0_64_gcc/lib64/openmpi:
>>> ...
>>> -rwxr-xr-x 1 root root  133096 Apr 19 17:42 mca_plm_rsh.so
>>> -rwxr-xr-x 1 root root    1007 Apr 19 17:41 mca_pmix_pmix112.la
>>> -rwxr-xr-x 1 root root 1320240 Apr 19 17:41 mca_pmix_pmix112.so
>>> -rwxr-xr-x 1 root root     971 Apr 19 17:46 mca_pml_cm.la
>>> -rwxr-xr-x 1 root root  419848 Apr 19 17:46 mca_pml_cm.so
>>> ...
>>>
>>>
>>> Yesterday I've installed openmpi-v2.x-dev-1290-gbd0e4e1 so that we
>>> have a current version for the investigation of the problem. Once
>>> more mca_pmix_pmix112.so isn't available on Solaris if I use the
>>> Sun C compiler.
>>>
>>> "config.log" for gcc-5.1.0 shows the following.
>>>
>>> ...
>>> configure:127799: /bin/bash
>>> '../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/
>>> pmix/configure' succeeded for opal/mca/pmix/pmix112/pmix
>>> configure:127916: checking if MCA component pmix:pmix112 can compile
>>> configure:127918: result: yes
>>> configure:5637: --- MCA component pmix:external (m4 configuration macro)
>>> configure:128523: checking for MCA component pmix:external compile mode
>>> configure:128529: result: dso
>>> configure:129054: checking if MCA component pmix:external can compile
>>> configure:129056: result: no
>>> ...
>>> config.status:3897: creating opal/mca/pmix/Makefile
>>> config.status:3897: creating opal/mca/pmix/s1/Makefile
>>> config.status:3897: creating opal/mca/pmix/cray/Makefile
>>> config.status:3897: creating opal/mca/pmix/s2/Makefile
>>> config.status:3897: creating opal/mca/pmix/pmix112/Makefile
>>> config.status:3897: creating opal/mca/pmix/external/Makefile
>>> ...
>>> MCA_BUILD_opal_pmix_cray_DSO_FALSE='#'
>>> MCA_BUILD_opal_pmix_cray_DSO_TRUE=''
>>> MCA_BUILD_opal_pmix_external_DSO_FALSE='#'
>>> MCA_BUILD_opal_pmix_external_DSO_TRUE=''
>>> MCA_BUILD_opal_pmix_pmix112_DSO_FALSE='#'
>>> MCA_BUILD_opal_pmix_pmix112_DSO_TRUE=''
>>> MCA_BUILD_opal_pmix_s1_DSO_FALSE='#'
>>> MCA_BUILD_opal_pmix_s1_DSO_TRUE=''
>>> MCA_BUILD_opal_pmix_s2_DSO_FALSE='#'
>>> MCA_BUILD_opal_pmix_s2_DSO_TRUE=''
>>> ...
>>> MCA_opal_FRAMEWORKS='common  allocator backtrace btl dl event hwloc if
>>> installdirs memchecker memcpy memory mpool pmix pstat rcache sec shmem
>>> timer'
>>> MCA_opal_FRAMEWORKS_SUBDIRS='mca/common  mca/allocator mca/backtrace
>>> mca/btl
>>> mca/dl mca/event mca/hwloc mca/if mca/installdirs mca/memchecker
>>> mca/memcpy
>>> mca/memory mca/mpool mca/pmix mca/pstat mca/rcache mca/sec mca/shmem
>>> mca/timer'
>>> MCA_opal_FRAMEWORK_COMPONENT_ALL_SUBDIRS='$(MCA_opal_common_ALL_SUBDIRS)
>>> $(MCA_opal_allocator_ALL_SUBDIRS) $(MCA_opal_backtrace_ALL_SUBDIRS)
>>> $(MCA_opal_btl_ALL_SUBDIRS) $(MCA_opal_dl_ALL_SUBDIRS)
>>> $(MCA_opal_event_ALL_SUBDIRS) $(MCA_opal_hwloc_ALL_SUBDIRS)
>>> $(MCA_opal_if_ALL_SUBDIRS) $(MCA_opal_installdirs_ALL_SUBDIRS)
>>> $(MCA_opal_memchecker_ALL_SUBDIRS) $(MCA_opal_memcpy_ALL_SUBDIRS)
>>> $(MCA_opal_memory_ALL_SUBDIRS) $(MCA_opal_mpool_ALL_SUBDIRS)
>>> $(MCA_opal_pmix_ALL_SUBDIRS) $(MCA_opal_pstat_ALL_SUBDIRS)
>>> $(MCA_opal_rcache_ALL_SUBDIRS) $(MCA_opal_sec_ALL_SUBDIRS)
>>> $(MCA_opal_shmem_ALL_SUBDIRS) $(MCA_opal_timer_ALL_SUBDIRS)'
>>> MCA_opal_FRAMEWORK_COMPONENT_DSO_SUBDIRS='$(MCA_opal_common_DSO_SUBDIRS)
>>> $(MCA_opal_allocator_DSO_SUBDIRS) $(MCA_opal_backtrace_DSO_SUBDIRS)
>>> $(MCA_opal_btl_DSO_SUBDIRS) $(MCA_opal_dl_DSO_SUBDIRS)
>>> $(MCA_opal_event_DSO_SUBDIRS) $(MCA_opal_hwloc_DSO_SUBDIRS)
>>> $(MCA_opal_if_DSO_SUBDIRS) $(MCA_opal_installdirs_DSO_SUBDIRS)
>>> $(MCA_opal_memchecker_DSO_SUBDIRS) $(MCA_opal_memcpy_DSO_SUBDIRS)
>>> $(MCA_opal_memory_DSO_SUBDIRS) $(MCA_opal_mpool_DSO_SUBDIRS)
>>> $(MCA_opal_pmix_DSO_SUBDIRS) $(MCA_opal_pstat_DSO_SUBDIRS)
>>> $(MCA_opal_rcache_DSO_SUBDIRS) $(MCA_opal_sec_DSO_SUBDIRS)
>>> $(MCA_opal_shmem_DSO_SUBDIRS) $(MCA_opal_timer_DSO_SUBDIRS)'
>>>
>>> MCA_opal_FRAMEWORK_COMPONENT_STATIC_SUBDIRS='$(MCA_opal_common_STATIC_SUBDIRS)
>>>  $(MCA_opal_allocator_STATIC_SUBDIRS)
>>> $(MCA_opal_backtrace_STATIC_SUBDIRS)
>>> $(MCA_opal_btl_STATIC_SUBDIRS) $(MCA_opal_dl_STATIC_SUBDIRS)
>>> $(MCA_opal_event_STATIC_SUBDIRS) $(MCA_opal_hwloc_STATIC_SUBDIRS)
>>> $(MCA_opal_if_STATIC_SUBDIRS) $(MCA_opal_installdirs_STATIC_SUBDIRS)
>>> $(MCA_opal_memchecker_STATIC_SUBDIRS) $(MCA_opal_memcpy_STATIC_SUBDIRS)
>>> $(MCA_opal_memory_STATIC_SUBDIRS) $(MCA_opal_mpool_STATIC_SUBDIRS)
>>> $(MCA_opal_pmix_STATIC_SUBDIRS) $(MCA_opal_pstat_STATIC_SUBDIRS)
>>> $(MCA_opal_rcache_STATIC_SUBDIRS) $(MCA_opal_sec_STATIC_SUBDIRS)
>>> $(MCA_opal_shmem_STATIC_SUBDIRS) $(MCA_opal_timer_STATIC_SUBDIRS)'
>>> MCA_opal_FRAMEWORK_LIBS=' $(MCA_opal_common_STATIC_LTLIBS)
>>> mca/allocator/libmca_allocator.la $(MCA_opal_allocator_STATIC_LTLIBS)
>>> mca/backtrace/libmca_backtrace.la $(MCA_opal_backtrace_STATIC_LTLIBS)
>>> mca/btl/libmca_btl.la $(MCA_opal_btl_STATIC_LTLIBS) mca/dl/libmca_dl.la
>>> $(MCA_opal_dl_STATIC_LTLIBS) mca/event/libmca_event.la
>>> $(MCA_opal_event_STATIC_LTLIBS) mca/hwloc/libmca_hwloc.la
>>> $(MCA_opal_hwloc_STATIC_LTLIBS) mca/if/libmca_if.la
>>> $(MCA_opal_if_STATIC_LTLIBS) mca/installdirs/libmca_installdirs.la
>>> $(MCA_opal_installdirs_STATIC_LTLIBS) mca/memchecker/
>>> libmca_memchecker.la
>>> $(MCA_opal_memchecker_STATIC_LTLIBS) mca/memcpy/libmca_memcpy.la
>>> $(MCA_opal_memcpy_STATIC_LTLIBS) mca/memory/libmca_memory.la
>>> $(MCA_opal_memory_STATIC_LTLIBS) mca/mpool/libmca_mpool.la
>>> $(MCA_opal_mpool_STATIC_LTLIBS) mca/pmix/libmca_pmix.la
>>> $(MCA_opal_pmix_STATIC_LTLIBS) mca/pstat/libmca_pstat.la
>>> $(MCA_opal_pstat_STATIC_LTLIBS) mca/rcache/libmca_rcache.la
>>> $(MCA_opal_rcache_STATIC_LTLIBS) mca/sec/libmca_sec.la
>>> $(MCA_opal_sec_STATIC_LTLIBS) mca/shmem/libmca_shmem.la
>>> $(MCA_opal_shmem_STATIC_LTLIBS) mca/timer/libmca_timer.la
>>> $(MCA_opal_timer_STATIC_LTLIBS)'
>>> ...
>>> MCA_opal_pmix_ALL_COMPONENTS=' s1 cray s2 pmix112 external'
>>> MCA_opal_pmix_ALL_SUBDIRS=' mca/pmix/s1 mca/pmix/cray mca/pmix/s2
>>> mca/pmix/pmix112 mca/pmix/external'
>>> MCA_opal_pmix_DSO_COMPONENTS=' pmix112'
>>> MCA_opal_pmix_DSO_SUBDIRS=' mca/pmix/pmix112'
>>> MCA_opal_pmix_STATIC_COMPONENTS=''
>>> MCA_opal_pmix_STATIC_LTLIBS=''
>>> MCA_opal_pmix_STATIC_SUBDIRS=''
>>> ...
>>> opal_pmix_ext_CPPFLAGS=''
>>> opal_pmix_ext_LDFLAGS=''
>>> opal_pmix_ext_LIBS=''
>>>
>>> opal_pmix_pmix112_CPPFLAGS='-I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/include/pmix
>>> -I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/include
>>> -I$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix
>>> -I$(OPAL_TOP_SRCDIR)/opal/mca/pmix/pmix112/pmix'
>>> opal_pmix_pmix112_LIBS='$(OPAL_TOP_BUILDDIR)/opal/mca/pmix/pmix112/pmix/
>>> libpmix.la'
>>>
>>> ...
>>>
>>>
>>>
>>> "config.log" for Sun C 5.13 shows the following.
>>>
>>> ...
>>> configure:127803: /bin/bash
>>> '../../../../../../openmpi-v2.x-dev-1290-gbd0e4e1/opal/mca/pmix/pmix112/
>>> pmix/configure' *failed* for opal/mca/pmix/pmix112/pmix
>>> configure:128379: checking if MCA component pmix:pmix112 can compile
>>> configure:128381: result: no
>>> configure:5637: --- MCA component pmix:external (m4 configuration macro)
>>> configure:128523: checking for MCA component pmix:external compile mode
>>> configure:128529: result: dso
>>> configure:129054: checking if MCA component pmix:external can compile
>>> configure:129056: result: no
>>> ...
>>> config.status:3887: creating opal/mca/pmix/Makefile
>>> config.status:3887: creating opal/mca/pmix/s1/Makefile
>>> config.status:3887: creating opal/mca/pmix/cray/Makefile
>>> config.status:3887: creating opal/mca/pmix/s2/Makefile
>>> config.status:3887: creating opal/mca/pmix/pmix112/Makefile
>>> config.status:3887: creating opal/mca/pmix/external/Makefile
>>> ...
>>> MCA_BUILD_opal_pmix_cray_DSO_FALSE='#'
>>> MCA_BUILD_opal_pmix_cray_DSO_TRUE=''
>>> MCA_BUILD_opal_pmix_external_DSO_FALSE='#'
>>> MCA_BUILD_opal_pmix_external_DSO_TRUE=''
>>> MCA_BUILD_opal_pmix_pmix112_DSO_FALSE='#'
>>> MCA_BUILD_opal_pmix_pmix112_DSO_TRUE=''
>>> MCA_BUILD_opal_pmix_s1_DSO_FALSE='#'
>>> MCA_BUILD_opal_pmix_s1_DSO_TRUE=''
>>> MCA_BUILD_opal_pmix_s2_DSO_FALSE='#'
>>> MCA_BUILD_opal_pmix_s2_DSO_TRUE=''
>>> ...
>>> MCA_opal_FRAMEWORKS='common  allocator backtrace btl dl event hwloc if
>>> installdirs memchecker memcpy memory mpool pmix pstat rcache sec shmem
>>> timer'
>>> MCA_opal_FRAMEWORKS_SUBDIRS='mca/common  mca/allocator mca/backtrace
>>> mca/btl
>>> mca/dl mca/event mca/hwloc mca/if mca/installdirs mca/memchecker
>>> mca/memcpy
>>> mca/memory mca/mpool mca/pmix mca/pstat mca/rcache mca/sec mca/shmem
>>> mca/timer'
>>> MCA_opal_FRAMEWORK_COMPONENT_ALL_SUBDIRS='$(MCA_opal_common_ALL_SUBDIRS)
>>> $(MCA_opal_allocator_ALL_SUBDIRS) $(MCA_opal_backtrace_ALL_SUBDIRS)
>>> $(MCA_opal_btl_ALL_SUBDIRS) $(MCA_opal_dl_ALL_SUBDIRS)
>>> $(MCA_opal_event_ALL_SUBDIRS) $(MCA_opal_hwloc_ALL_SUBDIRS)
>>> $(MCA_opal_if_ALL_SUBDIRS) $(MCA_opal_installdirs_ALL_SUBDIRS)
>>> $(MCA_opal_memchecker_ALL_SUBDIRS) $(MCA_opal_memcpy_ALL_SUBDIRS)
>>> $(MCA_opal_memory_ALL_SUBDIRS) $(MCA_opal_mpool_ALL_SUBDIRS)
>>> $(MCA_opal_pmix_ALL_SUBDIRS) $(MCA_opal_pstat_ALL_SUBDIRS)
>>> $(MCA_opal_rcache_ALL_SUBDIRS) $(MCA_opal_sec_ALL_SUBDIRS)
>>> $(MCA_opal_shmem_ALL_SUBDIRS) $(MCA_opal_timer_ALL_SUBDIRS)'
>>> MCA_opal_FRAMEWORK_COMPONENT_DSO_SUBDIRS='$(MCA_opal_common_DSO_SUBDIRS)
>>> $(MCA_opal_allocator_DSO_SUBDIRS) $(MCA_opal_backtrace_DSO_SUBDIRS)
>>> $(MCA_opal_btl_DSO_SUBDIRS) $(MCA_opal_dl_DSO_SUBDIRS)
>>> $(MCA_opal_event_DSO_SUBDIRS) $(MCA_opal_hwloc_DSO_SUBDIRS)
>>> $(MCA_opal_if_DSO_SUBDIRS) $(MCA_opal_installdirs_DSO_SUBDIRS)
>>> $(MCA_opal_memchecker_DSO_SUBDIRS) $(MCA_opal_memcpy_DSO_SUBDIRS)
>>> $(MCA_opal_memory_DSO_SUBDIRS) $(MCA_opal_mpool_DSO_SUBDIRS)
>>> $(MCA_opal_pmix_DSO_SUBDIRS) $(MCA_opal_pstat_DSO_SUBDIRS)
>>> $(MCA_opal_rcache_DSO_SUBDIRS) $(MCA_opal_sec_DSO_SUBDIRS)
>>> $(MCA_opal_shmem_DSO_SUBDIRS) $(MCA_opal_timer_DSO_SUBDIRS)'
>>>
>>> MCA_opal_FRAMEWORK_COMPONENT_STATIC_SUBDIRS='$(MCA_opal_common_STATIC_SUBDIRS)
>>>  $(MCA_opal_allocator_STATIC_SUBDIRS)
>>> $(MCA_opal_backtrace_STATIC_SUBDIRS)
>>> $(MCA_opal_btl_STATIC_SUBDIRS) $(MCA_opal_dl_STATIC_SUBDIRS)
>>> $(MCA_opal_event_STATIC_SUBDIRS) $(MCA_opal_hwloc_STATIC_SUBDIRS)
>>> $(MCA_opal_if_STATIC_SUBDIRS) $(MCA_opal_installdirs_STATIC_SUBDIRS)
>>> $(MCA_opal_memchecker_STATIC_SUBDIRS) $(MCA_opal_memcpy_STATIC_SUBDIRS)
>>> $(MCA_opal_memory_STATIC_SUBDIRS) $(MCA_opal_mpool_STATIC_SUBDIRS)
>>> $(MCA_opal_pmix_STATIC_SUBDIRS) $(MCA_opal_pstat_STATIC_SUBDIRS)
>>> $(MCA_opal_rcache_STATIC_SUBDIRS) $(MCA_opal_sec_STATIC_SUBDIRS)
>>> $(MCA_opal_shmem_STATIC_SUBDIRS) $(MCA_opal_timer_STATIC_SUBDIRS)'
>>> MCA_opal_FRAMEWORK_LIBS=' $(MCA_opal_common_STATIC_LTLIBS)
>>> mca/allocator/libmca_allocator.la $(MCA_opal_allocator_STATIC_LTLIBS)
>>> mca/backtrace/libmca_backtrace.la $(MCA_opal_backtrace_STATIC_LTLIBS)
>>> mca/btl/libmca_btl.la $(MCA_opal_btl_STATIC_LTLIBS) mca/dl/libmca_dl.la
>>> $(MCA_opal_dl_STATIC_LTLIBS) mca/event/libmca_event.la
>>> $(MCA_opal_event_STATIC_LTLIBS) mca/hwloc/libmca_hwloc.la
>>> $(MCA_opal_hwloc_STATIC_LTLIBS) mca/if/libmca_if.la
>>> $(MCA_opal_if_STATIC_LTLIBS) mca/installdirs/libmca_installdirs.la
>>> $(MCA_opal_installdirs_STATIC_LTLIBS) mca/memchecker/
>>> libmca_memchecker.la
>>> $(MCA_opal_memchecker_STATIC_LTLIBS) mca/memcpy/libmca_memcpy.la
>>> $(MCA_opal_memcpy_STATIC_LTLIBS) mca/memory/libmca_memory.la
>>> $(MCA_opal_memory_STATIC_LTLIBS) mca/mpool/libmca_mpool.la
>>> $(MCA_opal_mpool_STATIC_LTLIBS) mca/pmix/libmca_pmix.la
>>> $(MCA_opal_pmix_STATIC_LTLIBS) mca/pstat/libmca_pstat.la
>>> $(MCA_opal_pstat_STATIC_LTLIBS) mca/rcache/libmca_rcache.la
>>> $(MCA_opal_rcache_STATIC_LTLIBS) mca/sec/libmca_sec.la
>>> $(MCA_opal_sec_STATIC_LTLIBS) mca/shmem/libmca_shmem.la
>>> $(MCA_opal_shmem_STATIC_LTLIBS) mca/timer/libmca_timer.la
>>> $(MCA_opal_timer_STATIC_LTLIBS)'
>>> ...
>>> MCA_opal_pmix_ALL_COMPONENTS=' s1 cray s2 pmix112 external'
>>> MCA_opal_pmix_ALL_SUBDIRS=' mca/pmix/s1 mca/pmix/cray mca/pmix/s2
>>> mca/pmix/pmix112 mca/pmix/external'
>>> MCA_opal_pmix_DSO_COMPONENTS=''
>>> MCA_opal_pmix_DSO_SUBDIRS=''
>>> MCA_opal_pmix_STATIC_COMPONENTS=''
>>> MCA_opal_pmix_STATIC_LTLIBS=''
>>> MCA_opal_pmix_STATIC_SUBDIRS=''
>>> ...
>>> opal_pmix_ext_CPPFLAGS=''
>>> opal_pmix_ext_LDFLAGS=''
>>> opal_pmix_ext_LIBS=''
>>> opal_pmix_pmix112_CPPFLAGS=''
>>> opal_pmix_pmix112_LIBS=''
>>> ...
>>>
>>>
>>>
>>>
>>> I've attached the config.log files for pmix.
>>>
>>> tyr openmpi-2.0.0 142 tar zvft pmix_config.log.tar.gz
>>> -rw-r--r-- root/root    136291 2016-04-25 08:05:34
>>>
>>> openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_cc/opal/mca/pmix/pmix112/pmix/config.log
>>> -rw-r--r-- root/root    528808 2016-04-25 08:07:54
>>>
>>> openmpi-v2.x-dev-1290-gbd0e4e1-SunOS.sparc.64_gcc/opal/mca/pmix/pmix112/pmix/config.log
>>> tyr openmpi-2.0.0 143
>>>
>>>
>>>
>>> I've also attached the output for the broken execution of
>>> "spawn_multiple_master" for my gcc-version of Open MPI.
>>> "spawn_master" works as expected with my gcc-version of Open MPI.
>>>
>>> Hopefully you can fix the problem.
>>>
>>>
>>> Kind regards and thank you very much for your help
>>>
>>> Siegmar
>>>
>>>
>>>
>>> Am 23.04.2016 um 21:34 schrieb Siegmar Gross:
>>>
>>>> Hi Gilles,
>>>>
>>>> I don't know what happened, but the files are not available now
>>>> and they were definitely available when I answered the email from
>>>> Ralph. The files also have a different timestamp now. This is an
>>>> extract from my email to Ralph for Solaris Sparc.
>>>>
>>>> -rwxr-xr-x 1 root root     977 Apr 19 19:49 mca_plm_rsh.la
>>>> -rwxr-xr-x 1 root root  153280 Apr 19 19:49 mca_plm_rsh.so
>>>> -rwxr-xr-x 1 root root    1007 Apr 19 19:47 mca_pmix_pmix112.la
>>>> -rwxr-xr-x 1 root root 1400512 Apr 19 19:47 mca_pmix_pmix112.so
>>>> -rwxr-xr-x 1 root root     971 Apr 19 19:52 mca_pml_cm.la
>>>> -rwxr-xr-x 1 root root  342440 Apr 19 19:52 mca_pml_cm.so
>>>>
>>>> Now I have the following output for these files.
>>>>
>>>> -rwxr-xr-x 1 root root     976 Apr 19 19:58 mca_plm_rsh.la
>>>> -rwxr-xr-x 1 root root  319816 Apr 19 19:58 mca_plm_rsh.so
>>>> -rwxr-xr-x 1 root root     970 Apr 19 20:00 mca_pml_cm.la
>>>> -rwxr-xr-x 1 root root 1507440 Apr 19 20:00 mca_pml_cm.so
>>>>
>>>> I'll try to find out what happened next week when I'm back in
>>>> my office.
>>>>
>>>>
>>>> Kind regards
>>>>
>>>> Siegmar
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Am 23.04.16 um 02:12 schrieb Gilles Gouaillardet:
>>>>
>>>>> Siegmar,
>>>>>
>>>>> I will try to reproduce this on my solaris11 x86_64 vm
>>>>>
>>>>> In the mean time, can you please double check mca_pmix_pmix_pmix112.so
>>>>> is a 64 bits library ?
>>>>> (E.g, confirm "-m64" was correctly passed to pmix)
>>>>>
>>>>> Cheers,
>>>>>
>>>>> Gilles
>>>>>
>>>>> On Friday, April 22, 2016, Siegmar Gross
>>>>> <siegmar.gr...@informatik.hs-fulda.de
>>>>> <mailto:siegmar.gr...@informatik.hs-fulda.de>> wrote:
>>>>>
>>>>>     Hi Ralph,
>>>>>
>>>>>     I've already used "-enable-debug". "SYSTEM_ENV" is "SunOS" or
>>>>>     "Linux" and "MACHINE_ENV" is "sparc" or "x86_84".
>>>>>
>>>>>     mkdir
>>>>> openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_gcc
>>>>>     cd
>>>>> openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_gcc
>>>>>
>>>>>     ../openmpi-v2.x-dev-1280-gc110ae8/configure \
>>>>>       --prefix=/usr/local/openmpi-2.0.0_64_gcc \
>>>>>       --libdir=/usr/local/openmpi-2.0.0_64_gcc/lib64 \
>>>>>       --with-jdk-bindir=/usr/local/jdk1.8.0/bin \
>>>>>       --with-jdk-headers=/usr/local/jdk1.8.0/include \
>>>>>       JAVA_HOME=/usr/local/jdk1.8.0 \
>>>>>       LDFLAGS="-m64" CC="gcc" CXX="g++" FC="gfortran" \
>>>>>       CFLAGS="-m64" CXXFLAGS="-m64" FCFLAGS="-m64" \
>>>>>       CPP="cpp" CXXCPP="cpp" \
>>>>>       --enable-mpi-cxx \
>>>>>       --enable-cxx-exceptions \
>>>>>       --enable-mpi-java \
>>>>>       --enable-heterogeneous \
>>>>>       --enable-mpi-thread-multiple \
>>>>>       --with-hwloc=internal \
>>>>>       --without-verbs \
>>>>>       --with-wrapper-cflags="-std=c11 -m64" \
>>>>>       --with-wrapper-cxxflags="-m64" \
>>>>>       --with-wrapper-fcflags="-m64" \
>>>>>       --enable-debug \
>>>>>       |& tee log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_gcc
>>>>>
>>>>>
>>>>>     mkdir
>>>>> openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_cc
>>>>>     cd
>>>>> openmpi-v2.x-dev-1280-gc110ae8-${SYSTEM_ENV}.${MACHINE_ENV}.64_cc
>>>>>
>>>>>     ../openmpi-v2.x-dev-1280-gc110ae8/configure \
>>>>>       --prefix=/usr/local/openmpi-2.0.0_64_cc \
>>>>>       --libdir=/usr/local/openmpi-2.0.0_64_cc/lib64 \
>>>>>       --with-jdk-bindir=/usr/local/jdk1.8.0/bin \
>>>>>       --with-jdk-headers=/usr/local/jdk1.8.0/include \
>>>>>       JAVA_HOME=/usr/local/jdk1.8.0 \
>>>>>       LDFLAGS="-m64" CC="cc" CXX="CC" FC="f95" \
>>>>>       CFLAGS="-m64" CXXFLAGS="-m64 -library=stlport4" FCFLAGS="-m64" \
>>>>>       CPP="cpp" CXXCPP="cpp" \
>>>>>       --enable-mpi-cxx \
>>>>>       --enable-cxx-exceptions \
>>>>>       --enable-mpi-java \
>>>>>       --enable-heterogeneous \
>>>>>       --enable-mpi-thread-multiple \
>>>>>       --with-hwloc=internal \
>>>>>       --without-verbs \
>>>>>       --with-wrapper-cflags="-m64" \
>>>>>       --with-wrapper-cxxflags="-m64 -library=stlport4" \
>>>>>       --with-wrapper-fcflags="-m64" \
>>>>>       --with-wrapper-ldflags="" \
>>>>>       --enable-debug \
>>>>>       |& tee log.configure.$SYSTEM_ENV.$MACHINE_ENV.64_cc
>>>>>
>>>>>
>>>>>     Kind regards
>>>>>
>>>>>     Siegmar
>>>>>
>>>>>     Am 21.04.2016 um 18:18 schrieb Ralph Castain:
>>>>>
>>>>>         Can you please rebuild OMPI with -enable-debug in the configure
>>>>>         cmd? It will let us see more error output
>>>>>
>>>>>
>>>>>             On Apr 21, 2016, at 8:52 AM, Siegmar Gross
>>>>>             <siegmar.gr...@informatik.hs-fulda.de> wrote:
>>>>>
>>>>>             Hi Ralph,
>>>>>
>>>>>             I don't see any additional information.
>>>>>
>>>>>             tyr hello_1 108 mpiexec -np 4 --host
>>>>>             tyr,sunpc1,linpc1,ruester -mca
>>>>>             mca_base_component_show_load_errors 1 hello_1_mpi
>>>>>             [tyr.informatik.hs-fulda.de:06211
>>>>>             <http://tyr.informatik.hs-fulda.de:06211>] [[48741,0],0]
>>>>>             ORTE_ERROR_LOG: Not found in file
>>>>>
>>>>>
>>>>> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c
>>>>>
>>>>>             at line 638
>>>>>
>>>>>
>>>>> --------------------------------------------------------------------------
>>>>>             It looks like orte_init failed for some reason; your
>>>>>             parallel process is
>>>>>             likely to abort.  There are many reasons that a parallel
>>>>>             process can
>>>>>             fail during orte_init; some of which are due to
>>>>> configuration or
>>>>>             environment problems.  This failure appears to be an
>>>>>             internal failure;
>>>>>             here's some additional information (which may only be
>>>>>             relevant to an
>>>>>             Open MPI developer):
>>>>>
>>>>>              opal_pmix_base_select failed
>>>>>              --> Returned value Not found (-13) instead of ORTE_SUCCESS
>>>>>
>>>>>
>>>>> --------------------------------------------------------------------------
>>>>>
>>>>>
>>>>>             tyr hello_1 109 mpiexec -np 4 --host
>>>>>             tyr,sunpc1,linpc1,ruester -mca
>>>>>             mca_base_component_show_load_errors 1 -mca
>>>>> pmix_base_verbose
>>>>>             10 -mca pmix_server_verbose 5 hello_1_mpi
>>>>>             [tyr.informatik.hs-fulda.de:06212
>>>>>             <http://tyr.informatik.hs-fulda.de:06212>] mca: base:
>>>>>             components_register: registering framework pmix components
>>>>>             [tyr.informatik.hs-fulda.de:06212
>>>>>             <http://tyr.informatik.hs-fulda.de:06212>] mca: base:
>>>>>             components_open: opening pmix components
>>>>>             [tyr.informatik.hs-fulda.de:06212
>>>>>             <http://tyr.informatik.hs-fulda.de:06212>]
>>>>> mca:base:select:
>>>>>             Auto-selecting pmix components
>>>>>             [tyr.informatik.hs-fulda.de:06212
>>>>>             <http://tyr.informatik.hs-fulda.de:06212>]
>>>>> mca:base:select:(
>>>>>             pmix) No component selected!
>>>>>             [tyr.informatik.hs-fulda.de:06212
>>>>>             <http://tyr.informatik.hs-fulda.de:06212>] [[48738,0],0]
>>>>>             ORTE_ERROR_LOG: Not found in file
>>>>>
>>>>>
>>>>> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c
>>>>>
>>>>>             at line 638
>>>>>
>>>>>
>>>>> --------------------------------------------------------------------------
>>>>>             It looks like orte_init failed for some reason; your
>>>>>             parallel process is
>>>>>             likely to abort.  There are many reasons that a parallel
>>>>>             process can
>>>>>             fail during orte_init; some of which are due to
>>>>> configuration or
>>>>>             environment problems.  This failure appears to be an
>>>>>             internal failure;
>>>>>             here's some additional information (which may only be
>>>>>             relevant to an
>>>>>             Open MPI developer):
>>>>>
>>>>>              opal_pmix_base_select failed
>>>>>              --> Returned value Not found (-13) instead of ORTE_SUCCESS
>>>>>
>>>>>
>>>>> --------------------------------------------------------------------------
>>>>>             tyr hello_1 110
>>>>>
>>>>>
>>>>>             Kind regards
>>>>>
>>>>>             Siegmar
>>>>>
>>>>>
>>>>>             Am 21.04.2016 um 17:24 schrieb Ralph Castain:
>>>>>
>>>>>                 Hmmm…it looks like you built the right components, but
>>>>>                 they are not being picked up. Can you run your mpiexec
>>>>>                 command again, adding “-mca
>>>>>                 mca_base_component_show_load_errors 1” to the cmd line?
>>>>>
>>>>>
>>>>>                     On Apr 21, 2016, at 8:16 AM, Siegmar Gross
>>>>>                     <siegmar.gr...@informatik.hs-fulda.de> wrote:
>>>>>
>>>>>                     Hi Ralph,
>>>>>
>>>>>                     I have attached ompi_info output for both compilers
>>>>>                     from my
>>>>>                     sparc machine and the listings for both compilers
>>>>>                     from the
>>>>>                     <prefix>/lib/openmpi directories. Hopefully that
>>>>>                     helps to
>>>>>                     find the problem.
>>>>>
>>>>>                     hermes tmp 3 tar zvft openmpi-2.x_info.tar.gz
>>>>>                     -rw-r--r-- root/root     10969 2016-04-21 17:06
>>>>>                     ompi_info_SunOS_sparc_cc.txt
>>>>>                     -rw-r--r-- root/root     11044 2016-04-21 17:06
>>>>>                     ompi_info_SunOS_sparc_gcc.txt
>>>>>                     -rw-r--r-- root/root     71252 2016-04-21 17:02
>>>>>                     lib64_openmpi.txt
>>>>>                     hermes tmp 4
>>>>>
>>>>>
>>>>>                     Kind regards and thank you very much once more for
>>>>>                     your help
>>>>>
>>>>>                     Siegmar
>>>>>
>>>>>
>>>>>                     Am 21.04.2016 um 15:54 schrieb Ralph Castain:
>>>>>
>>>>>                         Odd - it would appear that none of the pmix
>>>>>                         components built? Can you send
>>>>>                         along the output from ompi_info? Or just send a
>>>>>                         listing of the files in the
>>>>>                         <prefix>/lib/openmpi directory?
>>>>>
>>>>>
>>>>>                             On Apr 21, 2016, at 1:27 AM, Siegmar Gross
>>>>>                             <siegmar.gr...@informatik.hs-fulda.de
>>>>>                             <mailto:
>>>>> siegmar.gr...@informatik.hs-fulda.de>>
>>>>>                             wrote:
>>>>>
>>>>>                             Hi Ralph,
>>>>>
>>>>>                             Am 21.04.2016 um 00:18 schrieb Ralph
>>>>> Castain:
>>>>>
>>>>>                                 Could you please rerun these test and
>>>>>                                 add “-mca pmix_base_verbose 10
>>>>>                                 -mca pmix_server_verbose 5” to your cmd
>>>>>                                 line? I need to see why the
>>>>>                                 pmix components failed.
>>>>>
>>>>>
>>>>>
>>>>>                             tyr spawn 111 mpiexec -np 1 --host
>>>>>                             tyr,sunpc1,linpc1,ruester -mca
>>>>>                             pmix_base_verbose 10 -mca
>>>>>                             pmix_server_verbose 5 spawn_multiple_master
>>>>>                             [tyr.informatik.hs-fulda.de
>>>>>                             <http://tyr.informatik.hs-fulda.de>
>>>>>                             <http://tyr.informatik.hs-fulda.de/>:26652]
>>>>> mca:
>>>>>                             base: components_register: registering
>>>>>                             framework pmix components
>>>>>                             [tyr.informatik.hs-fulda.de
>>>>>                             <http://tyr.informatik.hs-fulda.de>
>>>>>                             <http://tyr.informatik.hs-fulda.de/>:26652]
>>>>> mca:
>>>>>                             base: components_open: opening pmix
>>>>> components
>>>>>                             [tyr.informatik.hs-fulda.de
>>>>>                             <http://tyr.informatik.hs-fulda.de>
>>>>>                             <http://tyr.informatik.hs-fulda.de/
>>>>> >:26652]
>>>>>                             mca:base:select: Auto-selecting pmix
>>>>> components
>>>>>                             [tyr.informatik.hs-fulda.de
>>>>>                             <http://tyr.informatik.hs-fulda.de>
>>>>>                             <http://tyr.informatik.hs-fulda.de/
>>>>> >:26652]
>>>>>                             mca:base:select:( pmix) No component
>>>>> selected!
>>>>>                             [tyr.informatik.hs-fulda.de
>>>>>                             <http://tyr.informatik.hs-fulda.de>
>>>>>                             <http://tyr.informatik.hs-fulda.de/
>>>>> >:26652]
>>>>>                             [[52794,0],0] ORTE_ERROR_LOG: Not found in
>>>>> file
>>>>>
>>>>>
>>>>> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c
>>>>>
>>>>>                             at line 638
>>>>>
>>>>>
>>>>> --------------------------------------------------------------------------
>>>>>                             It looks like orte_init failed for some
>>>>>                             reason; your parallel process is
>>>>>                             likely to abort.  There are many reasons
>>>>>                             that a parallel process can
>>>>>                             fail during orte_init; some of which are
>>>>> due
>>>>>                             to configuration or
>>>>>                             environment problems.  This failure appears
>>>>>                             to be an internal failure;
>>>>>                             here's some additional information (which
>>>>>                             may only be relevant to an
>>>>>                             Open MPI developer):
>>>>>
>>>>>                             opal_pmix_base_select failed
>>>>>                             --> Returned value Not found (-13) instead
>>>>>                             of ORTE_SUCCESS
>>>>>
>>>>>
>>>>> --------------------------------------------------------------------------
>>>>>                             tyr spawn 112
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>                             tyr hello_1 116 mpiexec -np 1 --host
>>>>>                             tyr,sunpc1,linpc1,ruester -mca
>>>>>                             pmix_base_verbose 10 -mca
>>>>>                             pmix_server_verbose 5 hello_1_mpi
>>>>>                             [tyr.informatik.hs-fulda.de
>>>>>                             <http://tyr.informatik.hs-fulda.de>
>>>>>                             <http://tyr.informatik.hs-fulda.de/>:27261]
>>>>> mca:
>>>>>                             base: components_register: registering
>>>>>                             framework pmix components
>>>>>                             [tyr.informatik.hs-fulda.de
>>>>>                             <http://tyr.informatik.hs-fulda.de>
>>>>>                             <http://tyr.informatik.hs-fulda.de/>:27261]
>>>>> mca:
>>>>>                             base: components_open: opening pmix
>>>>> components
>>>>>                             [tyr.informatik.hs-fulda.de
>>>>>                             <http://tyr.informatik.hs-fulda.de>
>>>>>                             <http://tyr.informatik.hs-fulda.de/
>>>>> >:27261]
>>>>>                             mca:base:select: Auto-selecting pmix
>>>>> components
>>>>>                             [tyr.informatik.hs-fulda.de
>>>>>                             <http://tyr.informatik.hs-fulda.de>
>>>>>                             <http://tyr.informatik.hs-fulda.de/
>>>>> >:27261]
>>>>>                             mca:base:select:( pmix) No component
>>>>> selected!
>>>>>                             [tyr.informatik.hs-fulda.de
>>>>>                             <http://tyr.informatik.hs-fulda.de>
>>>>>                             <http://tyr.informatik.hs-fulda.de/
>>>>> >:27261]
>>>>>                             [[52315,0],0] ORTE_ERROR_LOG: Not found in
>>>>> file
>>>>>
>>>>>
>>>>> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c
>>>>>
>>>>>                             at line 638
>>>>>
>>>>>
>>>>> --------------------------------------------------------------------------
>>>>>                             It looks like orte_init failed for some
>>>>>                             reason; your parallel process is
>>>>>                             likely to abort.  There are many reasons
>>>>>                             that a parallel process can
>>>>>                             fail during orte_init; some of which are
>>>>> due
>>>>>                             to configuration or
>>>>>                             environment problems.  This failure appears
>>>>>                             to be an internal failure;
>>>>>                             here's some additional information (which
>>>>>                             may only be relevant to an
>>>>>                             Open MPI developer):
>>>>>
>>>>>                             opal_pmix_base_select failed
>>>>>                             --> Returned value Not found (-13) instead
>>>>>                             of ORTE_SUCCESS
>>>>>
>>>>>
>>>>> --------------------------------------------------------------------------
>>>>>                             tyr hello_1 117
>>>>>
>>>>>
>>>>>
>>>>>                             Thank you very much for your help.
>>>>>
>>>>>
>>>>>                             Kind regards
>>>>>
>>>>>                             Siegmar
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>                                 Thanks
>>>>>                                 Ralph
>>>>>
>>>>>                                     On Apr 20, 2016, at 10:12 AM,
>>>>>                                     Siegmar Gross
>>>>>                                     <
>>>>> siegmar.gr...@informatik.hs-fulda.de
>>>>>
>>>>> <mailto:siegmar.gr...@informatik.hs-fulda.de>>
>>>>>                                     wrote:
>>>>>
>>>>>                                     Hi,
>>>>>
>>>>>                                     I have built
>>>>>                                     openmpi-v2.x-dev-1280-gc110ae8 on
>>>>> my
>>>>>                                     machines
>>>>>                                     (Solaris 10 Sparc, Solaris 10
>>>>>                                     x86_64, and openSUSE Linux
>>>>>                                     12.1 x86_64) with gcc-5.1.0 and Sun
>>>>>                                     C 5.13. Unfortunately I get
>>>>>                                     runtime errors for some programs.
>>>>>
>>>>>
>>>>>                                     Sun C 5.13:
>>>>>                                     ===========
>>>>>
>>>>>                                     For all my test programs I get the
>>>>>                                     same error on Solaris Sparc and
>>>>>                                     Solaris x86_64, while the programs
>>>>>                                     work fine on Linux.
>>>>>
>>>>>                                     tyr hello_1 115 mpiexec -np 2
>>>>>                                     hello_1_mpi
>>>>>                                     [tyr.informatik.hs-fulda.de
>>>>>                                     <http://tyr.informatik.hs-fulda.de
>>>>> >
>>>>>
>>>>> <http://tyr.informatik.hs-fulda.de>:22373]
>>>>>                                     [[61763,0],0] ORTE_ERROR_LOG: Not
>>>>>                                     found in file
>>>>>
>>>>>
>>>>> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c
>>>>>
>>>>>                                     at line 638
>>>>>
>>>>>
>>>>> --------------------------------------------------------------------------
>>>>>                                     It looks like orte_init failed for
>>>>>                                     some reason; your parallel process
>>>>> is
>>>>>                                     likely to abort.  There are many
>>>>>                                     reasons that a parallel process can
>>>>>                                     fail during orte_init; some of
>>>>> which
>>>>>                                     are due to configuration or
>>>>>                                     environment problems.  This failure
>>>>>                                     appears to be an internal failure;
>>>>>                                     here's some additional information
>>>>>                                     (which may only be relevant to an
>>>>>                                     Open MPI developer):
>>>>>
>>>>>                                     opal_pmix_base_select failed
>>>>>                                     --> Returned value Not found (-13)
>>>>>                                     instead of ORTE_SUCCESS
>>>>>
>>>>>
>>>>> --------------------------------------------------------------------------
>>>>>                                     tyr hello_1 116
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>                                     GCC-5.1.0:
>>>>>                                     ==========
>>>>>
>>>>>                                     tyr spawn 121 mpiexec -np 1 --host
>>>>>                                     tyr,sunpc1,linpc1,ruester
>>>>>                                     spawn_multiple_master
>>>>>
>>>>>                                     Parent process 0 running on
>>>>>                                     tyr.informatik.hs-fulda.de
>>>>>                                     <http://tyr.informatik.hs-fulda.de
>>>>> >
>>>>>                                     <http://tyr.informatik.hs-fulda.de
>>>>> >
>>>>>                                     I create 3 slave processes.
>>>>>
>>>>>                                     [tyr.informatik.hs-fulda.de
>>>>>                                     <http://tyr.informatik.hs-fulda.de
>>>>> >
>>>>>
>>>>> <http://tyr.informatik.hs-fulda.de>:25366]
>>>>>                                     PMIX ERROR: UNPACK-PAST-END in file
>>>>>
>>>>>
>>>>> ../../../../../../openmpi-v2.x-dev-1280-gc110ae8/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c
>>>>>
>>>>>
>>>>>                                     at line 829
>>>>>                                     [tyr.informatik.hs-fulda.de
>>>>>                                     <http://tyr.informatik.hs-fulda.de
>>>>> >
>>>>>
>>>>> <http://tyr.informatik.hs-fulda.de>:25366]
>>>>>                                     PMIX ERROR: UNPACK-PAST-END in file
>>>>>
>>>>>
>>>>> ../../../../../../openmpi-v2.x-dev-1280-gc110ae8/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c
>>>>>
>>>>>
>>>>>                                     at line 2176
>>>>>                                     [tyr:25377] *** An error occurred
>>>>> in
>>>>>                                     MPI_Comm_spawn_multiple
>>>>>                                     [tyr:25377] *** reported by process
>>>>>                                     [3308257281,0]
>>>>>                                     [tyr:25377] *** on communicator
>>>>>                                     MPI_COMM_WORLD
>>>>>                                     [tyr:25377] *** MPI_ERR_SPAWN:
>>>>> could
>>>>>                                     not spawn processes
>>>>>                                     [tyr:25377] ***
>>>>> MPI_ERRORS_ARE_FATAL
>>>>>                                     (processes in this communicator
>>>>> will
>>>>>                                     now abort,
>>>>>                                     [tyr:25377] ***    and potentially
>>>>>                                     your MPI job)
>>>>>                                     tyr spawn 122
>>>>>
>>>>>
>>>>>                                     I would be grateful if somebody can
>>>>>                                     fix the problems. Thank you very
>>>>>                                     much for any help in advance.
>>>>>
>>>>>
>>>>>                                     Kind regards
>>>>>
>>>>>                                     Siegmar
>>>>>
>>>>>
>>>>> <hello_1_mpi.c><spawn_multiple_master.c>_______________________________________________
>>>>>
>>>>>
>>>>>                                     users mailing list
>>>>>                                     us...@open-mpi.org
>>>>> <mailto:us...@open-mpi.org>
>>>>>                                     Subscription:
>>>>>
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>                                     Link to this post:
>>>>>
>>>>> http://www.open-mpi.org/community/lists/users/2016/04/28983.php
>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>>                                 users mailing list
>>>>>                                 us...@open-mpi.org
>>>>> <mailto:us...@open-mpi.org>
>>>>>                                 Subscription:
>>>>>
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>                                 Link to this
>>>>>                                 post:
>>>>>
>>>>> http://www.open-mpi.org/community/lists/users/2016/04/28986.php
>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>>                             users mailing list
>>>>>                             us...@open-mpi.org <mailto:
>>>>> us...@open-mpi.org>
>>>>>                             Subscription:
>>>>>
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>                             Link to this
>>>>>                             post:
>>>>>
>>>>> http://www.open-mpi.org/community/lists/users/2016/04/28987.php
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>                         _______________________________________________
>>>>>                         users mailing list
>>>>>                         us...@open-mpi.org
>>>>>                         Subscription:
>>>>>
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>                         Link to this post:
>>>>>
>>>>> http://www.open-mpi.org/community/lists/users/2016/04/28988.php
>>>>>
>>>>>
>>>>>
>>>>> <openmpi-2.x_info.tar.gz>_______________________________________________
>>>>>                     users mailing list
>>>>>                     us...@open-mpi.org
>>>>>                     Subscription:
>>>>>                     http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>                     Link to this post:
>>>>>
>>>>> http://www.open-mpi.org/community/lists/users/2016/04/28989.php
>>>>>
>>>>>
>>>>>                 _______________________________________________
>>>>>                 users mailing list
>>>>>                 us...@open-mpi.org
>>>>>                 Subscription:
>>>>>                 http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>                 Link to this post:
>>>>>
>>>>> http://www.open-mpi.org/community/lists/users/2016/04/28990.php
>>>>>
>>>>>
>>>>>             _______________________________________________
>>>>>             users mailing list
>>>>>             us...@open-mpi.org
>>>>>             Subscription:
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>             Link to this post:
>>>>>
>>>>> http://www.open-mpi.org/community/lists/users/2016/04/28991.php
>>>>>
>>>>>
>>>>>         _______________________________________________
>>>>>         users mailing list
>>>>>         us...@open-mpi.org
>>>>>         Subscription:
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>         Link to this post:
>>>>>
>>>>> http://www.open-mpi.org/community/lists/users/2016/04/28992.php
>>>>>
>>>>>     _______________________________________________
>>>>>     users mailing list
>>>>>     us...@open-mpi.org
>>>>>     Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>     Link to this post:
>>>>>     http://www.open-mpi.org/community/lists/users/2016/04/28993.php
>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> Link to this post:
>>>>> http://www.open-mpi.org/community/lists/users/2016/04/28999.php
>>>>>
>>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> Link to this post:
>>>> http://www.open-mpi.org/community/lists/users/2016/04/29009.php
>>>>
>>>
>>>
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> Link to this post:
>>> http://www.open-mpi.org/community/lists/users/2016/04/29033.php
>>>
>>
>>
>>
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users
>> Link to this post:
>> http://www.open-mpi.org/community/lists/users/2016/04/29038.php
>>
>>
> --
>
> ########################################################################
> #                                                                      #
> # Hochschule Fulda          University of Applied Sciences             #
> # FB Angewandte Informatik  Department of Applied Computer Science     #
> #                                                                      #
> # Prof. Dr. Siegmar Gross   Tel.: +49 (0)661 9640 - 333                #
> #                           Fax:  +49 (0)661 9640 - 349                #
> # Leipziger Str. 123        WWW:  http://www.hs-fulda.de/~gross        #
> # D-36037 Fulda             Mail: siegmar.gr...@informatik.hs-fulda.de #
> #                                                                      #
> #                                                                      #
> # IT-Sicherheit: http://www.hs-fulda.de/it-sicherheit                  #
> #                                                                      #
> ########################################################################
>
>

Reply via email to