[OMPI users] opal_pmix_base_select failed for master and 4.0.0

2018-10-02 Thread Siegmar Gross

Hi,

yesterday I've installed openmpi-v4.0.x-201809290241-a7e275c and
openmpi-master-201805080348-b39bbfb on my "SUSE Linux Enterprise Server
12.3 (x86_64)" with Sun C 5.15, gcc 6.4.0, Intel icc 18.0.3, and Portland
Group pgcc 18.4-0. Unfortunately, I get the following error for all seven
installed versions (Sun C couldn't built master as I mentioned in another
email).


loki hello_1 118 mpiexec -np 4 --host loki:2,nfs2:2 hello_1_mpi
[loki:11423] [[45859,0],0] ORTE_ERROR_LOG: Not found in file 
../../../../../openmpi-v4.0.x-201809290241-a7e275c/orte/mca/ess/hnp/ess_hnp_module.c 
at line 321

--
It looks like orte_init failed for some reason; your parallel process is
likely to abort.  There are many reasons that a parallel process can
fail during orte_init; some of which are due to configuration or
environment problems.  This failure appears to be an internal failure;
here's some additional information (which may only be relevant to an
Open MPI developer):

  opal_pmix_base_select failed
  --> Returned value Not found (-13) instead of ORTE_SUCCESS
--
loki hello_1 119



I would be grateful, if somebody can fix the problem. Do you need anything
else? Thank you very much for any help in advance.


Kind regards

Siegmar
___
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users


Re: [OMPI users] [open-mpi/ompi] vader compile issue (#5814)

2018-10-02 Thread Nathan Hjelm via users
hmm. Add

#include 

to the test and try it again.

-Nathan

> On Oct 2, 2018, at 12:41 AM, Siegmar Gross 
>  wrote:
> 
> Hi Jeff, hi Nathan,
> 
> the compilers (Sun C 5.15, Sun C 5.14, Sun C 5.13) don't like the code.
> 
> loki tmp 110 cc -V
> cc: Studio 12.6 Sun C 5.15 Linux_i386 2017/05/30
> loki tmp 111 \cc -std=c11 atomic_test.c
> "atomic_test.c", line 5: warning: no explicit type given
> "atomic_test.c", line 5: syntax error before or at: test
> "atomic_test.c", line 8: undefined symbol: test
> "atomic_test.c", line 8: undefined symbol: x
> cc: acomp failed for atomic_test.c
> loki tmp 112
> 
> 
> loki tmp 111 cc -V
> cc: Studio 12.5 Sun C 5.14 Linux_i386 2016/05/31
> loki tmp 112 \cc -std=c11 atomic_test.c
> "atomic_test.c", line 5: warning: no explicit type given
> "atomic_test.c", line 5: syntax error before or at: test
> "atomic_test.c", line 8: undefined symbol: test
> "atomic_test.c", line 8: undefined symbol: x
> cc: acomp failed for atomic_test.c
> loki tmp 113
> 
> 
> loki tmp 108 cc -V
> cc: Sun C 5.13 Linux_i386 2014/10/20
> loki tmp 109 \cc -std=c11 atomic_test.c
> "atomic_test.c", line 2: cannot find include file: 
> "atomic_test.c", line 5: warning: _Atomic is a keyword in ISO C11
> "atomic_test.c", line 5: undefined symbol: _Atomic
> "atomic_test.c", line 5: syntax error before or at: intptr_t
> "atomic_test.c", line 6: undefined symbol: intptr_t
> "atomic_test.c", line 8: undefined symbol: test
> "atomic_test.c", line 8: undefined symbol: x
> cc: acomp failed for atomic_test.c
> loki tmp 110
> 
> 
> I have attached the file config.log.gist from my master build, although I
> didn't know what the gist is. Let me know if you need something different
> from that file. By the way, I was able to build the upcoming version 4.0.0.
> 
> loki openmpi-v4.0.x-201809290241-a7e275c-Linux.x86_64.64_cc 124 grep Error 
> log.*
> log.make-install.Linux.x86_64.64_cc: /usr/bin/install -c -m 644 
> mpi/man/man3/MPI_Compare_and_swap.3 mpi/man/man3/MPI_Dims_create.3 
> mpi/man/man3/MPI_Dist_graph_create.3 
> mpi/man/man3/MPI_Dist_graph_create_adjacent.3 
> mpi/man/man3/MPI_Dist_graph_neighbors.3 
> mpi/man/man3/MPI_Dist_graph_neighbors_count.3 
> mpi/man/man3/MPI_Errhandler_create.3 mpi/man/man3/MPI_Errhandler_free.3 
> mpi/man/man3/MPI_Errhandler_get.3 mpi/man/man3/MPI_Errhandler_set.3 
> mpi/man/man3/MPI_Error_class.3 mpi/man/man3/MPI_Error_string.3 
> mpi/man/man3/MPI_Exscan.3 mpi/man/man3/MPI_Iexscan.3 
> mpi/man/man3/MPI_Fetch_and_op.3 mpi/man/man3/MPI_File_c2f.3 
> mpi/man/man3/MPI_File_call_errhandler.3 mpi/man/man3/MPI_File_close.3 
> mpi/man/man3/MPI_File_create_errhandler.3 mpi/man/man3/MPI_File_delete.3 
> mpi/man/man3/MPI_File_f2c.3 mpi/man/man3/MPI_File_get_amode.3 
> mpi/man/man3/MPI_File_get_atomicity.3 mpi/man/man3/MPI_File_get_byte_offset.3 
> mpi/man/man3/MPI_File_get_errhandler.3 mpi/man/man3/MPI_File_get_group.3 
> mpi/man/man3/MPI_File_get_info.3 mpi/man/man3/MPI_File_get_position.3 
> mpi/man/man3/MPI_File_get_position_shared.3 mpi/man/man3/MPI_File_get_size.3 
> mpi/man/man3/MPI_File_get_type_extent.3 mpi/man/man3/MPI_File_get_view.3 
> mpi/man/man3/MPI_File_iread.3 mpi/man/man3/MPI_File_iread_at.3 
> mpi/man/man3/MPI_File_iread_all.3 mpi/man/man3/MPI_File_iread_at_all.3 
> mpi/man/man3/MPI_File_iread_shared.3 mpi/man/man3/MPI_File_iwrite.3 
> mpi/man/man3/MPI_File_iwrite_at.3 mpi/man/man3/MPI_File_iwrite_all.3 
> '/usr/local/openmpi-4.0.0_64_cc/share/man/man3'
> log.make.Linux.x86_64.64_cc:  GENERATE mpi/man/man3/MPI_Error_class.3
> log.make.Linux.x86_64.64_cc:  GENERATE mpi/man/man3/MPI_Error_string.3
> loki openmpi-v4.0.x-201809290241-a7e275c-Linux.x86_64.64_cc 125
> 
> 
> 
> Best regards and thank you very much for your help
> 
> Siegmar
> 
> 
>> On 10/01/18 22:07, Jeff Squyres wrote:
>> @siegmargross  Nathan posted a sample 
>> program (via editing his prior comment), so you didn't get the mail about 
>> it. Can you check #5814 (comment) 
>>  and 
>> compile/run the sample program he proposed and see what happens?
>> —
>> You are receiving this because you were mentioned.
>> Reply to this email directly, view it on GitHub 
>> , or 
>> mute the thread 
>> .
> 
> ___
> users mailing list
> users@lists.open-mpi.org
> https://lists.open-mpi.org/mailman/listinfo/users
___
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

Re: [OMPI users] opal_pmix_base_select failed for master and 4.0.0

2018-10-02 Thread Ralph H Castain
Looks like PMIx failed to build - can you send the config.log?

> On Oct 2, 2018, at 12:00 AM, Siegmar Gross 
>  wrote:
> 
> Hi,
> 
> yesterday I've installed openmpi-v4.0.x-201809290241-a7e275c and
> openmpi-master-201805080348-b39bbfb on my "SUSE Linux Enterprise Server
> 12.3 (x86_64)" with Sun C 5.15, gcc 6.4.0, Intel icc 18.0.3, and Portland
> Group pgcc 18.4-0. Unfortunately, I get the following error for all seven
> installed versions (Sun C couldn't built master as I mentioned in another
> email).
> 
> 
> loki hello_1 118 mpiexec -np 4 --host loki:2,nfs2:2 hello_1_mpi
> [loki:11423] [[45859,0],0] ORTE_ERROR_LOG: Not found in file 
> ../../../../../openmpi-v4.0.x-201809290241-a7e275c/orte/mca/ess/hnp/ess_hnp_module.c
>  at line 321
> --
> It looks like orte_init failed for some reason; your parallel process is
> likely to abort.  There are many reasons that a parallel process can
> fail during orte_init; some of which are due to configuration or
> environment problems.  This failure appears to be an internal failure;
> here's some additional information (which may only be relevant to an
> Open MPI developer):
> 
>  opal_pmix_base_select failed
>  --> Returned value Not found (-13) instead of ORTE_SUCCESS
> --
> loki hello_1 119
> 
> 
> 
> I would be grateful, if somebody can fix the problem. Do you need anything
> else? Thank you very much for any help in advance.
> 
> 
> Kind regards
> 
> Siegmar
> ___
> users mailing list
> users@lists.open-mpi.org
> https://lists.open-mpi.org/mailman/listinfo/users

___
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users


Re: [OMPI users] [open-mpi/ompi] vader compile issue (#5814)

2018-10-02 Thread Siegmar Gross

Hi Nathan,

it works for Sun C 5.14 and Sun C 5.15.

loki tmp 111 cc atomic_test.c
loki tmp 112 a.out
loki tmp 113 cc -V
cc: Studio 12.5 Sun C 5.14 Linux_i386 2016/05/31
loki tmp 114 exit


loki tmp 113 cc -V
cc: Studio 12.6 Sun C 5.15 Linux_i386 2017/05/30
loki tmp 114 cc atomic_test.c
loki tmp 115 a.out

loki tmp 116 more atomic_test.c
#include 
#include 

int main (int argc, char *argv[]) {
_Atomic intptr_t test;
intptr_t x = 0;

test = x;

return 0;
}
loki tmp 117


Best regards

Siegmar


Am 02.10.2018 um 13:08 schrieb Nathan Hjelm via users:

hmm. Add

#include 

to the test and try it again.

-Nathan


On Oct 2, 2018, at 12:41 AM, Siegmar Gross 
 wrote:

Hi Jeff, hi Nathan,

the compilers (Sun C 5.15, Sun C 5.14, Sun C 5.13) don't like the code.

loki tmp 110 cc -V
cc: Studio 12.6 Sun C 5.15 Linux_i386 2017/05/30
loki tmp 111 \cc -std=c11 atomic_test.c
"atomic_test.c", line 5: warning: no explicit type given
"atomic_test.c", line 5: syntax error before or at: test
"atomic_test.c", line 8: undefined symbol: test
"atomic_test.c", line 8: undefined symbol: x
cc: acomp failed for atomic_test.c
loki tmp 112


loki tmp 111 cc -V
cc: Studio 12.5 Sun C 5.14 Linux_i386 2016/05/31
loki tmp 112 \cc -std=c11 atomic_test.c
"atomic_test.c", line 5: warning: no explicit type given
"atomic_test.c", line 5: syntax error before or at: test
"atomic_test.c", line 8: undefined symbol: test
"atomic_test.c", line 8: undefined symbol: x
cc: acomp failed for atomic_test.c
loki tmp 113


loki tmp 108 cc -V
cc: Sun C 5.13 Linux_i386 2014/10/20
loki tmp 109 \cc -std=c11 atomic_test.c
"atomic_test.c", line 2: cannot find include file: 
"atomic_test.c", line 5: warning: _Atomic is a keyword in ISO C11
"atomic_test.c", line 5: undefined symbol: _Atomic
"atomic_test.c", line 5: syntax error before or at: intptr_t
"atomic_test.c", line 6: undefined symbol: intptr_t
"atomic_test.c", line 8: undefined symbol: test
"atomic_test.c", line 8: undefined symbol: x
cc: acomp failed for atomic_test.c
loki tmp 110


I have attached the file config.log.gist from my master build, although I
didn't know what the gist is. Let me know if you need something different
from that file. By the way, I was able to build the upcoming version 4.0.0.

loki openmpi-v4.0.x-201809290241-a7e275c-Linux.x86_64.64_cc 124 grep Error log.*
log.make-install.Linux.x86_64.64_cc: /usr/bin/install -c -m 644 
mpi/man/man3/MPI_Compare_and_swap.3 mpi/man/man3/MPI_Dims_create.3 
mpi/man/man3/MPI_Dist_graph_create.3 
mpi/man/man3/MPI_Dist_graph_create_adjacent.3 
mpi/man/man3/MPI_Dist_graph_neighbors.3 
mpi/man/man3/MPI_Dist_graph_neighbors_count.3 
mpi/man/man3/MPI_Errhandler_create.3 mpi/man/man3/MPI_Errhandler_free.3 
mpi/man/man3/MPI_Errhandler_get.3 mpi/man/man3/MPI_Errhandler_set.3 
mpi/man/man3/MPI_Error_class.3 mpi/man/man3/MPI_Error_string.3 
mpi/man/man3/MPI_Exscan.3 mpi/man/man3/MPI_Iexscan.3 
mpi/man/man3/MPI_Fetch_and_op.3 mpi/man/man3/MPI_File_c2f.3 
mpi/man/man3/MPI_File_call_errhandler.3 mpi/man/man3/MPI_File_close.3 
mpi/man/man3/MPI_File_create_errhandler.3 mpi/man/man3/MPI_File_delete.3 
mpi/man/man3/MPI_File_f2c.3 mpi/man/man3/MPI_File_get_amode.3 
mpi/man/man3/MPI_File_get_atomicity.3 mpi/man/man3/MPI_File_get_byte_offset.3 
mpi/man/man3/MPI_File_get_errhandler.3 mpi/man/man3/MPI_File_get_group.3 
mpi/man/man3/MPI_File_get_info.3 mpi/man/man3/MPI_File_get_position.3 
mpi/man/man3/MPI_File_get_position_shared.3 mpi/man/man3/MPI_File_get_size.3 
mpi/man/man3/MPI_File_get_type_extent.3 mpi/man/man3/MPI_File_get_view.3 
mpi/man/man3/MPI_File_iread.3 mpi/man/man3/MPI_File_iread_at.3 
mpi/man/man3/MPI_File_iread_all.3 mpi/man/man3/MPI_File_iread_at_all.3 
mpi/man/man3/MPI_File_iread_shared.3 mpi/man/man3/MPI_File_iwrite.3 
mpi/man/man3/MPI_File_iwrite_at.3 mpi/man/man3/MPI_File_iwrite_all.3 
'/usr/local/openmpi-4.0.0_64_cc/share/man/man3'
log.make.Linux.x86_64.64_cc:  GENERATE mpi/man/man3/MPI_Error_class.3
log.make.Linux.x86_64.64_cc:  GENERATE mpi/man/man3/MPI_Error_string.3
loki openmpi-v4.0.x-201809290241-a7e275c-Linux.x86_64.64_cc 125



Best regards and thank you very much for your help

Siegmar



On 10/01/18 22:07, Jeff Squyres wrote:
@siegmargross  Nathan posted a sample program (via 
editing his prior comment), so you didn't get the mail about it. Can you check #5814 
(comment)  and 
compile/run the sample program he proposed and see what happens?
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub 
, or mute the 
thread 
.


___
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

___
use

Re: [OMPI users] opal_pmix_base_select failed for master and 4.0.0

2018-10-02 Thread Ralph H Castain
So the problem is here when configuring the internal PMIx code:

configure:3383: === HWLOC
configure:36189: checking for hwloc in
configure:36201: result: Could not find internal/lib or internal/lib64
configure:36203: error: Can not continue

Can you confirm that HWLOC built? I believe we require it, but perhaps 
something is different about this environment.


> On Oct 2, 2018, at 6:36 AM, Ralph H Castain  wrote:
> 
> Looks like PMIx failed to build - can you send the config.log?
> 
>> On Oct 2, 2018, at 12:00 AM, Siegmar Gross 
>>  wrote:
>> 
>> Hi,
>> 
>> yesterday I've installed openmpi-v4.0.x-201809290241-a7e275c and
>> openmpi-master-201805080348-b39bbfb on my "SUSE Linux Enterprise Server
>> 12.3 (x86_64)" with Sun C 5.15, gcc 6.4.0, Intel icc 18.0.3, and Portland
>> Group pgcc 18.4-0. Unfortunately, I get the following error for all seven
>> installed versions (Sun C couldn't built master as I mentioned in another
>> email).
>> 
>> 
>> loki hello_1 118 mpiexec -np 4 --host loki:2,nfs2:2 hello_1_mpi
>> [loki:11423] [[45859,0],0] ORTE_ERROR_LOG: Not found in file 
>> ../../../../../openmpi-v4.0.x-201809290241-a7e275c/orte/mca/ess/hnp/ess_hnp_module.c
>>  at line 321
>> --
>> It looks like orte_init failed for some reason; your parallel process is
>> likely to abort.  There are many reasons that a parallel process can
>> fail during orte_init; some of which are due to configuration or
>> environment problems.  This failure appears to be an internal failure;
>> here's some additional information (which may only be relevant to an
>> Open MPI developer):
>> 
>> opal_pmix_base_select failed
>> --> Returned value Not found (-13) instead of ORTE_SUCCESS
>> --
>> loki hello_1 119
>> 
>> 
>> 
>> I would be grateful, if somebody can fix the problem. Do you need anything
>> else? Thank you very much for any help in advance.
>> 
>> 
>> Kind regards
>> 
>> Siegmar
>> ___
>> users mailing list
>> users@lists.open-mpi.org
>> https://lists.open-mpi.org/mailman/listinfo/users
> 
> ___
> users mailing list
> users@lists.open-mpi.org
> https://lists.open-mpi.org/mailman/listinfo/users

___
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users


Re: [OMPI users] [open-mpi/ompi] vader compile issue (#5814)

2018-10-02 Thread Nathan Hjelm via users
Definitely a compiler bug. I opened a PR to work around it and posted a question on the Oracle forums.-NathanOn Oct 02, 2018, at 12:48 AM, Siegmar Gross  wrote:Hi Jeff, hi Nathan,the compilers (Sun C 5.15, Sun C 5.14, Sun C 5.13) don't like the code.loki tmp 110 cc -Vcc: Studio 12.6 Sun C 5.15 Linux_i386 2017/05/30loki tmp 111 \cc -std=c11 atomic_test.c"atomic_test.c", line 5: warning: no explicit type given"atomic_test.c", line 5: syntax error before or at: test"atomic_test.c", line 8: undefined symbol: test"atomic_test.c", line 8: undefined symbol: xcc: acomp failed for atomic_test.cloki tmp 112loki tmp 111 cc -Vcc: Studio 12.5 Sun C 5.14 Linux_i386 2016/05/31loki tmp 112 \cc -std=c11 atomic_test.c"atomic_test.c", line 5: warning: no explicit type given"atomic_test.c", line 5: syntax error before or at: test"atomic_test.c", line 8: undefined symbol: test"atomic_test.c", line 8: undefined symbol: xcc: acomp failed for atomic_test.cloki tmp 113loki tmp 108 cc -Vcc: Sun C 5.13 Linux_i386 2014/10/20loki tmp 109 \cc -std=c11 atomic_test.c"atomic_test.c", line 2: cannot find include file: "atomic_test.c", line 5: warning: _Atomic is a keyword in ISO C11"atomic_test.c", line 5: undefined symbol: _Atomic"atomic_test.c", line 5: syntax error before or at: intptr_t"atomic_test.c", line 6: undefined symbol: intptr_t"atomic_test.c", line 8: undefined symbol: test"atomic_test.c", line 8: undefined symbol: xcc: acomp failed for atomic_test.cloki tmp 110I have attached the file config.log.gist from my master build, although Ididn't know what the gist is. Let me know if you need something differentfrom that file. By the way, I was able to build the upcoming version 4.0.0.loki openmpi-v4.0.x-201809290241-a7e275c-Linux.x86_64.64_cc 124 grep Error log.*log.make-install.Linux.x86_64.64_cc: /usr/bin/install -c -m 644 mpi/man/man3/MPI_Compare_and_swap.3 mpi/man/man3/MPI_Dims_create.3 mpi/man/man3/MPI_Dist_graph_create.3 mpi/man/man3/MPI_Dist_graph_create_adjacent.3 mpi/man/man3/MPI_Dist_graph_neighbors.3 mpi/man/man3/MPI_Dist_graph_neighbors_count.3 mpi/man/man3/MPI_Errhandler_create.3 mpi/man/man3/MPI_Errhandler_free.3 mpi/man/man3/MPI_Errhandler_get.3 mpi/man/man3/MPI_Errhandler_set.3 mpi/man/man3/MPI_Error_class.3 mpi/man/man3/MPI_Error_string.3 mpi/man/man3/MPI_Exscan.3 mpi/man/man3/MPI_Iexscan.3 mpi/man/man3/MPI_Fetch_and_op.3 mpi/man/man3/MPI_File_c2f.3 mpi/man/man3/MPI_File_call_errhandler.3 mpi/man/man3/MPI_File_close.3 mpi/man/man3/MPI_File_create_errhandler.3 mpi/man/man3/MPI_File_delete.3 mpi/man/man3/MPI_File_f2c.3 mpi/man/man3/MPI_File_get_amode.3 mpi/man/man3/MPI_File_get_atomicity.3 mpi/man/man3/MPI_File_get_byte_offset.3 mpi/man/man3/MPI_File_get_errhandler.3 mpi/man/man3/MPI_File_get_group.3 mpi/man/man3/MPI_File_get_info.3 mpi/man/man3/MPI_File_get_position.3 mpi/man/man3/MPI_File_get_position_shared.3 mpi/man/man3/MPI_File_get_size.3 mpi/man/man3/MPI_File_get_type_extent.3 mpi/man/man3/MPI_File_get_view.3 mpi/man/man3/MPI_File_iread.3 mpi/man/man3/MPI_File_iread_at.3 mpi/man/man3/MPI_File_iread_all.3 mpi/man/man3/MPI_File_iread_at_all.3 mpi/man/man3/MPI_File_iread_shared.3 mpi/man/man3/MPI_File_iwrite.3 mpi/man/man3/MPI_File_iwrite_at.3 mpi/man/man3/MPI_File_iwrite_all.3 '/usr/local/openmpi-4.0.0_64_cc/share/man/man3'log.make.Linux.x86_64.64_cc: GENERATE mpi/man/man3/MPI_Error_class.3log.make.Linux.x86_64.64_cc: GENERATE mpi/man/man3/MPI_Error_string.3loki openmpi-v4.0.x-201809290241-a7e275c-Linux.x86_64.64_cc 125Best regards and thank you very much for your helpSiegmarOn 10/01/18 22:07, Jeff Squyres wrote:@siegmargross  Nathan posted a sample program(via editing his prior comment), so you didn't get the mail about it. Can youcheck #5814 (comment) andcompile/run the sample program he proposed and see what happens?—You are receiving this because you were mentioned.Reply to this email directly, view it on GitHub, or mutethe thread.___users mailing listusers@lists.open-mpi.orghttps://lists.open-mpi.org/mailman/listinfo/usersThis file contains any messages produced by compilers while
running configure, to aid debugging if configure makes a mistake.

It was created by Open MPI configure master-201809290304-73075b8, which was
generated by GNU Autoconf 2.69.  Invocation command line was

  $ ../openmpi-master-201809290304-73075b8/configure 
--prefix=/usr/local/openmpi-master_64_cc 
--libdir=/usr/local/openmpi-master_64_cc/lib64 
--with-jdk-bindir=/usr/local/jdk-10.0.1/bin 
--with-jdk-headers=/usr/local/jdk-10.0.1/include 
JAVA_HOME=/usr/local/jdk-10.0.1 LDFLAGS=-m64 -mt -Wl,-z -Wl,noexecstack 
-L/usr/local/lib64 CC=cc CXX=CC FC=f95 CFLAGS=-m64 -mt CXXFLAGS=-m64 
FCFLAGS=-m64 CPP=cpp CXXCPP=cpp --disab

Re: [OMPI users] opal_pmix_base_select failed for master and 4.0.0

2018-10-02 Thread Ralph H Castain
Hi Siegmar

I honestly have no idea - for some reason, the PMIx component isn’t seeing the 
internal hwloc code in your environment.

Jeff, Brice - any ideas?


> On Oct 2, 2018, at 1:18 PM, Siegmar Gross 
>  wrote:
> 
> Hi Ralph,
> 
> how can I confirm that HWLOC built? Some hwloc files are available
> in the built directory.
> 
> loki openmpi-master-201809290304-73075b8-Linux.x86_64.64_gcc 111 find . -name 
> '*hwloc*'
> ./opal/mca/btl/usnic/.deps/btl_usnic_hwloc.Plo
> ./opal/mca/hwloc
> ./opal/mca/hwloc/external/.deps/hwloc_external_component.Plo
> ./opal/mca/hwloc/base/hwloc_base_frame.lo
> ./opal/mca/hwloc/base/.deps/hwloc_base_dt.Plo
> ./opal/mca/hwloc/base/.deps/hwloc_base_maffinity.Plo
> ./opal/mca/hwloc/base/.deps/hwloc_base_frame.Plo
> ./opal/mca/hwloc/base/.deps/hwloc_base_util.Plo
> ./opal/mca/hwloc/base/hwloc_base_dt.lo
> ./opal/mca/hwloc/base/hwloc_base_util.lo
> ./opal/mca/hwloc/base/hwloc_base_maffinity.lo
> ./opal/mca/hwloc/base/.libs/hwloc_base_util.o
> ./opal/mca/hwloc/base/.libs/hwloc_base_dt.o
> ./opal/mca/hwloc/base/.libs/hwloc_base_maffinity.o
> ./opal/mca/hwloc/base/.libs/hwloc_base_frame.o
> ./opal/mca/hwloc/.libs/libmca_hwloc.la
> ./opal/mca/hwloc/.libs/libmca_hwloc.a
> ./opal/mca/hwloc/libmca_hwloc.la
> ./opal/mca/hwloc/hwloc201
> ./opal/mca/hwloc/hwloc201/.deps/hwloc201_component.Plo
> ./opal/mca/hwloc/hwloc201/hwloc201_component.lo
> ./opal/mca/hwloc/hwloc201/hwloc
> ./opal/mca/hwloc/hwloc201/hwloc/include/hwloc
> ./opal/mca/hwloc/hwloc201/hwloc/hwloc
> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/libhwloc_embedded.la
> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_pci_la-topology-pci.Plo
> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_gl_la-topology-gl.Plo
> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_cuda_la-topology-cuda.Plo
> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_xml_libxml_la-topology-xml-libxml.Plo
> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_opencl_la-topology-opencl.Plo
> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_nvml_la-topology-nvml.Plo
> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.libs/libhwloc_embedded.la
> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.libs/libhwloc_embedded.a
> ./opal/mca/hwloc/hwloc201/.libs/hwloc201_component.o
> ./opal/mca/hwloc/hwloc201/.libs/libmca_hwloc_hwloc201.la
> ./opal/mca/hwloc/hwloc201/.libs/libmca_hwloc_hwloc201.a
> ./opal/mca/hwloc/hwloc201/libmca_hwloc_hwloc201.la
> ./orte/mca/rtc/hwloc
> ./orte/mca/rtc/hwloc/rtc_hwloc.lo
> ./orte/mca/rtc/hwloc/.deps/rtc_hwloc.Plo
> ./orte/mca/rtc/hwloc/.deps/rtc_hwloc_component.Plo
> ./orte/mca/rtc/hwloc/mca_rtc_hwloc.la
> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.so
> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.la
> ./orte/mca/rtc/hwloc/.libs/rtc_hwloc.o
> ./orte/mca/rtc/hwloc/.libs/rtc_hwloc_component.o
> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.soT
> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.lai
> ./orte/mca/rtc/hwloc/rtc_hwloc_component.lo
> loki openmpi-master-201809290304-73075b8-Linux.x86_64.64_gcc 112
> 
> And some files are available in the install directory.
> 
> loki openmpi-master_64_gcc 116 find . -name '*hwloc*'
> ./share/openmpi/help-orte-rtc-hwloc.txt
> ./share/openmpi/help-opal-hwloc-base.txt
> ./lib64/openmpi/mca_rtc_hwloc.so
> ./lib64/openmpi/mca_rtc_hwloc.la
> loki openmpi-master_64_gcc 117
> 
> I don't see any unavailable libraries so that the only available
> hwloc library should work.
> 
> loki openmpi 126 ldd -v mca_rtc_hwloc.so
>linux-vdso.so.1 (0x7ffd2df5b000)
>libopen-rte.so.0 => 
> /usr/local/openmpi-master_64_gcc/lib64/libopen-rte.so.0 (0x7f082b7fb000)
>libopen-pal.so.0 => 
> /usr/local/openmpi-master_64_gcc/lib64/libopen-pal.so.0 (0x7f082b493000)
>libdl.so.2 => /lib64/libdl.so.2 (0x7f082b28f000)
>libudev.so.1 => /usr/lib64/libudev.so.1 (0x7f082b06e000)
>libpciaccess.so.0 => /usr/lib64/libpciaccess.so.0 (0x7f082ae64000)
>librt.so.1 => /lib64/librt.so.1 (0x7f082ac5c000)
>libm.so.6 => /lib64/libm.so.6 (0x7f082a95f000)
>libutil.so.1 => /lib64/libutil.so.1 (0x7f082a75c000)
>libz.so.1 => /lib64/libz.so.1 (0x7f082a546000)
>libpthread.so.0 => /lib64/libpthread.so.0 (0x7f082a329000)
>libc.so.6 => /lib64/libc.so.6 (0x7f0829f84000)
>libgcc_s.so.1 => /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1 
> (0x7f0829d6c000)
>/lib64/ld-linux-x86-64.so.2 (0x7f082bd24000)
>libselinux.so.1 => /lib64/libselinux.so.1 (0x7f0829b46000)
>libcap.so.2 => /lib64/libcap.so.2 (0x7f0829941000)
>libresolv.so.2 => /lib64/libresolv.so.2 (0x7f082972a000)
>libpcre.so.1 => /usr/lib64/libpcre.so.1 (0x7f08294bb000)
> 
>Version information:
>./mca_rtc_hwloc.so:
>libpthread.so.0 (GLIBC_2.2.5) => /lib64/libpthread.so.0
>libc.so.6 (GLIBC_2.2.5) => /lib64/libc.so.6
>/usr/local/openmpi-master_64_gcc/lib64/libopen-rte.so.0:
>   

Re: [OMPI users] opal_pmix_base_select failed for master and 4.0.0

2018-10-02 Thread Jeff Squyres (jsquyres) via users
(Ralph sent me Siegmar's pmix config.log, which Siegmar sent to him off-list)

It looks like Siegmar passed --with-hwloc=internal.

Open MPI's configure understood this and did the appropriate things.
PMIX's configure didn't.

I think we need to add an adjustment into the PMIx configure.m4 in OMPI...


> On Oct 2, 2018, at 5:25 PM, Ralph H Castain  wrote:
> 
> Hi Siegmar
> 
> I honestly have no idea - for some reason, the PMIx component isn’t seeing 
> the internal hwloc code in your environment.
> 
> Jeff, Brice - any ideas?
> 
> 
>> On Oct 2, 2018, at 1:18 PM, Siegmar Gross 
>>  wrote:
>> 
>> Hi Ralph,
>> 
>> how can I confirm that HWLOC built? Some hwloc files are available
>> in the built directory.
>> 
>> loki openmpi-master-201809290304-73075b8-Linux.x86_64.64_gcc 111 find . 
>> -name '*hwloc*'
>> ./opal/mca/btl/usnic/.deps/btl_usnic_hwloc.Plo
>> ./opal/mca/hwloc
>> ./opal/mca/hwloc/external/.deps/hwloc_external_component.Plo
>> ./opal/mca/hwloc/base/hwloc_base_frame.lo
>> ./opal/mca/hwloc/base/.deps/hwloc_base_dt.Plo
>> ./opal/mca/hwloc/base/.deps/hwloc_base_maffinity.Plo
>> ./opal/mca/hwloc/base/.deps/hwloc_base_frame.Plo
>> ./opal/mca/hwloc/base/.deps/hwloc_base_util.Plo
>> ./opal/mca/hwloc/base/hwloc_base_dt.lo
>> ./opal/mca/hwloc/base/hwloc_base_util.lo
>> ./opal/mca/hwloc/base/hwloc_base_maffinity.lo
>> ./opal/mca/hwloc/base/.libs/hwloc_base_util.o
>> ./opal/mca/hwloc/base/.libs/hwloc_base_dt.o
>> ./opal/mca/hwloc/base/.libs/hwloc_base_maffinity.o
>> ./opal/mca/hwloc/base/.libs/hwloc_base_frame.o
>> ./opal/mca/hwloc/.libs/libmca_hwloc.la
>> ./opal/mca/hwloc/.libs/libmca_hwloc.a
>> ./opal/mca/hwloc/libmca_hwloc.la
>> ./opal/mca/hwloc/hwloc201
>> ./opal/mca/hwloc/hwloc201/.deps/hwloc201_component.Plo
>> ./opal/mca/hwloc/hwloc201/hwloc201_component.lo
>> ./opal/mca/hwloc/hwloc201/hwloc
>> ./opal/mca/hwloc/hwloc201/hwloc/include/hwloc
>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc
>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/libhwloc_embedded.la
>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_pci_la-topology-pci.Plo
>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_gl_la-topology-gl.Plo
>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_cuda_la-topology-cuda.Plo
>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_xml_libxml_la-topology-xml-libxml.Plo
>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_opencl_la-topology-opencl.Plo
>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.deps/hwloc_nvml_la-topology-nvml.Plo
>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.libs/libhwloc_embedded.la
>> ./opal/mca/hwloc/hwloc201/hwloc/hwloc/.libs/libhwloc_embedded.a
>> ./opal/mca/hwloc/hwloc201/.libs/hwloc201_component.o
>> ./opal/mca/hwloc/hwloc201/.libs/libmca_hwloc_hwloc201.la
>> ./opal/mca/hwloc/hwloc201/.libs/libmca_hwloc_hwloc201.a
>> ./opal/mca/hwloc/hwloc201/libmca_hwloc_hwloc201.la
>> ./orte/mca/rtc/hwloc
>> ./orte/mca/rtc/hwloc/rtc_hwloc.lo
>> ./orte/mca/rtc/hwloc/.deps/rtc_hwloc.Plo
>> ./orte/mca/rtc/hwloc/.deps/rtc_hwloc_component.Plo
>> ./orte/mca/rtc/hwloc/mca_rtc_hwloc.la
>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.so
>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.la
>> ./orte/mca/rtc/hwloc/.libs/rtc_hwloc.o
>> ./orte/mca/rtc/hwloc/.libs/rtc_hwloc_component.o
>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.soT
>> ./orte/mca/rtc/hwloc/.libs/mca_rtc_hwloc.lai
>> ./orte/mca/rtc/hwloc/rtc_hwloc_component.lo
>> loki openmpi-master-201809290304-73075b8-Linux.x86_64.64_gcc 112
>> 
>> And some files are available in the install directory.
>> 
>> loki openmpi-master_64_gcc 116 find . -name '*hwloc*'
>> ./share/openmpi/help-orte-rtc-hwloc.txt
>> ./share/openmpi/help-opal-hwloc-base.txt
>> ./lib64/openmpi/mca_rtc_hwloc.so
>> ./lib64/openmpi/mca_rtc_hwloc.la
>> loki openmpi-master_64_gcc 117
>> 
>> I don't see any unavailable libraries so that the only available
>> hwloc library should work.
>> 
>> loki openmpi 126 ldd -v mca_rtc_hwloc.so
>>   linux-vdso.so.1 (0x7ffd2df5b000)
>>   libopen-rte.so.0 => 
>> /usr/local/openmpi-master_64_gcc/lib64/libopen-rte.so.0 (0x7f082b7fb000)
>>   libopen-pal.so.0 => 
>> /usr/local/openmpi-master_64_gcc/lib64/libopen-pal.so.0 (0x7f082b493000)
>>   libdl.so.2 => /lib64/libdl.so.2 (0x7f082b28f000)
>>   libudev.so.1 => /usr/lib64/libudev.so.1 (0x7f082b06e000)
>>   libpciaccess.so.0 => /usr/lib64/libpciaccess.so.0 (0x7f082ae64000)
>>   librt.so.1 => /lib64/librt.so.1 (0x7f082ac5c000)
>>   libm.so.6 => /lib64/libm.so.6 (0x7f082a95f000)
>>   libutil.so.1 => /lib64/libutil.so.1 (0x7f082a75c000)
>>   libz.so.1 => /lib64/libz.so.1 (0x7f082a546000)
>>   libpthread.so.0 => /lib64/libpthread.so.0 (0x7f082a329000)
>>   libc.so.6 => /lib64/libc.so.6 (0x7f0829f84000)
>>   libgcc_s.so.1 => /usr/local/gcc-8.2.0/lib64/libgcc_s.so.1 
>> (0x7f0829d6c000)
>>   /lib64/ld-linux-x86-64.so.2 (0x7f082bd24000)
>>   libselinux.so.1 => /lib64/libselinux.so.1 (0x7f0829b46000)
>>   libcap.so.2 => /lib