Thank you for all the pointers. I was able to
build openmpi-v2.x-dev-96-g918650a without problems on Edison, and also on
other systems.

I'm circumventing the OS X warning by ignoring it via "grep -v"; the other
suggestion (--mca oob ^usock) did not work for me. I've
tried openmpi-v2.x-dev-100-g26c3f03, but it still leads to the same warning.

-erik


On Mon, Jul 27, 2015 at 10:17 AM, Gilles Gouaillardet <
gilles.gouaillar...@gmail.com> wrote:

> Eric,
>
> these warnings are not important and you can simply ignore them.
> fwiw, this is a race condition evidenced by recent "asynchrousity".
>
> i will push a fix tomorrow.
>
> in the mean time, you can
> mpirun --mca oob ^tcp ...
> (if you run on one node only)
> or
> mpirun --mca oob ^usock
> (if you have an OS X cluster ...)
>
> Cheers,
>
> Gilles
>
> On Sunday, July 26, 2015, Erik Schnetter <schnet...@gmail.com> wrote:
>
>> Mark
>>
>> No, it doesn't need to be 1.8.7.
>>
>> I just tried v2.x-dev-96-g918650a. This leads to run-time warnings on OS
>> X; I see messages such as
>>
>> [warn] select: Bad file descriptor
>>
>> Are these important? If not, how can I suppress them?
>>
>> -erik
>>
>>
>> On Sat, Jul 25, 2015 at 7:49 AM, Mark Santcroos <
>> mark.santcr...@rutgers.edu> wrote:
>>
>>> Hi Erik,
>>>
>>> Do you really want 1.8.7, otherwise you might want to give latest master
>>> a try. Other including myself had more luck with that on Cray's, including
>>> Edison.
>>>
>>> Mark
>>>
>>> > On 25 Jul 2015, at 1:35 , Erik Schnetter <schnet...@gmail.com> wrote:
>>> >
>>> > I want to build OpenMPI 1.8.7 on a Cray XC30 (Edison at NERSC). I've
>>> tried various configuration options, but I am always encountering either
>>> OpenMPI build errors, application build errors, or run-time errors.
>>> >
>>> > I'm currently looking at <
>>> http://www.open-mpi.org/community/lists/users/2015/06/27230.php>, which
>>> seems to describe my case. I'm now configuring OpenMPI without any options,
>>> except setting compilers to clang/gfortran and pointing it to a self-built
>>> hwloc. For completeness, here are my configure options as recorded by
>>> config.status:
>>> >
>>> >
>>> '/project/projectdirs/m152/schnette/edison/software/src/openmpi-1.8.7/src/openmpi-1.8.7/configure'
>>> '--prefix=/project/projectdirs/m152/schnette/edison/software/openmpi-1.8.7'
>>> '--with-hwloc=/project/projectdirs/m152/schnette/edison/software/hwloc-1.11.0'
>>> '--disable-vt'
>>> 'CC=/project/projectdirs/m152/schnette/edison/software/llvm-3.6.2/bin/wrap-clang'
>>> 'CXX=/project/projectdirs/m152/schnette/edison/software/llvm-3.6.2/bin/wrap-clang++'
>>> 'FC=/project/projectdirs/m152/schnette/edison/software/gcc-5.2.0/bin/wrap-gfortran'
>>> 'CFLAGS=-I/opt/ofed/include
>>> -I/project/projectdirs/m152/schnette/edison/software/hwloc-1.11.0/include'
>>> 'CXXFLAGS=-I/opt/ofed/include
>>> -I/project/projectdirs/m152/schnette/edison/software/hwloc-1.11.0/include'
>>> 'LDFLAGS=-L/opt/ofed/lib64
>>> -L/project/projectdirs/m152/schnette/edison/software/hwloc-1.11.0/lib
>>> -Wl,-rpath,/project/projectdirs/m152/schnette/edison/software/hwloc-1.11.0/lib'
>>> 'LIBS=-lhwloc -lpthread -lpthread'
>>> '--with-wrapper-ldflags=-L/project/projectdirs/
>>>  m152/schnette/edison/software/hwloc-1.11.0/lib
>>> -Wl,-rpath,/project/projectdirs/m152/schnette/edison/software/hwloc-1.11.0/lib'
>>> '--with-wrapper-libs=-lhwloc -lpthread'
>>> >
>>> > This builds and installs fine, and works when running on a single
>>> node. However, multi-node runs are stalling: The queue starts the job, but
>>> mpirun produces no output. The "-v" option to mpirun doesn't help.
>>> >
>>> > When I use aprun instead of mpirun to start my application, then all
>>> processes think they are rank 0.
>>> >
>>> > Do you have any pointers for how to debug this?
>>> >
>>> > -erik
>>> >
>>> > --
>>> > Erik Schnetter <schnet...@gmail.com>
>>> http://www.perimeterinstitute.ca/personal/eschnetter/
>>> > _______________________________________________
>>> > users mailing list
>>> > us...@open-mpi.org
>>> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> > Link to this post:
>>> http://www.open-mpi.org/community/lists/users/2015/07/27324.php
>>>
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> Link to this post:
>>> http://www.open-mpi.org/community/lists/users/2015/07/27327.php
>>>
>>
>>
>>
>> --
>> Erik Schnetter <schnet...@gmail.com>
>> http://www.perimeterinstitute.ca/personal/eschnetter/
>>
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post:
> http://www.open-mpi.org/community/lists/users/2015/07/27334.php
>



-- 
Erik Schnetter <schnet...@gmail.com>
http://www.perimeterinstitute.ca/personal/eschnetter/

Reply via email to