Erik,

the OS X warning (which should not be OS X specific) is fixed in https://github.com/open-mpi/ompi-release/pull/430
it will land into the v2.x series once reviewed
in the mean time, feel free to manually apply the patch on the tarball

Cheers,

Gilles

On 7/29/2015 10:35 AM, Erik Schnetter wrote:
Thank you for all the pointers. I was able to build openmpi-v2.x-dev-96-g918650a without problems on Edison, and also on other systems.

I'm circumventing the OS X warning by ignoring it via "grep -v"; the other suggestion (--mca oob ^usock) did not work for me. I've tried openmpi-v2.x-dev-100-g26c3f03, but it still leads to the same warning.

-erik


On Mon, Jul 27, 2015 at 10:17 AM, Gilles Gouaillardet <gilles.gouaillar...@gmail.com <mailto:gilles.gouaillar...@gmail.com>> wrote:

    Eric,

    these warnings are not important and you can simply ignore them.
    fwiw, this is a race condition evidenced by recent "asynchrousity".

    i will push a fix tomorrow.

    in the mean time, you can
    mpirun --mca oob ^tcp ...
    (if you run on one node only)
    or
    mpirun --mca oob ^usock
    (if you have an OS X cluster ...)

    Cheers,

    Gilles

    On Sunday, July 26, 2015, Erik Schnetter <schnet...@gmail.com
    <mailto:schnet...@gmail.com>> wrote:

        Mark

        No, it doesn't need to be 1.8.7.

        I just tried v2.x-dev-96-g918650a. This leads to run-time
        warnings on OS X; I see messages such as

        [warn] select: Bad file descriptor

        Are these important? If not, how can I suppress them?

        -erik


        On Sat, Jul 25, 2015 at 7:49 AM, Mark Santcroos
        <mark.santcr...@rutgers.edu> wrote:

            Hi Erik,

            Do you really want 1.8.7, otherwise you might want to give
            latest master a try. Other including myself had more luck
            with that on Cray's, including Edison.

            Mark

            > On 25 Jul 2015, at 1:35 , Erik Schnetter
            <schnet...@gmail.com> wrote:
            >
            > I want to build OpenMPI 1.8.7 on a Cray XC30 (Edison at
            NERSC). I've tried various configuration options, but I am
            always encountering either OpenMPI build errors,
            application build errors, or run-time errors.
            >
            > I'm currently looking at 
<http://www.open-mpi.org/community/lists/users/2015/06/27230.php>,
            which seems to describe my case. I'm now configuring
            OpenMPI without any options, except setting compilers to
            clang/gfortran and pointing it to a self-built hwloc. For
            completeness, here are my configure options as recorded by
            config.status:
            >
            >
            
'/project/projectdirs/m152/schnette/edison/software/src/openmpi-1.8.7/src/openmpi-1.8.7/configure'
            
'--prefix=/project/projectdirs/m152/schnette/edison/software/openmpi-1.8.7'
            
'--with-hwloc=/project/projectdirs/m152/schnette/edison/software/hwloc-1.11.0'
            '--disable-vt'
            
'CC=/project/projectdirs/m152/schnette/edison/software/llvm-3.6.2/bin/wrap-clang'
            
'CXX=/project/projectdirs/m152/schnette/edison/software/llvm-3.6.2/bin/wrap-clang++'
            
'FC=/project/projectdirs/m152/schnette/edison/software/gcc-5.2.0/bin/wrap-gfortran'
            'CFLAGS=-I/opt/ofed/include
            
-I/project/projectdirs/m152/schnette/edison/software/hwloc-1.11.0/include'
            'CXXFLAGS=-I/opt/ofed/include
            
-I/project/projectdirs/m152/schnette/edison/software/hwloc-1.11.0/include'
            'LDFLAGS=-L/opt/ofed/lib64
            
-L/project/projectdirs/m152/schnette/edison/software/hwloc-1.11.0/lib
            
-Wl,-rpath,/project/projectdirs/m152/schnette/edison/software/hwloc-1.11.0/lib'
            'LIBS=-lhwloc -lpthread -lpthread'
            '--with-wrapper-ldflags=-L/project/projectdirs/
             m152/schnette/edison/software/hwloc-1.11.0/lib
            
-Wl,-rpath,/project/projectdirs/m152/schnette/edison/software/hwloc-1.11.0/lib'
            '--with-wrapper-libs=-lhwloc -lpthread'
            >
            > This builds and installs fine, and works when running on
            a single node. However, multi-node runs are stalling: The
            queue starts the job, but mpirun produces no output. The
            "-v" option to mpirun doesn't help.
            >
            > When I use aprun instead of mpirun to start my
            application, then all processes think they are rank 0.
            >
            > Do you have any pointers for how to debug this?
            >
            > -erik
            >
            > --
            > Erik Schnetter <schnet...@gmail.com>
            http://www.perimeterinstitute.ca/personal/eschnetter/
            > _______________________________________________
            > users mailing list
            > us...@open-mpi.org
            > Subscription:
            http://www.open-mpi.org/mailman/listinfo.cgi/users
            > Link to this post:
            http://www.open-mpi.org/community/lists/users/2015/07/27324.php

            _______________________________________________
            users mailing list
            us...@open-mpi.org
            Subscription:
            http://www.open-mpi.org/mailman/listinfo.cgi/users
            Link to this post:
            http://www.open-mpi.org/community/lists/users/2015/07/27327.php




-- Erik Schnetter <schnet...@gmail.com>
        http://www.perimeterinstitute.ca/personal/eschnetter/


    _______________________________________________
    users mailing list
    us...@open-mpi.org <mailto:us...@open-mpi.org>
    Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
    Link to this post:
    http://www.open-mpi.org/community/lists/users/2015/07/27334.php




--
Erik Schnetter <schnet...@gmail.com <mailto:schnet...@gmail.com>> http://www.perimeterinstitute.ca/personal/eschnetter/


_______________________________________________
users mailing list
us...@open-mpi.org
Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
Link to this post: 
http://www.open-mpi.org/community/lists/users/2015/07/27353.php

Reply via email to