This is really odd...

you can run
ompi_info --all
and search coll_ml_priority

it will display the current value and the origin
(e.g. default, system wide config, user config, cli, environment variable)

Cheers,

Gilles

On Thursday, June 18, 2015, Daniel Letai <d...@letai.org.il> wrote:

>  No, that's the issue.
> I had to disable it to get things working.
>
> That's why I included my config settings - I couldn't figure out which
> option enabled it, so I could remove it from the configuration...
>
> On 06/18/2015 02:43 PM, Gilles Gouaillardet wrote:
>
> Daniel,
>
>  ML module is not ready for production and is disabled by default.
>
>  Did you explicitly enable this module ?
> If yes, I encourage you to disable it
>
>  Cheers,
>
>  Gilles
>
> On Thursday, June 18, 2015, Daniel Letai <d...@letai.org.il
> <javascript:_e(%7B%7D,'cvml','d...@letai.org.il');>> wrote:
>
>> given a simple hello.c:
>>
>> #include <stdio.h>
>> #include <mpi.h>
>>
>> int main(int argc, char* argv[])
>> {
>>         int size, rank, len;
>>         char name[MPI_MAX_PROCESSOR_NAME];
>>
>>         MPI_Init(&argc, &argv);
>>         MPI_Comm_size(MPI_COMM_WORLD, &size);
>>         MPI_Comm_rank(MPI_COMM_WORLD, &rank);
>>         MPI_Get_processor_name(name, &len);
>>
>>         printf("%s: Process %d out of %d\n", name, rank, size);
>>
>>         MPI_Finalize();ffff
>> }
>>
>> for n=1
>> mpirun -n 1 ./hello
>> it works correctly.
>>
>> for n>1 it segfaults with signal 11
>> used gdb to trace the problem to ml coll:
>>
>> Program received signal SIGSEGV, Segmentation fault.
>> 0x00007ffff6750845 in ml_coll_hier_barrier_setup()
>>     from <path to openmpi 1.8.5>/lib/openmpi/mca_coll_ml.so
>>
>> running with
>> mpirun -n 2 --mca coll ^ml ./hello
>> works correctly
>>
>> using mellanox ofed 2.3-2.0.5-rhel6.4-x86_64, if it's at all relevant.
>> openmpi 1.8.5 was built with following options:
>> rpmbuild --rebuild --define 'configure_options --with-verbs=/usr
>> --with-verbs-libdir=/usr/lib64 CC=gcc CXX=g++ FC=gfortran CFLAGS="-g -O3"
>> --enable-mpirun-prefix-by-default --enable-orterun-prefix-by-default
>> --disable-debug --with-knem=/opt/knem-1.1.1.90mlnx
>> --with-platform=optimized --without-mpi-param-check
>> --with-contrib-vt-flags=--disable-iotrace --enable-builtin-atomics
>> --enable-cxx-exceptions --enable-sparse-groups --enable-mpi-thread-multiple
>> --enable-memchecker --enable-btl-openib-failover --with-hwloc=internal
>> --with-verbs --with-x --with-slurm --with-pmi=/opt/slurm
>> --with-fca=/opt/mellanox/fca --with-mxm=/opt/mellanox/mxm
>> --with-hcoll=/opt/mellanox/hcoll' openmpi-1.8.5-1.src.rpm
>>
>> gcc version 5.1.1
>>
>> Thanks in advance
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>> Link to this post:
>> http://www.open-mpi.org/community/lists/users/2015/06/27154.php
>>
>
>
> _______________________________________________
> users mailing listus...@open-mpi.org 
> <javascript:_e(%7B%7D,'cvml','us...@open-mpi.org');>
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2015/06/27155.php
>
>
>

Reply via email to