Can you try the latest 1.8.2 rc tarball? (just released yesterday) http://www.open-mpi.org/software/ompi/v1.8/
On Aug 14, 2014, at 8:39 AM, Maxime Boissonneault <maxime.boissonnea...@calculquebec.ca> wrote: > Hi, > I compiled Charm++ 6.6.0rc3 using > ./build charm++ mpi-linux-x86_64 smp --with-production > > When compiling the simple example > mpi-linux-x86_64-smp/tests/charm++/simplearrayhello/ > > I get a segmentation fault that traces back to OpenMPI : > [mboisson@helios-login1 simplearrayhello]$ ./hello > [helios-login1:01813] *** Process received signal *** > [helios-login1:01813] Signal: Segmentation fault (11) > [helios-login1:01813] Signal code: Address not mapped (1) > [helios-login1:01813] Failing at address: 0x30 > [helios-login1:01813] [ 0] /lib64/libpthread.so.0[0x381c00f710] > [helios-login1:01813] [ 1] > /software-gpu/mpi/openmpi/1.8.1_gcc4.8_cuda6.0.37/lib/libmpi.so.1(+0xf78f8)[0x7f2cd1f6b8f8] > [helios-login1:01813] [ 2] > /software-gpu/mpi/openmpi/1.8.1_gcc4.8_cuda6.0.37/lib/libmpi.so.1(+0xf8f64)[0x7f2cd1f6cf64] > [helios-login1:01813] [ 3] > /software-gpu/mpi/openmpi/1.8.1_gcc4.8_cuda6.0.37/lib/libmpi.so.1(ompi_btl_openib_connect_base_select_for_local_port+0xcf)[0x7f2cd1f672af] > [helios-login1:01813] [ 4] > /software-gpu/mpi/openmpi/1.8.1_gcc4.8_cuda6.0.37/lib/libmpi.so.1(+0xe1ad7)[0x7f2cd1f55ad7] > [helios-login1:01813] [ 5] > /software-gpu/mpi/openmpi/1.8.1_gcc4.8_cuda6.0.37/lib/libmpi.so.1(mca_btl_base_select+0x168)[0x7f2cd1f4bf28] > [helios-login1:01813] [ 6] > /software-gpu/mpi/openmpi/1.8.1_gcc4.8_cuda6.0.37/lib/libmpi.so.1(mca_bml_r2_component_init+0x11)[0x7f2cd1f4b851] > [helios-login1:01813] [ 7] > /software-gpu/mpi/openmpi/1.8.1_gcc4.8_cuda6.0.37/lib/libmpi.so.1(mca_bml_base_init+0x7f)[0x7f2cd1f4a03f] > [helios-login1:01813] [ 8] > /software-gpu/mpi/openmpi/1.8.1_gcc4.8_cuda6.0.37/lib/libmpi.so.1(+0x1e0d17)[0x7f2cd2054d17] > [helios-login1:01813] [ 9] > /software-gpu/mpi/openmpi/1.8.1_gcc4.8_cuda6.0.37/lib/libmpi.so.1(mca_pml_base_select+0x3b6)[0x7f2cd20529d6] > [helios-login1:01813] [10] > /software-gpu/mpi/openmpi/1.8.1_gcc4.8_cuda6.0.37/lib/libmpi.so.1(ompi_mpi_init+0x4e4)[0x7f2cd1ef0c14] > [helios-login1:01813] [11] > /software-gpu/mpi/openmpi/1.8.1_gcc4.8_cuda6.0.37/lib/libmpi.so.1(MPI_Init_thread+0x15d)[0x7f2cd1f1065d] > [helios-login1:01813] [12] ./hello(LrtsInit+0x72)[0x4fcf02] > [helios-login1:01813] [13] ./hello(ConverseInit+0x70)[0x4ff680] > [helios-login1:01813] [14] ./hello(main+0x27)[0x470767] > [helios-login1:01813] [15] > /lib64/libc.so.6(__libc_start_main+0xfd)[0x381bc1ed1d] > [helios-login1:01813] [16] ./hello[0x470b71] > > > Anyone has a clue how to fix this ? > > Thanks, > > -- > --------------------------------- > Maxime Boissonneault > Analyste de calcul - Calcul Québec, Université Laval > Ph. D. en physique > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > http://www.open-mpi.org/community/lists/users/2014/08/25014.php -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/