Re: [OMPI users] any deadlocks in this sets of MPI_send and MPI_recv ?
On Sep 14, 2013, at 12:21 PM, Huangwei wrote: > do i=1, N-1 > allocate (QRECS(A(i))) > itag = i > call mpi_recv(QRECS..., i, itag, .) > deallocate(QRECS) > enddo I assume you're doing more than just allocate/mpi_recv/deallocate. :-) I don't know what A(i) is, so it's hard to say whether this is correct or not. > Will this cause any problem using this way to introducing QRECS? I don't know. Can you make a small program that reproduces the problem? > Besides, the reasonable why I did not choose mpi_gatherv is that the QRECS > will put into YVAR in a non-consecutive way. for instance, if I have 4 > processors, the first element in YVAR is from rank 0, second from rank 1 > ..fourth from rank 3, and then fifth from rank 0 again, sixth from rank 1 > again... But I will try your suggestion. Depending on how big the arrays are, it may still be faster/simpler to gather(v) to a single array and copy to the final array. Shrug. It depends on what your program is doing, and whether that is wort it to you. To be clear: either way should work fine; there may or may not be coding and/or performance implications of each. -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/
Re: [OMPI users] any deadlocks in this sets of MPI_send and MPI_recv ?
Thanks very much, have a nice weekend! best regards, Huangwei On 15 September 2013 11:29, Jeff Squyres (jsquyres) wrote: > On Sep 14, 2013, at 12:21 PM, Huangwei wrote: > > > do i=1, N-1 > > allocate (QRECS(A(i))) > > itag = i > > call mpi_recv(QRECS..., i, itag, .) > > deallocate(QRECS) > > enddo > > I assume you're doing more than just allocate/mpi_recv/deallocate. :-) > > I don't know what A(i) is, so it's hard to say whether this is correct or > not. > > > Will this cause any problem using this way to introducing QRECS? > > I don't know. > > Can you make a small program that reproduces the problem? > > > Besides, the reasonable why I did not choose mpi_gatherv is that the > QRECS will put into YVAR in a non-consecutive way. for instance, if I have > 4 processors, the first element in YVAR is from rank 0, second from rank 1 > ..fourth from rank 3, and then fifth from rank 0 again, sixth from rank > 1 again... But I will try your suggestion. > > Depending on how big the arrays are, it may still be faster/simpler to > gather(v) to a single array and copy to the final array. Shrug. It > depends on what your program is doing, and whether that is wort it to you. > > To be clear: either way should work fine; there may or may not be coding > and/or performance implications of each. > > -- > Jeff Squyres > jsquy...@cisco.com > For corporate legal information go to: > http://www.cisco.com/web/about/doing_business/legal/cri/ > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users >
Re: [OMPI users] Segmentation fault with fresh compilation of 1.7.2
I'm not entirely sure - I don't see anything that would cause that problem in that location. How did you configure this? On Sep 12, 2013, at 3:17 AM, Matthieu Brucher wrote: > Hi, > > I compiled OpenMPI on a RHEL6 box with LSF support, but when I run > sonthing, it crashes. Also orte-info crashes: > > Package: Open MPI mbruc...@xxx.com Distribution >Open RTE: 1.7.2 > Open RTE repo revision: r28673 > Open RTE release date: Jun 26, 2013 >OPAL: 1.7.2 > OPAL repo revision: r28673 > OPAL release date: Jun 26, 2013 >Ident string: 1.7.2 > Prefix: /xxx/mbrucher/openmpi > Configured architecture: x86_64-unknown-linux-gnu > Configure host: xxx.xxx.com > Configured by: mbrucher > Configured on: Thu Sep 12 10:22:06 BST 2013 > Configure host: xxx.xxx.com >Built by: mbrucher >Built on: Thu Sep 12 10:24:59 BST 2013 > Built host: xxx.xxx.com > C compiler: gcc > C compiler absolute: /usr/bin/gcc > C compiler family name: GNU > C compiler version: 4.4.6 > Internal debug support: no > Memory profiling support: no > Memory debugging support: no > libltdl support: yes > Heterogeneous support: no > orterun default --prefix: no > MPI_WTIME support: gettimeofday > Symbol vis. support: yes > FT Checkpoint support: no (checkpoint thread: no) > [abgengcluster:45509] *** Process received signal *** > [abgengcluster:45509] Signal: Segmentation fault (11) > [abgengcluster:45509] Signal code: Address not mapped (1) > [abgengcluster:45509] Failing at address: 0xf8 > [abgengcluster:45509] [ 0] /lib64/libpthread.so.0() [0x3ffc00f4a0] > [abgengcluster:45509] [ 1] > /xxx/mbrucher/openmpi/lib/libopen-pal.so.5(opal_libevent2019_event_priority_set+0x6f) > [0x2aae84a736ef] > [abgengcluster:45509] [ 2] > /xxx/mbrucher/openmpi/lib/libopen-rte.so.5(orte_iof_base_open+0x31c) > [0x2aae847edfbc] > [abgengcluster:45509] [ 3] orte-info(orte_info_open_components+0x71f) > [0x406b8f] > [abgengcluster:45509] [ 4] orte-info(main+0x93d) [0x40450d] > [abgengcluster:45509] [ 5] /lib64/libc.so.6(__libc_start_main+0xfd) > [0x3ffb81ecdd] > [abgengcluster:45509] [ 6] orte-info() [0x403b09] > [abgengcluster:45509] *** End of error message *** > Segmentation fault (core dumped) > > Is there something that I missed? > > Cheers, > > Matthieu > -- > Information System Engineer, Ph.D. > Blog: http://matt.eifelle.com > LinkedIn: http://www.linkedin.com/in/matthieubrucher > Music band: http://liliejay.com/ > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users
Re: [OMPI users] Segmentation fault with fresh compilation of 1.7.2
Just --with-lsf. Perhaps because then it must be launched through lsf? Le 15 sept. 2013 18:02, "Ralph Castain" a écrit : > I'm not entirely sure - I don't see anything that would cause that problem > in that location. How did you configure this? > > > On Sep 12, 2013, at 3:17 AM, Matthieu Brucher > wrote: > > > Hi, > > > > I compiled OpenMPI on a RHEL6 box with LSF support, but when I run > > sonthing, it crashes. Also orte-info crashes: > > > > Package: Open MPI mbruc...@xxx.com Distribution > >Open RTE: 1.7.2 > > Open RTE repo revision: r28673 > > Open RTE release date: Jun 26, 2013 > >OPAL: 1.7.2 > > OPAL repo revision: r28673 > > OPAL release date: Jun 26, 2013 > >Ident string: 1.7.2 > > Prefix: /xxx/mbrucher/openmpi > > Configured architecture: x86_64-unknown-linux-gnu > > Configure host: xxx.xxx.com > > Configured by: mbrucher > > Configured on: Thu Sep 12 10:22:06 BST 2013 > > Configure host: xxx.xxx.com > >Built by: mbrucher > >Built on: Thu Sep 12 10:24:59 BST 2013 > > Built host: xxx.xxx.com > > C compiler: gcc > > C compiler absolute: /usr/bin/gcc > > C compiler family name: GNU > > C compiler version: 4.4.6 > > Internal debug support: no > > Memory profiling support: no > > Memory debugging support: no > > libltdl support: yes > > Heterogeneous support: no > > orterun default --prefix: no > > MPI_WTIME support: gettimeofday > > Symbol vis. support: yes > > FT Checkpoint support: no (checkpoint thread: no) > > [abgengcluster:45509] *** Process received signal *** > > [abgengcluster:45509] Signal: Segmentation fault (11) > > [abgengcluster:45509] Signal code: Address not mapped (1) > > [abgengcluster:45509] Failing at address: 0xf8 > > [abgengcluster:45509] [ 0] /lib64/libpthread.so.0() [0x3ffc00f4a0] > > [abgengcluster:45509] [ 1] > > > /xxx/mbrucher/openmpi/lib/libopen-pal.so.5(opal_libevent2019_event_priority_set+0x6f) > > [0x2aae84a736ef] > > [abgengcluster:45509] [ 2] > > /xxx/mbrucher/openmpi/lib/libopen-rte.so.5(orte_iof_base_open+0x31c) > > [0x2aae847edfbc] > > [abgengcluster:45509] [ 3] orte-info(orte_info_open_components+0x71f) > [0x406b8f] > > [abgengcluster:45509] [ 4] orte-info(main+0x93d) [0x40450d] > > [abgengcluster:45509] [ 5] /lib64/libc.so.6(__libc_start_main+0xfd) > > [0x3ffb81ecdd] > > [abgengcluster:45509] [ 6] orte-info() [0x403b09] > > [abgengcluster:45509] *** End of error message *** > > Segmentation fault (core dumped) > > > > Is there something that I missed? > > > > Cheers, > > > > Matthieu > > -- > > Information System Engineer, Ph.D. > > Blog: http://matt.eifelle.com > > LinkedIn: http://www.linkedin.com/in/matthieubrucher > > Music band: http://liliejay.com/ > > ___ > > users mailing list > > us...@open-mpi.org > > http://www.open-mpi.org/mailman/listinfo.cgi/users > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users >
Re: [OMPI users] Segmentation fault with fresh compilation of 1.7.2
No - out of curiosity, does ompi_info work? I'm wondering if this is strictly an orte-info problem. On Sep 15, 2013, at 10:03 AM, Matthieu Brucher wrote: > Just --with-lsf. Perhaps because then it must be launched through lsf? > > Le 15 sept. 2013 18:02, "Ralph Castain" a écrit : > I'm not entirely sure - I don't see anything that would cause that problem in > that location. How did you configure this? > > > On Sep 12, 2013, at 3:17 AM, Matthieu Brucher > wrote: > > > Hi, > > > > I compiled OpenMPI on a RHEL6 box with LSF support, but when I run > > sonthing, it crashes. Also orte-info crashes: > > > > Package: Open MPI mbruc...@xxx.com Distribution > >Open RTE: 1.7.2 > > Open RTE repo revision: r28673 > > Open RTE release date: Jun 26, 2013 > >OPAL: 1.7.2 > > OPAL repo revision: r28673 > > OPAL release date: Jun 26, 2013 > >Ident string: 1.7.2 > > Prefix: /xxx/mbrucher/openmpi > > Configured architecture: x86_64-unknown-linux-gnu > > Configure host: xxx.xxx.com > > Configured by: mbrucher > > Configured on: Thu Sep 12 10:22:06 BST 2013 > > Configure host: xxx.xxx.com > >Built by: mbrucher > >Built on: Thu Sep 12 10:24:59 BST 2013 > > Built host: xxx.xxx.com > > C compiler: gcc > > C compiler absolute: /usr/bin/gcc > > C compiler family name: GNU > > C compiler version: 4.4.6 > > Internal debug support: no > > Memory profiling support: no > > Memory debugging support: no > > libltdl support: yes > > Heterogeneous support: no > > orterun default --prefix: no > > MPI_WTIME support: gettimeofday > > Symbol vis. support: yes > > FT Checkpoint support: no (checkpoint thread: no) > > [abgengcluster:45509] *** Process received signal *** > > [abgengcluster:45509] Signal: Segmentation fault (11) > > [abgengcluster:45509] Signal code: Address not mapped (1) > > [abgengcluster:45509] Failing at address: 0xf8 > > [abgengcluster:45509] [ 0] /lib64/libpthread.so.0() [0x3ffc00f4a0] > > [abgengcluster:45509] [ 1] > > /xxx/mbrucher/openmpi/lib/libopen-pal.so.5(opal_libevent2019_event_priority_set+0x6f) > > [0x2aae84a736ef] > > [abgengcluster:45509] [ 2] > > /xxx/mbrucher/openmpi/lib/libopen-rte.so.5(orte_iof_base_open+0x31c) > > [0x2aae847edfbc] > > [abgengcluster:45509] [ 3] orte-info(orte_info_open_components+0x71f) > > [0x406b8f] > > [abgengcluster:45509] [ 4] orte-info(main+0x93d) [0x40450d] > > [abgengcluster:45509] [ 5] /lib64/libc.so.6(__libc_start_main+0xfd) > > [0x3ffb81ecdd] > > [abgengcluster:45509] [ 6] orte-info() [0x403b09] > > [abgengcluster:45509] *** End of error message *** > > Segmentation fault (core dumped) > > > > Is there something that I missed? > > > > Cheers, > > > > Matthieu > > -- > > Information System Engineer, Ph.D. > > Blog: http://matt.eifelle.com > > LinkedIn: http://www.linkedin.com/in/matthieubrucher > > Music band: http://liliejay.com/ > > ___ > > users mailing list > > us...@open-mpi.org > > http://www.open-mpi.org/mailman/listinfo.cgi/users > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users
Re: [OMPI users] Segmentation fault with fresh compilation of 1.7.2
Yes, ompi_info does not crash. Le 15 sept. 2013 18:05, "Ralph Castain" a écrit : > No - out of curiosity, does ompi_info work? I'm wondering if this is > strictly an orte-info problem. > > On Sep 15, 2013, at 10:03 AM, Matthieu Brucher > wrote: > > Just --with-lsf. Perhaps because then it must be launched through lsf? > Le 15 sept. 2013 18:02, "Ralph Castain" a écrit : > >> I'm not entirely sure - I don't see anything that would cause that >> problem in that location. How did you configure this? >> >> >> On Sep 12, 2013, at 3:17 AM, Matthieu Brucher >> wrote: >> >> > Hi, >> > >> > I compiled OpenMPI on a RHEL6 box with LSF support, but when I run >> > sonthing, it crashes. Also orte-info crashes: >> > >> > Package: Open MPI mbruc...@xxx.com Distribution >> >Open RTE: 1.7.2 >> > Open RTE repo revision: r28673 >> > Open RTE release date: Jun 26, 2013 >> >OPAL: 1.7.2 >> > OPAL repo revision: r28673 >> > OPAL release date: Jun 26, 2013 >> >Ident string: 1.7.2 >> > Prefix: /xxx/mbrucher/openmpi >> > Configured architecture: x86_64-unknown-linux-gnu >> > Configure host: xxx.xxx.com >> > Configured by: mbrucher >> > Configured on: Thu Sep 12 10:22:06 BST 2013 >> > Configure host: xxx.xxx.com >> >Built by: mbrucher >> >Built on: Thu Sep 12 10:24:59 BST 2013 >> > Built host: xxx.xxx.com >> > C compiler: gcc >> > C compiler absolute: /usr/bin/gcc >> > C compiler family name: GNU >> > C compiler version: 4.4.6 >> > Internal debug support: no >> > Memory profiling support: no >> > Memory debugging support: no >> > libltdl support: yes >> > Heterogeneous support: no >> > orterun default --prefix: no >> > MPI_WTIME support: gettimeofday >> > Symbol vis. support: yes >> > FT Checkpoint support: no (checkpoint thread: no) >> > [abgengcluster:45509] *** Process received signal *** >> > [abgengcluster:45509] Signal: Segmentation fault (11) >> > [abgengcluster:45509] Signal code: Address not mapped (1) >> > [abgengcluster:45509] Failing at address: 0xf8 >> > [abgengcluster:45509] [ 0] /lib64/libpthread.so.0() [0x3ffc00f4a0] >> > [abgengcluster:45509] [ 1] >> > >> /xxx/mbrucher/openmpi/lib/libopen-pal.so.5(opal_libevent2019_event_priority_set+0x6f) >> > [0x2aae84a736ef] >> > [abgengcluster:45509] [ 2] >> > /xxx/mbrucher/openmpi/lib/libopen-rte.so.5(orte_iof_base_open+0x31c) >> > [0x2aae847edfbc] >> > [abgengcluster:45509] [ 3] orte-info(orte_info_open_components+0x71f) >> [0x406b8f] >> > [abgengcluster:45509] [ 4] orte-info(main+0x93d) [0x40450d] >> > [abgengcluster:45509] [ 5] /lib64/libc.so.6(__libc_start_main+0xfd) >> > [0x3ffb81ecdd] >> > [abgengcluster:45509] [ 6] orte-info() [0x403b09] >> > [abgengcluster:45509] *** End of error message *** >> > Segmentation fault (core dumped) >> > >> > Is there something that I missed? >> > >> > Cheers, >> > >> > Matthieu >> > -- >> > Information System Engineer, Ph.D. >> > Blog: http://matt.eifelle.com >> > LinkedIn: http://www.linkedin.com/in/matthieubrucher >> > Music band: http://liliejay.com/ >> > ___ >> > users mailing list >> > us...@open-mpi.org >> > http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> ___ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users >
Re: [OMPI users] Segmentation fault with fresh compilation of 1.7.2
Could you try the current 1.7.3 nightly tarball instead? I don't see a problem there, and I'm wondering if this is something we already fixed. We will be releasing 1.7.3 shortly and it is mostly complete at this time. On Sep 15, 2013, at 10:43 AM, Matthieu Brucher wrote: > Yes, ompi_info does not crash. > > Le 15 sept. 2013 18:05, "Ralph Castain" a écrit : > No - out of curiosity, does ompi_info work? I'm wondering if this is strictly > an orte-info problem. > > On Sep 15, 2013, at 10:03 AM, Matthieu Brucher > wrote: > >> Just --with-lsf. Perhaps because then it must be launched through lsf? >> >> Le 15 sept. 2013 18:02, "Ralph Castain" a écrit : >> I'm not entirely sure - I don't see anything that would cause that problem >> in that location. How did you configure this? >> >> >> On Sep 12, 2013, at 3:17 AM, Matthieu Brucher >> wrote: >> >> > Hi, >> > >> > I compiled OpenMPI on a RHEL6 box with LSF support, but when I run >> > sonthing, it crashes. Also orte-info crashes: >> > >> > Package: Open MPI mbruc...@xxx.com Distribution >> >Open RTE: 1.7.2 >> > Open RTE repo revision: r28673 >> > Open RTE release date: Jun 26, 2013 >> >OPAL: 1.7.2 >> > OPAL repo revision: r28673 >> > OPAL release date: Jun 26, 2013 >> >Ident string: 1.7.2 >> > Prefix: /xxx/mbrucher/openmpi >> > Configured architecture: x86_64-unknown-linux-gnu >> > Configure host: xxx.xxx.com >> > Configured by: mbrucher >> > Configured on: Thu Sep 12 10:22:06 BST 2013 >> > Configure host: xxx.xxx.com >> >Built by: mbrucher >> >Built on: Thu Sep 12 10:24:59 BST 2013 >> > Built host: xxx.xxx.com >> > C compiler: gcc >> > C compiler absolute: /usr/bin/gcc >> > C compiler family name: GNU >> > C compiler version: 4.4.6 >> > Internal debug support: no >> > Memory profiling support: no >> > Memory debugging support: no >> > libltdl support: yes >> > Heterogeneous support: no >> > orterun default --prefix: no >> > MPI_WTIME support: gettimeofday >> > Symbol vis. support: yes >> > FT Checkpoint support: no (checkpoint thread: no) >> > [abgengcluster:45509] *** Process received signal *** >> > [abgengcluster:45509] Signal: Segmentation fault (11) >> > [abgengcluster:45509] Signal code: Address not mapped (1) >> > [abgengcluster:45509] Failing at address: 0xf8 >> > [abgengcluster:45509] [ 0] /lib64/libpthread.so.0() [0x3ffc00f4a0] >> > [abgengcluster:45509] [ 1] >> > /xxx/mbrucher/openmpi/lib/libopen-pal.so.5(opal_libevent2019_event_priority_set+0x6f) >> > [0x2aae84a736ef] >> > [abgengcluster:45509] [ 2] >> > /xxx/mbrucher/openmpi/lib/libopen-rte.so.5(orte_iof_base_open+0x31c) >> > [0x2aae847edfbc] >> > [abgengcluster:45509] [ 3] orte-info(orte_info_open_components+0x71f) >> > [0x406b8f] >> > [abgengcluster:45509] [ 4] orte-info(main+0x93d) [0x40450d] >> > [abgengcluster:45509] [ 5] /lib64/libc.so.6(__libc_start_main+0xfd) >> > [0x3ffb81ecdd] >> > [abgengcluster:45509] [ 6] orte-info() [0x403b09] >> > [abgengcluster:45509] *** End of error message *** >> > Segmentation fault (core dumped) >> > >> > Is there something that I missed? >> > >> > Cheers, >> > >> > Matthieu >> > -- >> > Information System Engineer, Ph.D. >> > Blog: http://matt.eifelle.com >> > LinkedIn: http://www.linkedin.com/in/matthieubrucher >> > Music band: http://liliejay.com/ >> > ___ >> > users mailing list >> > us...@open-mpi.org >> > http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> ___ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> ___ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users
Re: [OMPI users] Segmentation fault with fresh compilation of 1.7.2
I can try later this week, yes. Thanks Le 15 sept. 2013 19:09, "Ralph Castain" a écrit : > Could you try the current 1.7.3 nightly tarball instead? I don't see a > problem there, and I'm wondering if this is something we already fixed. We > will be releasing 1.7.3 shortly and it is mostly complete at this time. > > > On Sep 15, 2013, at 10:43 AM, Matthieu Brucher > wrote: > > Yes, ompi_info does not crash. > Le 15 sept. 2013 18:05, "Ralph Castain" a écrit : > >> No - out of curiosity, does ompi_info work? I'm wondering if this is >> strictly an orte-info problem. >> >> On Sep 15, 2013, at 10:03 AM, Matthieu Brucher < >> matthieu.bruc...@gmail.com> wrote: >> >> Just --with-lsf. Perhaps because then it must be launched through lsf? >> Le 15 sept. 2013 18:02, "Ralph Castain" a écrit : >> >>> I'm not entirely sure - I don't see anything that would cause that >>> problem in that location. How did you configure this? >>> >>> >>> On Sep 12, 2013, at 3:17 AM, Matthieu Brucher < >>> matthieu.bruc...@gmail.com> wrote: >>> >>> > Hi, >>> > >>> > I compiled OpenMPI on a RHEL6 box with LSF support, but when I run >>> > sonthing, it crashes. Also orte-info crashes: >>> > >>> > Package: Open MPI mbruc...@xxx.com Distribution >>> >Open RTE: 1.7.2 >>> > Open RTE repo revision: r28673 >>> > Open RTE release date: Jun 26, 2013 >>> >OPAL: 1.7.2 >>> > OPAL repo revision: r28673 >>> > OPAL release date: Jun 26, 2013 >>> >Ident string: 1.7.2 >>> > Prefix: /xxx/mbrucher/openmpi >>> > Configured architecture: x86_64-unknown-linux-gnu >>> > Configure host: xxx.xxx.com >>> > Configured by: mbrucher >>> > Configured on: Thu Sep 12 10:22:06 BST 2013 >>> > Configure host: xxx.xxx.com >>> >Built by: mbrucher >>> >Built on: Thu Sep 12 10:24:59 BST 2013 >>> > Built host: xxx.xxx.com >>> > C compiler: gcc >>> > C compiler absolute: /usr/bin/gcc >>> > C compiler family name: GNU >>> > C compiler version: 4.4.6 >>> > Internal debug support: no >>> > Memory profiling support: no >>> > Memory debugging support: no >>> > libltdl support: yes >>> > Heterogeneous support: no >>> > orterun default --prefix: no >>> > MPI_WTIME support: gettimeofday >>> > Symbol vis. support: yes >>> > FT Checkpoint support: no (checkpoint thread: no) >>> > [abgengcluster:45509] *** Process received signal *** >>> > [abgengcluster:45509] Signal: Segmentation fault (11) >>> > [abgengcluster:45509] Signal code: Address not mapped (1) >>> > [abgengcluster:45509] Failing at address: 0xf8 >>> > [abgengcluster:45509] [ 0] /lib64/libpthread.so.0() [0x3ffc00f4a0] >>> > [abgengcluster:45509] [ 1] >>> > >>> /xxx/mbrucher/openmpi/lib/libopen-pal.so.5(opal_libevent2019_event_priority_set+0x6f) >>> > [0x2aae84a736ef] >>> > [abgengcluster:45509] [ 2] >>> > /xxx/mbrucher/openmpi/lib/libopen-rte.so.5(orte_iof_base_open+0x31c) >>> > [0x2aae847edfbc] >>> > [abgengcluster:45509] [ 3] orte-info(orte_info_open_components+0x71f) >>> [0x406b8f] >>> > [abgengcluster:45509] [ 4] orte-info(main+0x93d) [0x40450d] >>> > [abgengcluster:45509] [ 5] /lib64/libc.so.6(__libc_start_main+0xfd) >>> > [0x3ffb81ecdd] >>> > [abgengcluster:45509] [ 6] orte-info() [0x403b09] >>> > [abgengcluster:45509] *** End of error message *** >>> > Segmentation fault (core dumped) >>> > >>> > Is there something that I missed? >>> > >>> > Cheers, >>> > >>> > Matthieu >>> > -- >>> > Information System Engineer, Ph.D. >>> > Blog: http://matt.eifelle.com >>> > LinkedIn: http://www.linkedin.com/in/matthieubrucher >>> > Music band: http://liliejay.com/ >>> > ___ >>> > users mailing list >>> > us...@open-mpi.org >>> > http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> ___ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >> ___ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> >> ___ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users >