Re: [OMPI users] any deadlocks in this sets of MPI_send and MPI_recv ?

2013-09-15 Thread Jeff Squyres (jsquyres)
On Sep 14, 2013, at 12:21 PM, Huangwei  wrote:

>  do i=1, N-1
>   allocate (QRECS(A(i)))
>   itag = i
>  call mpi_recv(QRECS..., i, itag, .)
>   deallocate(QRECS)
>  enddo

I assume you're doing more than just allocate/mpi_recv/deallocate.  :-)

I don't know what A(i) is, so it's hard to say whether this is correct or not.

> Will this cause any problem using this way to introducing QRECS?

I don't know.  

Can you make a small program that reproduces the problem?

> Besides, the reasonable why I did not choose mpi_gatherv is that the QRECS 
> will put into YVAR in a non-consecutive way. for instance, if I have 4 
> processors, the first element in YVAR is from rank 0, second from rank 1 
> ..fourth from rank 3, and then fifth from rank 0 again, sixth from rank 1 
> again... But I will try your suggestion. 

Depending on how big the arrays are, it may still be faster/simpler to 
gather(v) to a single array and copy to the final array.  Shrug.  It depends on 
what your program is doing, and whether that is wort it to you.

To be clear: either way should work fine; there may or may not be coding and/or 
performance implications of each.

-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to: 
http://www.cisco.com/web/about/doing_business/legal/cri/



Re: [OMPI users] any deadlocks in this sets of MPI_send and MPI_recv ?

2013-09-15 Thread Huangwei
Thanks very much, have a nice weekend!

best regards,
Huangwei






On 15 September 2013 11:29, Jeff Squyres (jsquyres) wrote:

> On Sep 14, 2013, at 12:21 PM, Huangwei  wrote:
>
> >  do i=1, N-1
> >   allocate (QRECS(A(i)))
> >   itag = i
> >  call mpi_recv(QRECS..., i, itag, .)
> >   deallocate(QRECS)
> >  enddo
>
> I assume you're doing more than just allocate/mpi_recv/deallocate.  :-)
>
> I don't know what A(i) is, so it's hard to say whether this is correct or
> not.
>
> > Will this cause any problem using this way to introducing QRECS?
>
> I don't know.
>
> Can you make a small program that reproduces the problem?
>
> > Besides, the reasonable why I did not choose mpi_gatherv is that the
> QRECS will put into YVAR in a non-consecutive way. for instance, if I have
> 4 processors, the first element in YVAR is from rank 0, second from rank 1
> ..fourth from rank 3, and then fifth from rank 0 again, sixth from rank
> 1 again... But I will try your suggestion.
>
> Depending on how big the arrays are, it may still be faster/simpler to
> gather(v) to a single array and copy to the final array.  Shrug.  It
> depends on what your program is doing, and whether that is wort it to you.
>
> To be clear: either way should work fine; there may or may not be coding
> and/or performance implications of each.
>
> --
> Jeff Squyres
> jsquy...@cisco.com
> For corporate legal information go to:
> http://www.cisco.com/web/about/doing_business/legal/cri/
>
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>


Re: [OMPI users] Segmentation fault with fresh compilation of 1.7.2

2013-09-15 Thread Ralph Castain
I'm not entirely sure - I don't see anything that would cause that problem in 
that location. How did you configure this?


On Sep 12, 2013, at 3:17 AM, Matthieu Brucher  
wrote:

> Hi,
> 
> I compiled OpenMPI on a RHEL6 box with LSF support, but when I run
> sonthing, it crashes. Also orte-info crashes:
> 
> Package: Open MPI mbruc...@xxx.com Distribution
>Open RTE: 1.7.2
>  Open RTE repo revision: r28673
>   Open RTE release date: Jun 26, 2013
>OPAL: 1.7.2
>  OPAL repo revision: r28673
>   OPAL release date: Jun 26, 2013
>Ident string: 1.7.2
>  Prefix: /xxx/mbrucher/openmpi
> Configured architecture: x86_64-unknown-linux-gnu
>  Configure host: xxx.xxx.com
>   Configured by: mbrucher
>   Configured on: Thu Sep 12 10:22:06 BST 2013
>  Configure host: xxx.xxx.com
>Built by: mbrucher
>Built on: Thu Sep 12 10:24:59 BST 2013
>  Built host: xxx.xxx.com
>  C compiler: gcc
> C compiler absolute: /usr/bin/gcc
>  C compiler family name: GNU
>  C compiler version: 4.4.6
>  Internal debug support: no
> Memory profiling support: no
> Memory debugging support: no
> libltdl support: yes
>   Heterogeneous support: no
> orterun default --prefix: no
>   MPI_WTIME support: gettimeofday
> Symbol vis. support: yes
>   FT Checkpoint support: no (checkpoint thread: no)
> [abgengcluster:45509] *** Process received signal ***
> [abgengcluster:45509] Signal: Segmentation fault (11)
> [abgengcluster:45509] Signal code: Address not mapped (1)
> [abgengcluster:45509] Failing at address: 0xf8
> [abgengcluster:45509] [ 0] /lib64/libpthread.so.0() [0x3ffc00f4a0]
> [abgengcluster:45509] [ 1]
> /xxx/mbrucher/openmpi/lib/libopen-pal.so.5(opal_libevent2019_event_priority_set+0x6f)
> [0x2aae84a736ef]
> [abgengcluster:45509] [ 2]
> /xxx/mbrucher/openmpi/lib/libopen-rte.so.5(orte_iof_base_open+0x31c)
> [0x2aae847edfbc]
> [abgengcluster:45509] [ 3] orte-info(orte_info_open_components+0x71f) 
> [0x406b8f]
> [abgengcluster:45509] [ 4] orte-info(main+0x93d) [0x40450d]
> [abgengcluster:45509] [ 5] /lib64/libc.so.6(__libc_start_main+0xfd)
> [0x3ffb81ecdd]
> [abgengcluster:45509] [ 6] orte-info() [0x403b09]
> [abgengcluster:45509] *** End of error message ***
> Segmentation fault (core dumped)
> 
> Is there something that I missed?
> 
> Cheers,
> 
> Matthieu
> -- 
> Information System Engineer, Ph.D.
> Blog: http://matt.eifelle.com
> LinkedIn: http://www.linkedin.com/in/matthieubrucher
> Music band: http://liliejay.com/
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users



Re: [OMPI users] Segmentation fault with fresh compilation of 1.7.2

2013-09-15 Thread Matthieu Brucher
Just --with-lsf. Perhaps because then it must be launched through lsf?
Le 15 sept. 2013 18:02, "Ralph Castain"  a écrit :

> I'm not entirely sure - I don't see anything that would cause that problem
> in that location. How did you configure this?
>
>
> On Sep 12, 2013, at 3:17 AM, Matthieu Brucher 
> wrote:
>
> > Hi,
> >
> > I compiled OpenMPI on a RHEL6 box with LSF support, but when I run
> > sonthing, it crashes. Also orte-info crashes:
> >
> > Package: Open MPI mbruc...@xxx.com Distribution
> >Open RTE: 1.7.2
> >  Open RTE repo revision: r28673
> >   Open RTE release date: Jun 26, 2013
> >OPAL: 1.7.2
> >  OPAL repo revision: r28673
> >   OPAL release date: Jun 26, 2013
> >Ident string: 1.7.2
> >  Prefix: /xxx/mbrucher/openmpi
> > Configured architecture: x86_64-unknown-linux-gnu
> >  Configure host: xxx.xxx.com
> >   Configured by: mbrucher
> >   Configured on: Thu Sep 12 10:22:06 BST 2013
> >  Configure host: xxx.xxx.com
> >Built by: mbrucher
> >Built on: Thu Sep 12 10:24:59 BST 2013
> >  Built host: xxx.xxx.com
> >  C compiler: gcc
> > C compiler absolute: /usr/bin/gcc
> >  C compiler family name: GNU
> >  C compiler version: 4.4.6
> >  Internal debug support: no
> > Memory profiling support: no
> > Memory debugging support: no
> > libltdl support: yes
> >   Heterogeneous support: no
> > orterun default --prefix: no
> >   MPI_WTIME support: gettimeofday
> > Symbol vis. support: yes
> >   FT Checkpoint support: no (checkpoint thread: no)
> > [abgengcluster:45509] *** Process received signal ***
> > [abgengcluster:45509] Signal: Segmentation fault (11)
> > [abgengcluster:45509] Signal code: Address not mapped (1)
> > [abgengcluster:45509] Failing at address: 0xf8
> > [abgengcluster:45509] [ 0] /lib64/libpthread.so.0() [0x3ffc00f4a0]
> > [abgengcluster:45509] [ 1]
> >
> /xxx/mbrucher/openmpi/lib/libopen-pal.so.5(opal_libevent2019_event_priority_set+0x6f)
> > [0x2aae84a736ef]
> > [abgengcluster:45509] [ 2]
> > /xxx/mbrucher/openmpi/lib/libopen-rte.so.5(orte_iof_base_open+0x31c)
> > [0x2aae847edfbc]
> > [abgengcluster:45509] [ 3] orte-info(orte_info_open_components+0x71f)
> [0x406b8f]
> > [abgengcluster:45509] [ 4] orte-info(main+0x93d) [0x40450d]
> > [abgengcluster:45509] [ 5] /lib64/libc.so.6(__libc_start_main+0xfd)
> > [0x3ffb81ecdd]
> > [abgengcluster:45509] [ 6] orte-info() [0x403b09]
> > [abgengcluster:45509] *** End of error message ***
> > Segmentation fault (core dumped)
> >
> > Is there something that I missed?
> >
> > Cheers,
> >
> > Matthieu
> > --
> > Information System Engineer, Ph.D.
> > Blog: http://matt.eifelle.com
> > LinkedIn: http://www.linkedin.com/in/matthieubrucher
> > Music band: http://liliejay.com/
> > ___
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>


Re: [OMPI users] Segmentation fault with fresh compilation of 1.7.2

2013-09-15 Thread Ralph Castain
No - out of curiosity, does ompi_info work? I'm wondering if this is strictly 
an orte-info problem.

On Sep 15, 2013, at 10:03 AM, Matthieu Brucher  
wrote:

> Just --with-lsf. Perhaps because then it must be launched through lsf?
> 
> Le 15 sept. 2013 18:02, "Ralph Castain"  a écrit :
> I'm not entirely sure - I don't see anything that would cause that problem in 
> that location. How did you configure this?
> 
> 
> On Sep 12, 2013, at 3:17 AM, Matthieu Brucher  
> wrote:
> 
> > Hi,
> >
> > I compiled OpenMPI on a RHEL6 box with LSF support, but when I run
> > sonthing, it crashes. Also orte-info crashes:
> >
> > Package: Open MPI mbruc...@xxx.com Distribution
> >Open RTE: 1.7.2
> >  Open RTE repo revision: r28673
> >   Open RTE release date: Jun 26, 2013
> >OPAL: 1.7.2
> >  OPAL repo revision: r28673
> >   OPAL release date: Jun 26, 2013
> >Ident string: 1.7.2
> >  Prefix: /xxx/mbrucher/openmpi
> > Configured architecture: x86_64-unknown-linux-gnu
> >  Configure host: xxx.xxx.com
> >   Configured by: mbrucher
> >   Configured on: Thu Sep 12 10:22:06 BST 2013
> >  Configure host: xxx.xxx.com
> >Built by: mbrucher
> >Built on: Thu Sep 12 10:24:59 BST 2013
> >  Built host: xxx.xxx.com
> >  C compiler: gcc
> > C compiler absolute: /usr/bin/gcc
> >  C compiler family name: GNU
> >  C compiler version: 4.4.6
> >  Internal debug support: no
> > Memory profiling support: no
> > Memory debugging support: no
> > libltdl support: yes
> >   Heterogeneous support: no
> > orterun default --prefix: no
> >   MPI_WTIME support: gettimeofday
> > Symbol vis. support: yes
> >   FT Checkpoint support: no (checkpoint thread: no)
> > [abgengcluster:45509] *** Process received signal ***
> > [abgengcluster:45509] Signal: Segmentation fault (11)
> > [abgengcluster:45509] Signal code: Address not mapped (1)
> > [abgengcluster:45509] Failing at address: 0xf8
> > [abgengcluster:45509] [ 0] /lib64/libpthread.so.0() [0x3ffc00f4a0]
> > [abgengcluster:45509] [ 1]
> > /xxx/mbrucher/openmpi/lib/libopen-pal.so.5(opal_libevent2019_event_priority_set+0x6f)
> > [0x2aae84a736ef]
> > [abgengcluster:45509] [ 2]
> > /xxx/mbrucher/openmpi/lib/libopen-rte.so.5(orte_iof_base_open+0x31c)
> > [0x2aae847edfbc]
> > [abgengcluster:45509] [ 3] orte-info(orte_info_open_components+0x71f) 
> > [0x406b8f]
> > [abgengcluster:45509] [ 4] orte-info(main+0x93d) [0x40450d]
> > [abgengcluster:45509] [ 5] /lib64/libc.so.6(__libc_start_main+0xfd)
> > [0x3ffb81ecdd]
> > [abgengcluster:45509] [ 6] orte-info() [0x403b09]
> > [abgengcluster:45509] *** End of error message ***
> > Segmentation fault (core dumped)
> >
> > Is there something that I missed?
> >
> > Cheers,
> >
> > Matthieu
> > --
> > Information System Engineer, Ph.D.
> > Blog: http://matt.eifelle.com
> > LinkedIn: http://www.linkedin.com/in/matthieubrucher
> > Music band: http://liliejay.com/
> > ___
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users



Re: [OMPI users] Segmentation fault with fresh compilation of 1.7.2

2013-09-15 Thread Matthieu Brucher
Yes, ompi_info does not crash.
Le 15 sept. 2013 18:05, "Ralph Castain"  a écrit :

> No - out of curiosity, does ompi_info work? I'm wondering if this is
> strictly an orte-info problem.
>
> On Sep 15, 2013, at 10:03 AM, Matthieu Brucher 
> wrote:
>
> Just --with-lsf. Perhaps because then it must be launched through lsf?
> Le 15 sept. 2013 18:02, "Ralph Castain"  a écrit :
>
>> I'm not entirely sure - I don't see anything that would cause that
>> problem in that location. How did you configure this?
>>
>>
>> On Sep 12, 2013, at 3:17 AM, Matthieu Brucher 
>> wrote:
>>
>> > Hi,
>> >
>> > I compiled OpenMPI on a RHEL6 box with LSF support, but when I run
>> > sonthing, it crashes. Also orte-info crashes:
>> >
>> > Package: Open MPI mbruc...@xxx.com Distribution
>> >Open RTE: 1.7.2
>> >  Open RTE repo revision: r28673
>> >   Open RTE release date: Jun 26, 2013
>> >OPAL: 1.7.2
>> >  OPAL repo revision: r28673
>> >   OPAL release date: Jun 26, 2013
>> >Ident string: 1.7.2
>> >  Prefix: /xxx/mbrucher/openmpi
>> > Configured architecture: x86_64-unknown-linux-gnu
>> >  Configure host: xxx.xxx.com
>> >   Configured by: mbrucher
>> >   Configured on: Thu Sep 12 10:22:06 BST 2013
>> >  Configure host: xxx.xxx.com
>> >Built by: mbrucher
>> >Built on: Thu Sep 12 10:24:59 BST 2013
>> >  Built host: xxx.xxx.com
>> >  C compiler: gcc
>> > C compiler absolute: /usr/bin/gcc
>> >  C compiler family name: GNU
>> >  C compiler version: 4.4.6
>> >  Internal debug support: no
>> > Memory profiling support: no
>> > Memory debugging support: no
>> > libltdl support: yes
>> >   Heterogeneous support: no
>> > orterun default --prefix: no
>> >   MPI_WTIME support: gettimeofday
>> > Symbol vis. support: yes
>> >   FT Checkpoint support: no (checkpoint thread: no)
>> > [abgengcluster:45509] *** Process received signal ***
>> > [abgengcluster:45509] Signal: Segmentation fault (11)
>> > [abgengcluster:45509] Signal code: Address not mapped (1)
>> > [abgengcluster:45509] Failing at address: 0xf8
>> > [abgengcluster:45509] [ 0] /lib64/libpthread.so.0() [0x3ffc00f4a0]
>> > [abgengcluster:45509] [ 1]
>> >
>> /xxx/mbrucher/openmpi/lib/libopen-pal.so.5(opal_libevent2019_event_priority_set+0x6f)
>> > [0x2aae84a736ef]
>> > [abgengcluster:45509] [ 2]
>> > /xxx/mbrucher/openmpi/lib/libopen-rte.so.5(orte_iof_base_open+0x31c)
>> > [0x2aae847edfbc]
>> > [abgengcluster:45509] [ 3] orte-info(orte_info_open_components+0x71f)
>> [0x406b8f]
>> > [abgengcluster:45509] [ 4] orte-info(main+0x93d) [0x40450d]
>> > [abgengcluster:45509] [ 5] /lib64/libc.so.6(__libc_start_main+0xfd)
>> > [0x3ffb81ecdd]
>> > [abgengcluster:45509] [ 6] orte-info() [0x403b09]
>> > [abgengcluster:45509] *** End of error message ***
>> > Segmentation fault (core dumped)
>> >
>> > Is there something that I missed?
>> >
>> > Cheers,
>> >
>> > Matthieu
>> > --
>> > Information System Engineer, Ph.D.
>> > Blog: http://matt.eifelle.com
>> > LinkedIn: http://www.linkedin.com/in/matthieubrucher
>> > Music band: http://liliejay.com/
>> > ___
>> > users mailing list
>> > us...@open-mpi.org
>> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
>> ___
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
>
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>


Re: [OMPI users] Segmentation fault with fresh compilation of 1.7.2

2013-09-15 Thread Ralph Castain
Could you try the current 1.7.3 nightly tarball instead? I don't see a problem 
there, and I'm wondering if this is something we already fixed. We will be 
releasing 1.7.3 shortly and it is mostly complete at this time.


On Sep 15, 2013, at 10:43 AM, Matthieu Brucher  
wrote:

> Yes, ompi_info does not crash.
> 
> Le 15 sept. 2013 18:05, "Ralph Castain"  a écrit :
> No - out of curiosity, does ompi_info work? I'm wondering if this is strictly 
> an orte-info problem.
> 
> On Sep 15, 2013, at 10:03 AM, Matthieu Brucher  
> wrote:
> 
>> Just --with-lsf. Perhaps because then it must be launched through lsf?
>> 
>> Le 15 sept. 2013 18:02, "Ralph Castain"  a écrit :
>> I'm not entirely sure - I don't see anything that would cause that problem 
>> in that location. How did you configure this?
>> 
>> 
>> On Sep 12, 2013, at 3:17 AM, Matthieu Brucher  
>> wrote:
>> 
>> > Hi,
>> >
>> > I compiled OpenMPI on a RHEL6 box with LSF support, but when I run
>> > sonthing, it crashes. Also orte-info crashes:
>> >
>> > Package: Open MPI mbruc...@xxx.com Distribution
>> >Open RTE: 1.7.2
>> >  Open RTE repo revision: r28673
>> >   Open RTE release date: Jun 26, 2013
>> >OPAL: 1.7.2
>> >  OPAL repo revision: r28673
>> >   OPAL release date: Jun 26, 2013
>> >Ident string: 1.7.2
>> >  Prefix: /xxx/mbrucher/openmpi
>> > Configured architecture: x86_64-unknown-linux-gnu
>> >  Configure host: xxx.xxx.com
>> >   Configured by: mbrucher
>> >   Configured on: Thu Sep 12 10:22:06 BST 2013
>> >  Configure host: xxx.xxx.com
>> >Built by: mbrucher
>> >Built on: Thu Sep 12 10:24:59 BST 2013
>> >  Built host: xxx.xxx.com
>> >  C compiler: gcc
>> > C compiler absolute: /usr/bin/gcc
>> >  C compiler family name: GNU
>> >  C compiler version: 4.4.6
>> >  Internal debug support: no
>> > Memory profiling support: no
>> > Memory debugging support: no
>> > libltdl support: yes
>> >   Heterogeneous support: no
>> > orterun default --prefix: no
>> >   MPI_WTIME support: gettimeofday
>> > Symbol vis. support: yes
>> >   FT Checkpoint support: no (checkpoint thread: no)
>> > [abgengcluster:45509] *** Process received signal ***
>> > [abgengcluster:45509] Signal: Segmentation fault (11)
>> > [abgengcluster:45509] Signal code: Address not mapped (1)
>> > [abgengcluster:45509] Failing at address: 0xf8
>> > [abgengcluster:45509] [ 0] /lib64/libpthread.so.0() [0x3ffc00f4a0]
>> > [abgengcluster:45509] [ 1]
>> > /xxx/mbrucher/openmpi/lib/libopen-pal.so.5(opal_libevent2019_event_priority_set+0x6f)
>> > [0x2aae84a736ef]
>> > [abgengcluster:45509] [ 2]
>> > /xxx/mbrucher/openmpi/lib/libopen-rte.so.5(orte_iof_base_open+0x31c)
>> > [0x2aae847edfbc]
>> > [abgengcluster:45509] [ 3] orte-info(orte_info_open_components+0x71f) 
>> > [0x406b8f]
>> > [abgengcluster:45509] [ 4] orte-info(main+0x93d) [0x40450d]
>> > [abgengcluster:45509] [ 5] /lib64/libc.so.6(__libc_start_main+0xfd)
>> > [0x3ffb81ecdd]
>> > [abgengcluster:45509] [ 6] orte-info() [0x403b09]
>> > [abgengcluster:45509] *** End of error message ***
>> > Segmentation fault (core dumped)
>> >
>> > Is there something that I missed?
>> >
>> > Cheers,
>> >
>> > Matthieu
>> > --
>> > Information System Engineer, Ph.D.
>> > Blog: http://matt.eifelle.com
>> > LinkedIn: http://www.linkedin.com/in/matthieubrucher
>> > Music band: http://liliejay.com/
>> > ___
>> > users mailing list
>> > us...@open-mpi.org
>> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> ___
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> ___
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users



Re: [OMPI users] Segmentation fault with fresh compilation of 1.7.2

2013-09-15 Thread Matthieu Brucher
I can try later this week, yes.
Thanks
Le 15 sept. 2013 19:09, "Ralph Castain"  a écrit :

> Could you try the current 1.7.3 nightly tarball instead? I don't see a
> problem there, and I'm wondering if this is something we already fixed. We
> will be releasing 1.7.3 shortly and it is mostly complete at this time.
>
>
> On Sep 15, 2013, at 10:43 AM, Matthieu Brucher 
> wrote:
>
> Yes, ompi_info does not crash.
> Le 15 sept. 2013 18:05, "Ralph Castain"  a écrit :
>
>> No - out of curiosity, does ompi_info work? I'm wondering if this is
>> strictly an orte-info problem.
>>
>> On Sep 15, 2013, at 10:03 AM, Matthieu Brucher <
>> matthieu.bruc...@gmail.com> wrote:
>>
>> Just --with-lsf. Perhaps because then it must be launched through lsf?
>> Le 15 sept. 2013 18:02, "Ralph Castain"  a écrit :
>>
>>> I'm not entirely sure - I don't see anything that would cause that
>>> problem in that location. How did you configure this?
>>>
>>>
>>> On Sep 12, 2013, at 3:17 AM, Matthieu Brucher <
>>> matthieu.bruc...@gmail.com> wrote:
>>>
>>> > Hi,
>>> >
>>> > I compiled OpenMPI on a RHEL6 box with LSF support, but when I run
>>> > sonthing, it crashes. Also orte-info crashes:
>>> >
>>> > Package: Open MPI mbruc...@xxx.com Distribution
>>> >Open RTE: 1.7.2
>>> >  Open RTE repo revision: r28673
>>> >   Open RTE release date: Jun 26, 2013
>>> >OPAL: 1.7.2
>>> >  OPAL repo revision: r28673
>>> >   OPAL release date: Jun 26, 2013
>>> >Ident string: 1.7.2
>>> >  Prefix: /xxx/mbrucher/openmpi
>>> > Configured architecture: x86_64-unknown-linux-gnu
>>> >  Configure host: xxx.xxx.com
>>> >   Configured by: mbrucher
>>> >   Configured on: Thu Sep 12 10:22:06 BST 2013
>>> >  Configure host: xxx.xxx.com
>>> >Built by: mbrucher
>>> >Built on: Thu Sep 12 10:24:59 BST 2013
>>> >  Built host: xxx.xxx.com
>>> >  C compiler: gcc
>>> > C compiler absolute: /usr/bin/gcc
>>> >  C compiler family name: GNU
>>> >  C compiler version: 4.4.6
>>> >  Internal debug support: no
>>> > Memory profiling support: no
>>> > Memory debugging support: no
>>> > libltdl support: yes
>>> >   Heterogeneous support: no
>>> > orterun default --prefix: no
>>> >   MPI_WTIME support: gettimeofday
>>> > Symbol vis. support: yes
>>> >   FT Checkpoint support: no (checkpoint thread: no)
>>> > [abgengcluster:45509] *** Process received signal ***
>>> > [abgengcluster:45509] Signal: Segmentation fault (11)
>>> > [abgengcluster:45509] Signal code: Address not mapped (1)
>>> > [abgengcluster:45509] Failing at address: 0xf8
>>> > [abgengcluster:45509] [ 0] /lib64/libpthread.so.0() [0x3ffc00f4a0]
>>> > [abgengcluster:45509] [ 1]
>>> >
>>> /xxx/mbrucher/openmpi/lib/libopen-pal.so.5(opal_libevent2019_event_priority_set+0x6f)
>>> > [0x2aae84a736ef]
>>> > [abgengcluster:45509] [ 2]
>>> > /xxx/mbrucher/openmpi/lib/libopen-rte.so.5(orte_iof_base_open+0x31c)
>>> > [0x2aae847edfbc]
>>> > [abgengcluster:45509] [ 3] orte-info(orte_info_open_components+0x71f)
>>> [0x406b8f]
>>> > [abgengcluster:45509] [ 4] orte-info(main+0x93d) [0x40450d]
>>> > [abgengcluster:45509] [ 5] /lib64/libc.so.6(__libc_start_main+0xfd)
>>> > [0x3ffb81ecdd]
>>> > [abgengcluster:45509] [ 6] orte-info() [0x403b09]
>>> > [abgengcluster:45509] *** End of error message ***
>>> > Segmentation fault (core dumped)
>>> >
>>> > Is there something that I missed?
>>> >
>>> > Cheers,
>>> >
>>> > Matthieu
>>> > --
>>> > Information System Engineer, Ph.D.
>>> > Blog: http://matt.eifelle.com
>>> > LinkedIn: http://www.linkedin.com/in/matthieubrucher
>>> > Music band: http://liliejay.com/
>>> > ___
>>> > users mailing list
>>> > us...@open-mpi.org
>>> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>
>>> ___
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>
>> ___
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
>>
>>
>> ___
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
>
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>