On 1/20/10 5:38 PM, Eloi Gaudry wrote:
Hi,
FYI, This issue is solved with the last version of the library
(v2-1.11), at least on my side.
Eloi
Hi Eloi,
Thanks a lot for the message. The issues are fixed on my side too.
Dorian
Hi,
FYI, This issue is solved with the last version of the library
(v2-1.11), at least on my side.
Eloi
Gus Correa wrote:
Hi Dorian
Dorian Krause wrote:
Hi,
@Gus I don't use any flags for the installed OpenMPI version. In fact
for this mail I used an OpenMPI version just installed with t
Hi Dorian
Dorian Krause wrote:
Hi,
@Gus I don't use any flags for the installed OpenMPI version. In fact
for this mail I used an OpenMPI version just installed with the
--enable-debug flag.
You are right, my guess was wrong.
Just that some people reported trouble with threaded
versions of
Hi,
@Gus I don't use any flags for the installed OpenMPI version. In fact
for this mail I used an OpenMPI version just installed with the
--enable-debug flag.
From what I can tell from stepping through the debugger the problem
happens in btl_openib_component_init:
#0 btl_openib_component_
Can you get a core dump, or otherwise see exactly where the seg fault is
occurring?
On Jan 18, 2010, at 8:34 AM, Dorian Krause wrote:
> Hi Eloi,
> >
> > Does the segmentation faults you're facing also happen in a sequential
> > environment (i.e. not linked against openmpi libraries) ?
>
> No, w
Yaakoub El Khamra wrote:
Greetings
Can we please verify this problem is with Gotoblas and not with
OpenMPI? if I read this correctly, without MPI and with other flavors
of MPI, you have normal execution. This would normally indicate the
problem is on the OpenMPI side.
I am 2 doors away from Kazu
Greetings
Can we please verify this problem is with Gotoblas and not with
OpenMPI? if I read this correctly, without MPI and with other flavors
of MPI, you have normal execution. This would normally indicate the
problem is on the OpenMPI side.
I am 2 doors away from Kazushige's office. Please do l
Gus Correa wrote:
Hi Dorian and Eloi
I wonder if this is really a Goto BLAS problem or related to
how OpenMPI was configured.
In a recent sequence of postings on this list
a colleague reported several errors which were fixed
after he removed the (non-default) "--enable-mpi-threads"
flag from hi
Hi Dorian and Eloi
I wonder if this is really a Goto BLAS problem or related to
how OpenMPI was configured.
In a recent sequence of postings on this list
a colleague reported several errors which were fixed
after he removed the (non-default) "--enable-mpi-threads"
flag from his OpenMPI configura
Dorian Krause wrote:
Hi Eloi,
Does the segmentation faults you're facing also happen in a
sequential environment (i.e. not linked against openmpi libraries) ?
No, without MPI everything works fine. Also, linking against mvapich
doesn't give any errors. I think there is a problem with GotoBL
Hi Eloi,
Does the segmentation faults you're facing also happen in a sequential
environment (i.e. not linked against openmpi libraries) ?
No, without MPI everything works fine. Also, linking against mvapich
doesn't give any errors. I think there is a problem with GotoBLAS and
the shared lib
Dorian Krause wrote:
Hi,
has any one successfully combined OpenMPI and GotoBLAS2? I'm facing
segfaults in any program which combines the two libraries (as shared
libs). The segmentation fault seems to occur in MPI_Init(). The gdb
backtrace is
Program received signal SIGSEGV, Segmentation fa
Hi,
has any one successfully combined OpenMPI and GotoBLAS2? I'm facing
segfaults in any program which combines the two libraries (as shared
libs). The segmentation fault seems to occur in MPI_Init(). The gdb
backtrace is
Program received signal SIGSEGV, Segmentation fault.
[Switching to Thr
13 matches
Mail list logo