From: Clay Kirkland [mailto:clay.kirkl...@versityinc.com]
Sent: Friday, May 02, 2014 03:24 PM
To: us...@open-mpi.org
Subject: [OMPI users] MPI_Barrier hangs on second attempt but only when
multiple hosts used.
I have been using MPI for many many years so I have very well debugged mpi
tests
Hmmm...just testing on my little cluster here on two nodes, it works just fine
with 1.8.2:
[rhc@bend001 v1.8]$ mpirun -n 2 --map-by node ./a.out
In rank 0 and host= bend001 Do Barrier call 1.
In rank 0 and host= bend001 Do Barrier call 2.
In rank 0 and host= bend001 Do Barrier call 3.
In r
I have been using MPI for many many years so I have very well debugged mpi
tests. I am
having trouble on either openmpi-1.4.5 or openmpi-1.6.5 versions though
with getting the
MPI_Barrier calls to work. It works fine when I run all processes on one
machine but when
I run with two or more host
Dear George,
Please see below.
On 29 September 2013 01:03, George Bosilca wrote:
>
> On Sep 29, 2013, at 01:19 , Huangwei wrote:
>
> Dear All,
>
> In my code I implement mpi_send/mpi_receive for an three dimensional real
> array, and process is as follows:
>
> all other processors send the ar
On Sep 29, 2013, at 01:19 , Huangwei wrote:
> Dear All,
>
> In my code I implement mpi_send/mpi_receive for an three dimensional real
> array, and process is as follows:
>
> all other processors send the array to rank 0 and then rank 0 receives the
> array and put these arrays into a comple
Dear All,
In my code I implement mpi_send/mpi_receive for an three dimensional real
array, and process is as follows:
all other processors send the array to rank 0 and then rank 0 receives the
array and put these arrays into a complete array. Then mpi_bcast is called
to send the complete array fr
When you receive that callback the MPI has ben put in a quiescent state. As
such it does not allow MPI communication until the checkpoint is completely
finished. So you cannot call barrier in the checkpoint callback. Since Open
MPI did doing a coordinated checkpoint, you can assume that all process
Dear Group,
I wanted to do a synchronization check with 'MPI_Barrier(MPI_COMM_WORLD)' in
'opal_crs_self_user_checkpoint(char **restart_cmd)' call. Although every
process is present in this call, it fails to synchronize. Is there any reason
why cant we use barrier?Thanks in advance.
Kind regards
P.P.S. I ran the same test with OpenMPI 1.5.4, the behaviour is the same.
Evgeniy
Message: 10
List-Post: users@lists.open-mpi.org
Date: Sat, 28 Jan 2012 08:24:39 -0500
From: Jeff Squyres
Subject: Re: [OMPI users] MPI_Barrier, again
To: Open MPI Users
Message-ID: <1859c141-813d-46ba-9
P.S. I have tested with OpenMPI 1.4.5rc4 and the problem is still there.
Evgeniy
./mpibarriertest.f90
OpenMPI: 1.4.3
hangs with 15 processes randomly as described.
Evgeniy
Message: 10
List-Post: users@lists.open-mpi.org
Date: Sat, 28 Jan 2012 08:24:39 -0500
From: Jeff Squyres
Subject: Re: [OMPI users] MPI_Barrier, again
To: Open MPI Users
Message-ID: <1859c141-813d-46ba-97bc-4b0290
Is there any chance you can make a small-ish reproducer of the issue that we
can run?
On Jan 27, 2012, at 10:45 AM, Evgeniy Shapiro wrote:
> Hi
>
> I have a strange problem with MPI_Barrier occurring when writing to a
> file. The output subroutine (the code is in FORTRAN) is called from
> the m
Hi
I have a strange problem with MPI_Barrier occurring when writing to a
file. The output subroutine (the code is in FORTRAN) is called from
the main program and there is an MPI_Barrier just before the call.
In the subroutine
1. Process 0 checks whether the first file exists and, if not, -
creat
Hi Jeff,
thanks for the explanation. Yes, some of the earlier discussion where in
fact very useful. In general I found this list to be very helpful, my thanks
to everyone here who is helping people like me out.
The suggestion to use messages and non-blocking receives with MPI_Test()
proved just w
On Dec 4, 2009, at 6:54 PM, Nicolas Bock wrote:
> in our code we use a very short front-end program to drive a larger set of
> codes that do our calculations. Right in the beginning of the front-end, we
> have an if() statement such that only the rank 0 front-end does something,
> and the other
Hello list,
in our code we use a very short front-end program to drive a larger set of
codes that do our calculations. Right in the beginning of the front-end, we
have an if() statement such that only the rank 0 front-end does something,
and the other ranks go right away to an MPI_Barrier() statem
This helped me track down a leaked file descriptor, but I think the
order of events is not desirable. If an MPIIO file descriptor is not
closed before MPI_Finalize, I get the following.
*** An error occurred in MPI_Barrier
*** after MPI was finalized
*** MPI_ERRORS_ARE_FATAL (your MPI job will n
I have no idea how this e-mail slipped by me -- I'm sorry for the
delay in replying.
Can you run your code through a memory-checking debugger such as
valgrind and see if it turns up any ancillary memory problems?
On Dec 1, 2006, at 1:52 PM, Cupp, Matthew R wrote:
Hi,
In my program I a
Hi,
In my program I am calling MPI_Barrier(MPI_COMM_WORLD), but it seems to
cause an error on one node. The node changes depending on how many
total nodes I have (it could be 4 or 2). I'm thinking it's MPI_Barrier
because I put print statements around it and that's where the program
terminate
19 matches
Mail list logo