Two notes for you:

1. Your program does necessarily guarantee what you might expect: since you use ANY_SOURCE/ANY_TAG in both the receives, you might actually get two receives from the same sender in a given iteration. The fact that you're effectively using yield_when_idle (which OMPI will automatically enable when you tell it "slots=1" but you run with - np 3) means that you probably *won't* have this happen (because every MPI process will yield on every iteration, effectively keeping all 3 in lock step), but it still *can* happen (and did frequently in my tests).

2. The problem you're seeing is an optimization called "early completion" where, for latency ping-pong optimizations, Open MPI may indicate that a send has "completed" before the message is actually placed on the network (shared memory, in your case). This can be a nice performance boost for applications that both a) dip into the MPI layer frequently and b) synchronize at some point. Your application is not necessarily doing this in the final iterations; it may reach MPI_FINALIZE while there's still a pile of messages that have been queued for delivery before they are actually progressed out the network to the receiver. In our upcoming 1.2.6 release, there is a run-time parameter to disable this early completion behavior (i.e., never signal completion of a send before the data is actually transmitted out on the network). You can try the 1.2.6rc2 tarball:

    http://www.open-mpi.org/software/ompi/v1.2/

And use the following MCA parameter:

    mpirun --mca pml_ob1_use_early_completion 0 ...

See if that works for you.


On Mar 18, 2008, at 7:11 AM, Giovani Faccin wrote:

Hi Andreas, thanks for the reply!

I'm using openmpi-1.2.5. It was installed using my distro's (Gentoo) default package:

sys-cluster/openmpi-1.2.5 USE="fortran ipv6 -debug -heterogeneous - nocxx -pbs -romio -smp -threads"

I've tried setting the mpi_yield_when_idle parameter as you asked. However, the program still hangs.

Just in case, the command line I'm using to call it is this:
/usr/bin/mpirun --hostfile mpi-config.txt --mca mpi_yield_when_idle 1 -np 3 /home/gfaccin/desenvolvimento/Eclipse/mpiplay/Debug/mpiplay

where mpi-config.txt contains the following line:
localhost slots=1

Anything else I could try?

Thank you!

Giovani

Andreas Schäfer <gent...@gmx.de> escreveu: Hmm, strange. It doesn't hang for me and AFAICS it shouldn't hang at
all. I'm using 1.2.5. Which version of Open MPI are you using?

Hanging with 100% CPU utilization often means that your processes are
caught in a busy wait. You could try to set mpi_yield_when_idle:

> gentryx@hex ~ $ cat .openmpi/mca-params.conf
> mpi_yield_when_idle=1

But I don't think this should be necessary.

HTH
-Andreas


On 21:35 Mon 17 Mar , Giovani Faccin wrote:
> Hi there!
>
> I'm learning MPI, and got really puzzled... Please take a look at this very short code:
>
> #include
> #include "mpicxx.h"
> using namespace std;
> int main(int argc, char *argv[])
> {
> MPI::Init();
>
> for (unsigned long t = 0; t < 10000000; t++)
> {
> //If we are process 0:
> if ( MPI::COMM_WORLD.Get_rank() == 0 )
> {
> MPI::Status mpi_status;
> unsigned long d = 0;
> unsigned long d2 = 0;
> MPI::COMM_WORLD.Recv(&d, 1, MPI::UNSIGNED_LONG, MPI::ANY_SOURCE, MPI::ANY_TAG, mpi_status ); > MPI::COMM_WORLD.Recv(&d2, 1, MPI::UNSIGNED_LONG, MPI::ANY_SOURCE, MPI::ANY_TAG, mpi_status ); > cout << "Time = " << t << "; Node 0 received: " << d << " and " << d2 << endl;
> }
> //Else:
> else
> {
> unsigned long d = MPI::COMM_WORLD.Get_rank();
> MPI::COMM_WORLD.Send( &d, 1, MPI::UNSIGNED_LONG, 0, 0);
> };
> };
> MPI::Finalize();
> }
>
> Ok, so what I'm trying to do is to make a gather operation using point to point communication. In my real application instead of sending an unsigned long I'd be calling an object's send and receive methods, which in turn would call their inner object's similar methods and so on until all data is syncronized. I'm using this loop because the number of objects to be sent to process rank 0 varies depending on the sender.
>
> When running this test with 3 processes on a dual core, oversubscribed node, I get this output:
> (skipped previous output)
> Time = 5873; Node 0 received: 1 and 2
> Time = 5874; Node 0 received: 1 and 2
> Time = 5875; Node 0 received: 1 and 2
> Time = 5876; Node 0 received: 1 and 2
>
> and then the application hangs, with processor usage at 100%. The exact time when this condition occurs varies on each run, but it usually happens quite fast.
>
> What would I have to modify, in this simple example, so that the application works as expected? Must I always use Gather, instead of point to point, to make a syncronization like this?
>
> Thank you very much!
>
> Giovani
>
>
>
>
>
>
>
> __________________________________________________
> Fale com seus amigos de graça com o novo Yahoo! Messenger
> http://br.messenger.yahoo.com/
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

--
============================================
Andreas Schäfer
Cluster and Metacomputing Working Group
Friedrich-Schiller-Universität Jena, Germany
PGP/GPG key via keyserver
I'm a bright... http://www.the-brights.net
============================================

(\___/)
(+'.'+)
(")_(")
This is Bunny. Copy and paste Bunny into your
signature to help him gain world domination!
_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users

Abra sua conta no Yahoo! Mail, o único sem limite de espaço para armazenamento! _______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users


--
Jeff Squyres
Cisco Systems


Reply via email to