Message: 9 Date: Wed, 1 Sep 2010 20:14:44 -0600 From: Ralph Castain<r...@open-mpi.org> Subject: Re: [OMPI users] MPI_Comm_accept and MPI_Comm_connect both use 100% one cpu core. Is it a bug? To: Open MPI Users<us...@open-mpi.org> Message-ID:<4e4bc153-b4e3-43e2-b980-904dabe78...@open-mpi.org> Content-Type: text/plain; charset="us-ascii"It's not a bug - that is normal behavior. The processes are polling hard to establish the connections as quickly as possible. On Sep 1, 2010, at 7:24 PM, lyb wrote:> Hi, All,> > I tested two sample applications on Windows 2003 Server, one use MPI_Comm_accept and other use MPI_Comm_connect,> when run into MPI_Comm_accept or MPI_Comm_connect, the application use 100% one cpu core. Is it a bug or some wrong?> > I tested with three version including Version 1.4 (stable), Version 1.5 (prerelease) and trunk 23706 version. > > ...> MPI_Open_port(MPI_INFO_NULL, port); > MPI_Comm_accept( port, MPI_INFO_NULL, 0, MPI_COMM_WORLD,&client ); > ...> > ...> MPI_Comm_connect( port, MPI_INFO_NULL, 0, MPI_COMM_WORLD,&server ); > ...> > thanks a lot. > > lyb > > > > > > _______________________________________________> users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users-------------- next part -------------- HTML attachment scrubbed and removed
Thanks for your answer, but I test with MPICH2, it doesn't have this
fault. Why?
- Re: [OMPI users] users Digest, Vol 1674, Issue 1 lyb
- Re: [OMPI users] users Digest, Vol 1674, Issue 1 Jeff Squyres