f Squyres wrote:
From: Jeff Squyres
Subject: Re: [OMPI users] MPI_Bcast issue
To: "Open MPI Users"
Received: Friday, 13 August, 2010, 3:03 AM
Dick / all --
I just had a phone call with Ralph Castain who has had some additional off-list
mails with Randolph. Apparently, none of us u
- MPI Team
> IBM Systems & Technology Group
> Dept X2ZA / MS P963 -- 2455 South Road -- Poughkeepsie, NY 12601
> Tele (845) 433-7846 Fax (845) 433-8363
>
>
> users-boun...@open-mpi.org wrote on 08/11/2010 08:59:16 PM:
>
> > [image removed]
> >
n...@open-mpi.org wrote on 08/11/2010 08:59:16 PM:
> [image removed]
>
> Re: [OMPI users] MPI_Bcast issue
>
> Randolph Pullen
>
> to:
>
> Open MPI Users
>
> 08/11/2010 09:01 PM
>
> Sent by:
>
> users-boun...@open-mpi.org
>
> Please respon
Interesting point.
--- On Thu, 12/8/10, Ashley Pittman wrote:
From: Ashley Pittman
Subject: Re: [OMPI users] MPI_Bcast issue
To: "Open MPI Users"
Received: Thursday, 12 August, 2010, 12:22 AM
On 11 Aug 2010, at 05:10, Randolph Pullen wrote:
> Sure, but broadcasts are faster -
question is why.
--- On Wed, 11/8/10, Richard Treumann wrote:
From: Richard Treumann
Subject: Re: [OMPI users] MPI_Bcast issue
To: "Open MPI Users"
Received: Wednesday, 11 August, 2010, 11:34 PM
Randolf
I am confused about using multiple,
concurrent mpirun operations. If there are
On 11 Aug 2010, at 05:10, Randolph Pullen wrote:
> Sure, but broadcasts are faster - less reliable apparently, but much faster
> for large clusters.
Going off-topic here but I think it's worth saying:
If you have a dataset that requires collective communication then use the
function call that
On Aug 11, 2010, at 12:10 AM, Randolph Pullen wrote:
> Sure, but broadcasts are faster - less reliable apparently, but much faster
> for large clusters.
Just to be totally clear: MPI_BCAST is defined to be "reliable", in the sense
that it will complete or invoke an error (vs. unreliable data
On Aug 11, 2010, at 9:54 AM, Jeff Squyres wrote:
> (I'll say that OMPI's ALLGATHER algorithm is probably not well optimized for
> massive data transfers like you describe)
Wrong wrong wrong -- I should have checked the code before sending. I made the
incorrect assumption that OMPI still only h
On Aug 10, 2010, at 10:09 PM, Randolph Pullen wrote:
> Jeff thanks for the clarification,
> What I am trying to do is run N concurrent copies of a 1 to N data movement
> program to affect an N to N solution. The actual mechanism I am using is to
> spawn N copies of mpirun from PVM across the cl
Randolf
I am confused about using multiple, concurrent mpirun operations. If
there are M uses of mpirun and each starts N tasks (carried out under pvm
or any other way) I would expect you to have M completely independent MPI
jobs with N tasks (processes) each. You could have some root in eac
wrote:
From: Terry Frankcombe
Subject: Re: [OMPI users] MPI_Bcast issue
To: "Open MPI Users"
Received: Wednesday, 11 August, 2010, 1:57 PM
On Tue, 2010-08-10 at 19:09 -0700, Randolph Pullen wrote:
> Jeff thanks for the clarification,
> What I am trying to do is run N concurre
On Tue, 2010-08-10 at 19:09 -0700, Randolph Pullen wrote:
> Jeff thanks for the clarification,
> What I am trying to do is run N concurrent copies of a 1 to N data
> movement program to affect an N to N solution.
I'm no MPI guru, nor do I completely understand what you are doing, but
isn't this an
rom: Jeff Squyres
Subject: Re: [OMPI users] MPI_Bcast issue
To: "Open MPI Users"
Received: Wednesday, 11 August, 2010, 6:24 AM
+1 on Eugene's comment that I don't fully understand what you are trying to do.
Can you send a short example code?
Some random points:
- Edgar alre
rote:
> The install was completly vanilla - no extras a plain .configure command line
> (on FC10 x8x_64 linux)
>
> Are you saying that all broadcast calls are actually implemented as serial
> point to point calls?
>
>
> --- On Tue, 10/8/10, Ralph Castain wrote:
>
st is implemented with multicast calls but does it use any
actual broadcast calls at all?
I
know I'm scraping the edges here looking for something but I just cant get my
head around why it should fail where it has.
--- On Mon, 9/8/10, Ralph Castain wrote:
From: Ralph Castain
Su
presume that bcast is implemented with multicast calls but does it use any
> actual broadcast calls at all?
> I know I'm scraping the edges here looking for something but I just cant get
> my head around why it should fail where it has.
>
> --- On Mon, 9/8/10, Ralph Castain
Personally, I've been having trouble following the explanations of the
problem. Perhaps it'd be helpful if you gave us an example of how to
reproduce the problem. E.g., short sample code and how you run the
example to produce the problem. The shorter the example, the greater
the odds of reso
Sorry -
I missed the statement that all works when you add sleeps. That probably
rules out any possible error in the way MPI_Bcast was used.
Dick Treumann - MPI Team
IBM Systems & Technology Group
Dept X2ZA / MS P963 -- 2455 South Road -- Poughkeepsie, NY 12601
Tele (845) 433-7846 F
I did not take the time to try to fully understand your approach so this
may sound like a dumb question;
Do you have an MPI_Bcast ROOT process in every MPI_COMM_WORLD and does
every non-ROOT MPI_Bcast call correctly identify the rank of ROOT in its
MPI_COMM_WORLD ?
An MPI_Bcast call when the
o or more copies are run at [exactly] the
> same time.
>
> Has anyone else seen similar behavior in concurrently running
> programs that perform lots of broadcasts perhaps?
>
> Randolph
>
>
> --- On Sun, 8/8/10, David Zhang wrote:
>
> From: David Zhang Subject: R
why it should fail where it has.
--- On Mon, 9/8/10, Ralph Castain wrote:
From: Ralph Castain
Subject: Re: [OMPI users] MPI_Bcast issue
To: "Open MPI Users"
Received: Monday, 9 August, 2010, 1:32 PM
Hi Randolph
Unless your code is doing a connect/accept between the copies, there
ram waits on broadcast reception forever when two or
> more copies are run at [exactly] the same time.
>
> Has anyone else seen similar behavior in concurrently running programs that
> perform lots of broadcasts perhaps?
>
> Randolph
>
>
> --- On Sun, 8/8/10, David
copies are run at [exactly] the same time.
Has anyone else seen similar behavior in concurrently running programs that
perform lots of broadcasts perhaps?
Randolph
--- On Sun, 8/8/10, David Zhang wrote:
From: David Zhang
Subject: Re: [OMPI users] MPI_Bcast issue
To: "Open MPI Users"
In particular, intercommunicators
On 8/7/10, Aurélien Bouteiller wrote:
> You should consider reading about communicators in MPI.
>
> Aurelien
> --
> Aurelien Bouteiller, Ph.D.
> Innovative Computing Laboratory, The University of Tennessee.
>
> Envoyé de mon iPad
>
> Le Aug 7, 2010 à 1:05, Randol
You should consider reading about communicators in MPI.
Aurelien
--
Aurelien Bouteiller, Ph.D.
Innovative Computing Laboratory, The University of Tennessee.
Envoyé de mon iPad
Le Aug 7, 2010 à 1:05, Randolph Pullen a écrit :
> I seem to be having a problem with MPI_Bcast.
> My massive I/O int
25 matches
Mail list logo