f Squyres wrote:
From: Jeff Squyres
Subject: Re: [OMPI users] MPI_Bcast issue
To: "Open MPI Users"
Received: Friday, 13 August, 2010, 3:03 AM
Dick / all --
I just had a phone call with Ralph Castain who has had some additional off-list
mails with Randolph. Apparently, none of us u
- MPI Team
> IBM Systems & Technology Group
> Dept X2ZA / MS P963 -- 2455 South Road -- Poughkeepsie, NY 12601
> Tele (845) 433-7846 Fax (845) 433-8363
>
>
> users-boun...@open-mpi.org wrote on 08/11/2010 08:59:16 PM:
>
> > [image removed]
> >
n...@open-mpi.org wrote on 08/11/2010 08:59:16 PM:
> [image removed]
>
> Re: [OMPI users] MPI_Bcast issue
>
> Randolph Pullen
>
> to:
>
> Open MPI Users
>
> 08/11/2010 09:01 PM
>
> Sent by:
>
> users-boun...@open-mpi.org
>
> Please respon
Interesting point.
--- On Thu, 12/8/10, Ashley Pittman wrote:
From: Ashley Pittman
Subject: Re: [OMPI users] MPI_Bcast issue
To: "Open MPI Users"
Received: Thursday, 12 August, 2010, 12:22 AM
On 11 Aug 2010, at 05:10, Randolph Pullen wrote:
> Sure, but broadcasts are faster -
question is why.
--- On Wed, 11/8/10, Richard Treumann wrote:
From: Richard Treumann
Subject: Re: [OMPI users] MPI_Bcast issue
To: "Open MPI Users"
Received: Wednesday, 11 August, 2010, 11:34 PM
Randolf
I am confused about using multiple,
concurrent mpirun operations. If there are
On 11 Aug 2010, at 05:10, Randolph Pullen wrote:
> Sure, but broadcasts are faster - less reliable apparently, but much faster
> for large clusters.
Going off-topic here but I think it's worth saying:
If you have a dataset that requires collective communication then use the
function call that
On Aug 11, 2010, at 12:10 AM, Randolph Pullen wrote:
> Sure, but broadcasts are faster - less reliable apparently, but much faster
> for large clusters.
Just to be totally clear: MPI_BCAST is defined to be "reliable", in the sense
that it will complete or invoke an error (vs. unreliable data
On Aug 11, 2010, at 9:54 AM, Jeff Squyres wrote:
> (I'll say that OMPI's ALLGATHER algorithm is probably not well optimized for
> massive data transfers like you describe)
Wrong wrong wrong -- I should have checked the code before sending. I made the
incorrect assumption that OMPI still only h
On Aug 10, 2010, at 10:09 PM, Randolph Pullen wrote:
> Jeff thanks for the clarification,
> What I am trying to do is run N concurrent copies of a 1 to N data movement
> program to affect an N to N solution. The actual mechanism I am using is to
> spawn N copies of mpirun from PVM across the cl
Randolf
I am confused about using multiple, concurrent mpirun operations. If
there are M uses of mpirun and each starts N tasks (carried out under pvm
or any other way) I would expect you to have M completely independent MPI
jobs with N tasks (processes) each. You could have some root in eac
wrote:
From: Terry Frankcombe
Subject: Re: [OMPI users] MPI_Bcast issue
To: "Open MPI Users"
Received: Wednesday, 11 August, 2010, 1:57 PM
On Tue, 2010-08-10 at 19:09 -0700, Randolph Pullen wrote:
> Jeff thanks for the clarification,
> What I am trying to do is run N concurre
On Tue, 2010-08-10 at 19:09 -0700, Randolph Pullen wrote:
> Jeff thanks for the clarification,
> What I am trying to do is run N concurrent copies of a 1 to N data
> movement program to affect an N to N solution.
I'm no MPI guru, nor do I completely understand what you are doing, but
isn't this an
rom: Jeff Squyres
Subject: Re: [OMPI users] MPI_Bcast issue
To: "Open MPI Users"
Received: Wednesday, 11 August, 2010, 6:24 AM
+1 on Eugene's comment that I don't fully understand what you are trying to do.
Can you send a short example code?
Some random points:
- Edgar alre
rote:
> The install was completly vanilla - no extras a plain .configure command line
> (on FC10 x8x_64 linux)
>
> Are you saying that all broadcast calls are actually implemented as serial
> point to point calls?
>
>
> --- On Tue, 10/8/10, Ralph Castain wrote:
>
st is implemented with multicast calls but does it use any
actual broadcast calls at all?
I
know I'm scraping the edges here looking for something but I just cant get my
head around why it should fail where it has.
--- On Mon, 9/8/10, Ralph Castain wrote:
From: Ralph Castain
Su
presume that bcast is implemented with multicast calls but does it use any
> actual broadcast calls at all?
> I know I'm scraping the edges here looking for something but I just cant get
> my head around why it should fail where it has.
>
> --- On Mon, 9/8/10, Ralph Castain
he odds of resolution.
From:
Randolph Pullen
To:
us...@open-mpi.org
Date:
08/07/2010 01:23 AM
Subject:
[OMPI users]
7846 Fax (845) 433-8363
From:
Randolph Pullen
To:
us...@open-mpi.org
Date:
08/07/2010 01:23 AM
Subject:
[OMPI users] MPI_Bcast issue
Sent by:
users-boun...@open-mpi.org
I seem to be having a problem with MPI_Bcast.
My massive I/O intensive data movement program must broadcast from n to n
n
llen
To:
us...@open-mpi.org
Date:
08/07/2010 01:23 AM
Subject:
[OMPI users] MPI_Bcast issue
Sent by:
users-boun...@open-mpi.org
I seem to be having a problem with MPI_Bcast.
My massive I/O intensive data movement program must broadcast from n to n
nodes. My problem starts because I requi
o or more copies are run at [exactly] the
> same time.
>
> Has anyone else seen similar behavior in concurrently running
> programs that perform lots of broadcasts perhaps?
>
> Randolph
>
>
> --- On Sun, 8/8/10, David Zhang wrote:
>
> From: David Zhang Subject: R
why it should fail where it has.
--- On Mon, 9/8/10, Ralph Castain wrote:
From: Ralph Castain
Subject: Re: [OMPI users] MPI_Bcast issue
To: "Open MPI Users"
Received: Monday, 9 August, 2010, 1:32 PM
Hi Randolph
Unless your code is doing a connect/accept between the copies, there
ram waits on broadcast reception forever when two or
> more copies are run at [exactly] the same time.
>
> Has anyone else seen similar behavior in concurrently running programs that
> perform lots of broadcasts perhaps?
>
> Randolph
>
>
> --- On Sun, 8/8/10, David
copies are run at [exactly] the same time.
Has anyone else seen similar behavior in concurrently running programs that
perform lots of broadcasts perhaps?
Randolph
--- On Sun, 8/8/10, David Zhang wrote:
From: David Zhang
Subject: Re: [OMPI users] MPI_Bcast issue
To: "Open MPI Users"
In particular, intercommunicators
On 8/7/10, Aurélien Bouteiller wrote:
> You should consider reading about communicators in MPI.
>
> Aurelien
> --
> Aurelien Bouteiller, Ph.D.
> Innovative Computing Laboratory, The University of Tennessee.
>
> Envoyé de mon iPad
>
> Le Aug 7, 2010 à 1:05, Randol
You should consider reading about communicators in MPI.
Aurelien
--
Aurelien Bouteiller, Ph.D.
Innovative Computing Laboratory, The University of Tennessee.
Envoyé de mon iPad
Le Aug 7, 2010 à 1:05, Randolph Pullen a écrit :
> I seem to be having a problem with MPI_Bcast.
> My massive I/O int
I seem to be having a problem with MPI_Bcast.
My massive I/O intensive data movement program must broadcast from n to n
nodes. My problem starts because I require 2 processes per node, a sender and a
receiver and I have implemented these using MPI processes rather than tackle
the complexities of
26 matches
Mail list logo