On Thu, 11 Mar 2010 12:44:01 -0500, "Cole, Derek E" <derek.e.c...@lmco.com> 
wrote:
> I am replying to this via the daily-digest message I got. Sorry it
> wasn't sooner... I didn't realize I was getting replies until I got
> the digest. Does anyone know how to change it so I get the emails as
> you all send them?

Log in at the bottom and edit options:

  http://www.open-mpi.org/mailman/listinfo.cgi/users

> I am doing a Red-black Gauss-Seidel algorithm.

Note that red-black Guss-Seidel is a terrible algorithm on cache-based
hardware, it only makes sense on vector hardware.  The reason for this
is that the whole point is to approximate a dense action (the inverse of
a matrix), but the red-black ordering causes this action to be purely
local.  A sequential ordering, on the other hand, is like a dense
lower-triangular operation, which tends to be much closer to a real
inverse.  In parallel, you do these sequential sweeps on each process,
and communicate when you're done.  The memory performance will be twice
as good, and the algorithm will converge in fewer iterations.

> The ghost points are what I was trying to figure for moving this into
> the 3rd dimension. Thanks for adding some concrete-ness to my idea of
> exactly how much overhead is involved. The test domains I am computing
> on are on the order of 100*50*50 or so. This is why I am trying to
> limit the overhead of the MPI communication. I am in the process of
> finding out exactly how big the domains may become, so that I can
> adjust the algorithm accordingly.

The difficulty is for small subdomains.  If you have large subdomains,
then parallel overhead will almost always be small.

> If I understand what you mean by pencils versus books, I don't think
> that will work for these types of calculations will it? Maybe a better
> explanation of what you mean by a pencil versus a book. If I was going
> to solve a sub-matrix of the XY plane for all Z-values, what is that
> considered?

That would be a "book" or "slab".

I still recommend using PETSc rather than reproducing standard code to
call MPI directly for this, it will handle all the decomposition and
updates, and has the advantage that you'll be able to use much better
algorithms than Gauss-Seidel.

Jed

Reply via email to