On Mon, Jun 16, 2008 at 9:46 PM, Brock Palen <bro...@umich.edu> wrote:

> Brad just curious.
> Did you tweak any other values for starting and running a job on such a
> large system?  You say unmodified,  but OpenMPI lets you tweak many values
> at runtime.
>

Ahh...sorry for the confusion.  By "unmodified", I meant at the source code
level.  The version of Open MPI used was from a straight checkout of the
trunk and did not have any special code modifications for Roadrunner.  We
certainly did change some of the default run-time settings.


> I would be curious to expand what I know from what you discovered.
>

The main run-time parameters used are as follows:

- For scalable process launch
  routed = binomial

- To reduce the overhead for IB connections:
  btl_openib_receive_queues =
P,128,64,32,32,32:S,2048,1024,128,32:S,12288,1024,128,32:S,65536,1024,128,32

- Bind processes to computational cores:
  opal_paffinity_alone = 1

- Leave buffers pinned and registered
  mpi_leave_pinned = 1

--brad



>
> Brock Palen
> www.umich.edu/~brockp <http://www.umich.edu/%7Ebrockp>
> Center for Advanced Computing
> bro...@umich.edu
> (734)936-1985
>
>
>
> On Jun 16, 2008, at 10:12 PM, Brad Benton wrote:
>
> Greetings Open MPI users; we thought you'd be interested in the
> following announcement...
>
> A new supercomputer, powered by Open MPI, has broken the petaflop
> barrier to become the world's fastest supercomputer.  The
> "Roadrunner" system was jointly developed by Los Alamos National
> Laboratories and IBM.  Roadrunner's design uses a cluster of AMD
> dual-core processors coupled with computational accelerators based
> on the IBM Cell Broadband Engine.  The cluster consists of 3,060
> nodes, each of which has 2 dual-core AMD processors and associated
> Cell accelerators.  The AMD nodes are connected with 4x DDR
> InfiniBand links.
>
> Open MPI was used as the communications library for the 12,240
> processes comprising the Linpack run which broke the Petaflop
> barrier at 1.026 Petaflop/s.  The version of Open MPI used in the
> run-for-record was a pre-release version of the upcoming 1.3
> release.  Enhancements in this release include modifications for
> efficient, scalable process launch.  As such, Open MPI was run
> unmodified from a snapshot of the pre-1.3 source base (meaning:
> there are no Roadrunner-specific enhancements that are unportable to
> other environments -- all Open MPI users benefit from the
> scalability and performance improvements contributed by the
> Roadrunner project).
>
> --Brad Benton
> Open MPI/Roadrunner Team
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>

Reply via email to