hi,

   We are about to start running GROMACS 4.0.4 with OpenMPI, in 8
nodes, quad core Rocks cluster. We made some tests, without PME and
found two notable things:

* We are getting the best speedup (6) with 2 nodes ( == 8 cores ). I read
the "Speeding Up Parallel GROMACS in High Latency networks" paper, and
thought that the culprit was the switch, but ifconfig shows no retransmits
(neither does ethtool -s or netstat -s). Does version 4 includes the
alltoall patch? Is the paper irrelevant with GROMACS 4?

* When running with the whole cluster ( 8 nodes, 32 cores ), top reports
in any node a 50% system CPU usage. Is that normal? Can it be accounted to
the use of the network? The sys usage gets a bit up when we configured the
Intel NICs with Interrupt Coalescense Off, so I'm tempted to think it is
just OpenMPI hammering the tcp stack, polling for packages.

Thanks in advance,

Dra.M.Florencia Martini

Laboratorio de Fisicoquímica de Membranas Lipídicas y Liposomas

Cátedra de Química General e Inorgánica

Facultad de Farmacia y Bioquímica

Universidad de Buenos Aires

Junín 956 2º (1113)

TE: 54 011 4964-8249 int 24


      Yahoo! Cocina

Encontra las mejores recetas con Yahoo! Cocina.


http://ar.mujer.yahoo.com/cocina/
_______________________________________________
gmx-users mailing list    gmx-users@gromacs.org
http://lists.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to gmx-users-requ...@gromacs.org.
Can't post? Read http://www.gromacs.org/mailing_lists/users.php

Reply via email to