hi,
We are about to start running GROMACS 4.0.4 with OpenMPI, in 8
nodes, quad core Rocks cluster. We made some tests, without PME and
found two notable things:
* We are getting the best speedup (6) with 2 nodes ( == 8 cores ). I read
the "Speeding Up Parallel GROMACS in High Latency networks" paper, and
thought that the culprit was the switch, but ifconfig shows no retransmits
(neither does ethtool -s or netstat -s). Does version 4 includes the
alltoall patch? Is the paper irrelevant with GROMACS 4?
* When running with the whole cluster ( 8 nodes, 32 cores ), top reports
in any node a 50% system CPU usage. Is that normal? Can it be accounted to
the use of the network? The sys usage gets a bit up when we configured the
Intel NICs with Interrupt Coalescense Off, so I'm tempted to think it is
just OpenMPI hammering the tcp stack, polling for packages.
Thanks in advance,
Dra.M.Florencia Martini
Laboratorio de Fisicoquímica de Membranas Lipídicas y Liposomas
Cátedra de Química General e Inorgánica
Facultad de Farmacia y Bioquímica
Universidad de Buenos Aires
Junín 956 2º (1113)
TE: 54 011 4964-8249 int 24
Yahoo! Cocina
Encontra las mejores recetas con Yahoo! Cocina.
http://ar.mujer.yahoo.com/cocina/
_______________________________________________
gmx-users mailing list gmx-users@gromacs.org
http://lists.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the
www interface or send it to gmx-users-requ...@gromacs.org.
Can't post? Read http://www.gromacs.org/mailing_lists/users.php