Hi Looks like your system simply runs out of memory. So power cycling nodes isnt needed. If your cluster runs linux then it already has OOM Killer that will kill processes that runs out of memory. Also having swap on nodes is a good idea even with huge amount of memory. Memory usage for mpi processes will strongly depend on mpi implentation because some of them are usualy caching slave process memory (like usualy do mvapich2)
So can you provide info about youre cluster setup. OS version (including kernel version) uname -a mpi version mpirun --version or mpiexec --version also compiler version that was used for compiling gromacs On Четверг 04 марта 2010 03:15:53 Amit Choubey wrote: > Hi Roland, > > I was using 32 nodes with 8 cores, each with 16 Gb memory. The system was > about 154 M particles. This should be feasible according to the numbers. > Assuming that it takes 50bytes per atoms and 1.76 KB per atom per core then > > Masternode -> (50*154 M + 8*1.06)bytes ~ 16GB (There is no leverage here) > All other nodes 8*1.06 ~ 8.5 GB > > I am planning to try the same run on 64 nodes with 8 cores each again but > not until i am a little more confident. The problem is if gromacs crashes > due to memory it makes the nodes to hang and people have to recycle the > power supply. > > > Thank you, > -- Best Regards, Alexey 'Alexxy' Shvetsov Petersburg Nuclear Physics Institute, Russia Department of Molecular and Radiation Biophysics Gentoo Team Ru Gentoo Linux Dev mailto:alexx...@gmail.com mailto:ale...@gentoo.org mailto:ale...@omrb.pnpi.spb.ru
signature.asc
Description: This is a digitally signed message part.
-- gmx-users mailing list gmx-users@gromacs.org http://lists.gromacs.org/mailman/listinfo/gmx-users Please search the archive at http://www.gromacs.org/search before posting! Please don't post (un)subscribe requests to the list. Use the www interface or send it to gmx-users-requ...@gromacs.org. Can't post? Read http://www.gromacs.org/mailing_lists/users.php