Am Dienstag 11 November 2008 12:06:06 schrieb vivek sharma:
> I have also tried scaling gromacs for a number of nodes ....but was not > able to optimize it beyond 20 processor..on 20 nodes i.e. 1 processor per As mentioned before, performance strongly depends on the type of interconnect you're using between your processes. Shared Memory, Ethernet, Infiniband, NumaLink, whatever... I assume you're using ethernet (100/1000 MBit?), you can tune here to some extend as described in: Kutzner, C.; Spoel, D. V. D.; Fechner, M.; Lindahl, E.; Schmitt, U. W.; Groot, B. L. D. & Grubmüller, H. Speeding up parallel GROMACS on high-latency networks Journal of Computational Chemistry, 2007 ...but be aware that principal limitations of ethernet remain. To come around this, you might consider to invest in the interconnect. If you can come out with <16 cores, shared memory nodes will give you the "biggest bang for the buck". Best Martin _______________________________________________ gmx-users mailing list gmx-users@gromacs.org http://www.gromacs.org/mailman/listinfo/gmx-users Please search the archive at http://www.gromacs.org/search before posting! Please don't post (un)subscribe requests to the list. Use the www interface or send it to [EMAIL PROTECTED] Can't post? Read http://www.gromacs.org/mailing_lists/users.php