Dear Szilárd, Many thanks for your reply. I've got following reply from my question from Linux-PowerEdge mailing list. I was wondering which one applies to GROMACS parallel computation (I mean CPU bound, disk bound, etc).
>There shouldn't be any linux compatibility issues with any PowerEdge >system. At Duke we have a large compute cluster using a variety of >PowerEdge blades (including M710's) all running on linux. >What interconnect are you using? And are your jobs memory bound, cpu >bound, disk bound, or network bound? >If your computation is more dependent on the interlink and communication >between the nodes, its more important to worry about your interconnect. >If Inter-node communication is highly important, you may also want to >consider something like the M910. The M910 can be configured with 4 >8-core CPUs, thus giving you 32 NUMA-connected cores. Or 64 logical >processors if your job is one that can benefit from HT. Note that when >going with more cores-per chip, your max clockrate tends to be lower. >As such, its really important to know how your jobs are bound so that >you can order a cluster configuration that'll be best for that job. Cheers, Maryam --- On Tue, 18/1/11, Szilárd Páll <szilard.p...@cbr.su.se> wrote: From: Szilárd Páll <szilard.p...@cbr.su.se> Subject: Re: [gmx-users] Dell PowerEdge M710 with Intel Xeon 5667 processor To: "Discussion list for GROMACS users" <gmx-users@gromacs.org> Received: Tuesday, 18 January, 2011, 10:31 PM Hi, Although the question is a bit fuzzy, I might be able to give you a useful answer. >From what I see on the whitepaper of the Poweredge m710 baldes, among other (not so interesting :) OS-es, Dell provides the options of Red Had or SUSE Linux as factory installed OS-es. If you have any of these, you can rest assured that Gromacs will run just fine -- on a single node. Parallel runs are little bit different story and depends on the interconnect. If you have Infiniband, than you'll have a very good scaling over multiple nodes. This is true especially if it's the I/O cards are the Mellanox QDR-s. Cheers, -- Szilárd On Tue, Jan 18, 2011 at 4:48 PM, Maryam Hamzehee <maryam_h_7...@yahoo.com> wrote: > > Dear list, > > I will appreciate it if I can get your expert opinion on doing > parallel computation (I will use GROMACS and AMBER molecular mechanics > packages and some other programs like CYANA, ARIA and CNS to do > structure calculations based on NMR experimental data) using a cluster based > on Dell PowerEdge M710 with Intel Xeon 5667 processor architecture which > apparently each blade has two quad-core cpus. I was wondering if I can get > some information about LINUX compatibility and parallel computation on this > system. > Cheers, > Maryam > > -- > gmx-users mailing list gmx-users@gromacs.org > http://lists.gromacs.org/mailman/listinfo/gmx-users > Please search the archive at > http://www.gromacs.org/Support/Mailing_Lists/Search before posting! > Please don't post (un)subscribe requests to the list. Use the > www interface or send it to gmx-users-requ...@gromacs.org. > Can't post? Read http://www.gromacs.org/Support/Mailing_Lists -- gmx-users mailing list gmx-users@gromacs.org http://lists.gromacs.org/mailman/listinfo/gmx-users Please search the archive at http://www.gromacs.org/Support/Mailing_Lists/Search before posting! Please don't post (un)subscribe requests to the list. Use the www interface or send it to gmx-users-requ...@gromacs.org. Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
-- gmx-users mailing list gmx-users@gromacs.org http://lists.gromacs.org/mailman/listinfo/gmx-users Please search the archive at http://www.gromacs.org/Support/Mailing_Lists/Search before posting! Please don't post (un)subscribe requests to the list. Use the www interface or send it to gmx-users-requ...@gromacs.org. Can't post? Read http://www.gromacs.org/Support/Mailing_Lists