A thing to look at is how much bandwidth the models require compared to
the CPU load.  You can redline gigabit ethernet with a 1GHz PIII and a
64-bit PCI bus.  Opterons on a decent motherboard will definitely keep a
gigabit line chock full.  With dual-core you get the advantage of very
fast processor-to-processor communication but you'll run the risk of
choking on the ethernet connection.  You might be OK if you can get
dual-ethernet connections on the motherboard and run channel-bonding to
increase the bandwidth, but your switch has to be able to handle it.

Damien

> You might want to take this question over to the Beowulf list -- they
> talk a lot more about cluster configurations than we do -- and/or the
> mm4 and wein2k support lists (since they know the details of those
> applications -- if you're going to have a cluster for a specific set of
> applications, it can be best to get input from the developers who know
> the applications best, and what their communication characteristics
> are).
>
>
>
>> -----Original Message-----
>> From: users-boun...@open-mpi.org
>> [mailto:users-boun...@open-mpi.org] On Behalf Of h...@gurban.org
>> Sent: Wednesday, April 26, 2006 12:23 PM
>> To: us...@open-mpi.org
>> Subject: [OMPI users] which is better: 64x1 or 32x2
>>
>> Hi,
>>
>> I want to build an hpc cluster for running mm5 and wien2k
>> scientific applications for my physics coledge. both of them
>> use MPI.
>>
>> Interconnection between nodes: GigEth (Cisco 24 port GigEth)
>>
>> It seems I have two choices for nodes:
>>  * 32 dual core opteron processors (1 GB ram for each node)
>>  * 64 single core opteron processors (2 GB ram for each node)
>>
>> Which is better (performance & price)?
>>
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>


Reply via email to