On 14-Jul-09, at 5:14 PM, Robert Kubrick wrote:

Jody,

Just to make sure, you did set processor affinity during your test right?

I'm not sure what that means in the context of OS X.

Hyperthreading was turned on.

Cheers,  Jody


On Jul 13, 2009, at 9:28 PM, Klymak Jody wrote:

Hi Robert,

I got inspired by your question to run a few more tests. They are crude, and I don't have actual cpu timing information because of a library mismatch. However:

Setup:
Xserve, 2x2.26 GHz Quad-core Intel Xeon
6.0 Gb memory 1067 MHz DDR3
Mac OS X 10.5.6

Nodes are connected with a dedicated gigabit ethernet switch.

I'm running the MITgcm, a nonhydrostatic global circulation model. The grid size is modest: 10x150x1600, so bear that in mind. Message passing is on the dimension that is 150x10, and typically is 3 grid cells in either direction. I'm not sure how many variables are passed, but I would guess on the order of 24.

I turned off all the I/O I knew of to reduce disk latency.

1 node:  8 processes:              54 minutes
1 node: 16 processes:             40 minutes (oversubscribed)
2 nodes, 16 processes:            29 minutes

So, oversubscribing was faster (in this case), but it didn't double the speed. Certainly spreading the load to another node was much faster.

I haven't had a chance to implement Warner's suggestion of turning hyperthreading off to see what affect that has on the speed.

Cheers,  Jody
_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users

_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to