On Dec 1, 2006, at 9:23 AM, Gleb Natapov wrote:

On Fri, Dec 01, 2006 at 04:14:31PM +0200, Gleb Natapov wrote:
On Fri, Dec 01, 2006 at 11:51:24AM +0100, Peter Kjellstrom wrote:
On Saturday 25 November 2006 15:31, shap...@isp.nsc.ru wrote:
Hello,
i cant figure out, is there a way with open-mpi to bind all
threads on a given node to a specified subset of CPUs.
For example, on a multi-socket multi-core machine, i want to use
only a single core on each CPU.
Thank You.

This might be a bit naive but, if you spawn two procs on a dual core dual socket system then the linux kernel should automagically schedule them this
way.

I actually think this applies to most of the situations discussed in this thread. Explicitly assigning processes to cores may actually get it wrong
more often than the normal linux scheduler.

If you run two single threaded ranks on the dual core dual socket node you better be placing them on the same core. Shared memory communication
Isn't this only valid with NUMA systems? (large systems or AMD Opteron) The intel multicores each must communicate along the bus to the north-bridge and back again. So all cores have the same path to memory. Correct me if im wrong. Though working on this would be good, i dont expect all systems to stick with bus, and more and more will be NUMA in the future.

On another note for systems that use pbs (and maybe other resource managers) It gives out the cpus in the hostlist (hostname/0 hostname/1 etc) Why cant OMPI read that info if its available?

Im prob totally off on these comments.

Brock

I mean "same socket" here and not "same core" of cause.

will be much faster (especially if two cores shares cache).

/Peter (who may be putting a bit too much faith in the linux scheduler...)
Linux scheduler does its best assuming the processes are unrelated. This is
not the case with MPI ranks.

--
                        Gleb.
_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users

--
                        Gleb.
_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users



Reply via email to