On Fri, Dec 01, 2006 at 09:35:09AM -0500, Brock Palen wrote:
> On Dec 1, 2006, at 9:23 AM, Gleb Natapov wrote:
> 
> > On Fri, Dec 01, 2006 at 04:14:31PM +0200, Gleb Natapov wrote:
> >> On Fri, Dec 01, 2006 at 11:51:24AM +0100, Peter Kjellstrom wrote:
> >>> On Saturday 25 November 2006 15:31, shap...@isp.nsc.ru wrote:
> >>>> Hello,
> >>>> i cant figure out, is there a way with open-mpi to bind all
> >>>> threads on a given node to a specified subset of CPUs.
> >>>> For example, on a multi-socket multi-core machine, i want to use
> >>>> only a single core on each CPU.
> >>>> Thank You.
> >>>
> >>> This might be a bit naive but, if you spawn two procs on a dual  
> >>> core dual
> >>> socket system then the linux kernel should automagically schedule  
> >>> them this
> >>> way.
> >>>
> >>> I actually think this applies to most of the situations discussed  
> >>> in this
> >>> thread. Explicitly assigning processes to cores may actually get  
> >>> it wrong
> >>> more often than the normal linux scheduler.
> >>>
> >> If you run two single threaded ranks on the dual core dual socket  
> >> node
> >> you better be placing them on the same core. Shared memory  
> >> communication
> Isn't this only valid with NUMA systems?  (large systems or AMD  
> Opteron)  The intel multicores each must communicate along the bus to  
> the north-bridge and back again.  So all cores have the same path to  
> memory.  Correct me if im wrong.  Though working on this would be  
> good, i dont expect all systems to stick with bus, and more and more  
> will be NUMA in the future.
AFAIK Core 2 Duo has shared L2 cache so shared memory communication should be
much faster if two ranks are on the same socket. But I don't have such a
setup to test the theory.

> 
> On another note for systems that use pbs (and maybe other resource  
> managers)  It gives out the cpus in the hostlist  (hostname/0  
> hostname/1 etc)   Why cant OMPI read that info if its available?
> 
> Im prob totally off on these comments.
> 
> Brock
> 
> > I mean "same socket" here and not "same core" of cause.
> >
> >> will be much faster (especially if two cores shares cache).
> >>
> >>> /Peter (who may be putting a bit too much faith in the linux  
> >>> scheduler...)
> >> Linux scheduler does its best assuming the processes are  
> >> unrelated. This is
> >> not the case with MPI ranks.
> >>
> >> --
> >>                    Gleb.
> >> _______________________________________________
> >> users mailing list
> >> us...@open-mpi.org
> >> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >
> > --
> >                     Gleb.
> > _______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
> >
> >
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

--
                        Gleb.

Reply via email to