Well, bummer -- there goes my theory. According to the hwloc info you posted earlier, this shows that OMPI is binding to the 1st hyperthread on each core; *not* to both hyperthreads on a single core. :-\
It would still be slightly interesting to see if there's any difference when you run with --bind-to-core instead of paffinity_alone. On Jan 7, 2011, at 9:56 AM, Gilbert Grosdidier wrote: > Yes, here it is : > > > mpirun -np 8 --mca mpi_paffinity_alone 1 > > /opt/software/SGI/hwloc/1.1rc6r3028/bin/hwloc-bind --get > 0x00000001 > 0x00000002 > 0x00000004 > 0x00000008 > 0x00000010 > 0x00000020 > 0x00000040 > 0x00000080 > > Gilbert. > > Le 7 janv. 11 à 15:50, Jeff Squyres a écrit : > >> Can you run with np=8? >> >> On Jan 7, 2011, at 9:49 AM, Gilbert Grosdidier wrote: >> >>> Hi Jeff, >>> >>> Thanks for taking care of this. >>> >>> Here is what I got on a worker node: >>> >>>> mpirun --mca mpi_paffinity_alone 1 >>>> /opt/software/SGI/hwloc/1.1rc6r3028/bin/hwloc-bind --get >>> 0x00000001 >>> >>> Is this what is expected, please ? Or should I try yet another command ? >>> >>> Thanks, Regards, Gilbert. >>> >>> >>> >>> Le 7 janv. 11 à 15:35, Jeff Squyres a écrit : >>> >>>> On Jan 6, 2011, at 11:23 PM, Gilbert Grosdidier wrote: >>>> >>>>>> lstopo >>>>> Machine (35GB) >>>>> NUMANode L#0 (P#0 18GB) + Socket L#0 + L3 L#0 (8192KB) >>>>> L2 L#0 (256KB) + L1 L#0 (32KB) + Core L#0 >>>>> PU L#0 (P#0) >>>>> PU L#1 (P#8) >>>>> L2 L#1 (256KB) + L1 L#1 (32KB) + Core L#1 >>>>> PU L#2 (P#1) >>>>> PU L#3 (P#9) >>>>> L2 L#2 (256KB) + L1 L#2 (32KB) + Core L#2 >>>>> PU L#4 (P#2) >>>>> PU L#5 (P#10) >>>>> L2 L#3 (256KB) + L1 L#3 (32KB) + Core L#3 >>>>> PU L#6 (P#3) >>>>> PU L#7 (P#11) >>>> [snip] -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/