On Tue, Apr 07, 2009 at 03:18:31PM -0400, George Bosilca wrote:
> Steve,
> 
> I spotted a strange value for the mpi_yield_when_idle MCA parameter. 1  
> means your processor is oversubscribed, and this trigger a call to  
> sched_yield after each check on the SM. Are you running the job  
> oversubscribed? If not it looks like somehow we don't correctly  
> identify that there are multiple cores ...
> 
>   george.
> 

The node is not oversubscribed.  Here's top(1) output

last pid: 90265;  load averages:  0.79,  0.40,  0.28   up 89+18:30:27  12:33:36
31 processes:  3 running, 28 sleeping
CPU:  2.3% user,  0.0% nice, 25.8% system,  0.4% interrupt, 71.5% idle
Mem: 27M Active, 28G Inact, 748M Wired, 1304M Cache, 617M Buf, 840M Free
Swap: 4096M Total, 4096M Free

  PID USERNAME    THR PRI NICE   SIZE    RES STATE   C   TIME   WCPU COMMAND
90264 kargl         1 101    0   170M 12012K CPU1    1   0:32 98.46% z
90265 kargl         1 101    0   170M  4320K CPU2    2   0:32 97.01% z
  756 root          1   4    0  4668K   928K -       7   8:36  0.00% nfsd
  757 root          1   4    0  4668K   932K -       7   7:57  0.00% nfsd


z is the netpipe executable.  This node has 2 quad-core opteron
processors.

I also see the slowdown if I used node19 instead of node20.  19 and
20 are identical blades.


-- 
Steve

Reply via email to