On Wed, Oct 03, 2012 at 03:30:01PM +0200, Mike Galbraith wrote:
> On Wed, 2012-10-03 at 10:13 +0200, Mike Galbraith wrote:
> 
> > Watching all cores instead. 
> > 
> > switch rate ~890KHz                            switch rate ~570KHz
> > NO_TTWU_QUEUE nohz=off                         TTWU_QUEUE nohz=off
> > 5.38%  [kernel]  [k] __schedule                4.81%  [kernel]   [k] 
> > _raw_spin_lock_irqsave
> > 4.29%  [kernel]  [k] _raw_spin_lock_irqsave    3.36%  [kernel]   [k] 
> > __skb_recv_datagram
> > 2.88%  [kernel]  [k] resched_task              2.71%  [kernel]   [k] 
> > copy_user_generic_string
> > 2.60%  [kernel]  [k] copy_user_generic_string  2.67%  [kernel]   [k] 
> > reschedule_interrupt
> > 2.38%  [kernel]  [k] __switch_to               2.62%  [kernel]   [k] 
> > sock_alloc_send_pskb
> > 2.15%  [kernel]  [k] sock_alloc_send_pskb      2.52%  [kernel]   [k] 
> > __schedule
> > 2.08%  [kernel]  [k] __skb_recv_datagram       2.31%  [kernel]   [k] 
> > try_to_wake_up
> > 1.81%  [kernel]  [k] udp_sendmsg               2.14%  [kernel]   [k] 
> > system_call
> > 1.76%  [kernel]  [k] system_call               1.98%  [kernel]   [k] 
> > udp_sendmsg
> > 1.73%  [kernel]  [k] __udp4_lib_lookup         1.96%  [kernel]   [k] 
> > __udp4_lib_lookup
> > 1.65%  [kernel]  [k] __slab_free.isra.42       1.78%  [kernel]   [k] 
> > sock_def_readable
> > 1.62%  [kernel]  [k] try_to_wake_up            1.63%  [kernel]   [k] 
> > __slab_free.isra.42
> > 1.43%  [kernel]  [k] update_rq_clock           1.60%  [kernel]   [k] 
> > __switch_to
> > 1.43%  [kernel]  [k] sock_def_readable         1.52%  [kernel]   [k] 
> > dma_issue_pending_all
> > 1.41%  [kernel]  [k] dma_issue_pending_all     1.48%  [kernel]   [k] 
> > __ip_append_data.isra.35
> > 1.40%  [kernel]  [k] menu_select               1.44%  [kernel]   [k] 
> > _raw_spin_lock
> > 1.36%  [kernel]  [k] finish_task_switch        1.38%  [kernel]   [k] 
> > _raw_spin_unlock_irqrestore
> > 1.30%  [kernel]  [k] ksize                     1.33%  [kernel]   [k] 
> > __udp4_lib_rcv
> > 
> > Strange.
> 
> nohz=off, pipe-test with one half pinned to CPU0, the other to CPU1.
> 
> procs -----------memory---------- ---swap-- -----io---- -system-- 
> -----cpu------
>  r  b   swpd   free   buff  cache   si   so    bi    bo   in   cs us sy id wa 
> st
> TTW_QUEUE
>  1  0      0 3039488  50948 444720    0    0     0     0 539724 1013417  1 15 
> 84  0  0
>  1  0      0 3039488  50956 444720    0    0     0     1 540853 1015679  1 15 
> 84  0  0
>  1  0      0 3039364  50956 444720    0    0     0     0 541630 1017239  1 16 
> 83  0  0
>  2  0      0 3038992  50956 444720    0    0     0     0 335550 1096569  4 20 
> 76  0  0
> NO_TTWU_QUEUE
>  1  0      0 3038992  50956 444720    0    0     0     0 33100 1318984  1 27 
> 71  0  0
>  1  0      0 3038868  50956 444720    0    0     0     0 33100 1319126  2 27 
> 71  0  0
>  1  0      0 3038868  50956 444720    0    0     0     0 33097 1317968  1 27 
> 72  0  0
>  2  0      0 3038868  50964 444720    0    0     0     1 33104 1318558  2 27 
> 71  0  0
> 
> We can switch faster with NO_TTWU_QUEUE, so we switch more, and that
> hurts netperf UDP_STREAM throughput.. somehow.  Fatter is better is not
> the way context switch happy benchmarks usually work.
> 

Do we really switch more though?

Look at the difference in interrupts vs context switch. IPIs are an interrupt
so if TTWU_QUEUE wakes process B using an IPI, does that count as a context
switch? It probably does not get accounted as a context switch even though
it's functionally similar in this case but I'd like to hear confirmation
of that.

If we did assume that these IPIs are effectively context switches then look
at the TTWU_QUEUE figures. There are 530K interrupts versus 33K interrupts
for NO_TTWU_QUEUE. If each one of those IPIs are effectively a context
switch then the actual switch rates are 1.5M switches versus 1.3 switches
and TTWU_QUEUE is actually switching faster.

-- 
Mel Gorman
SUSE Labs
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to