Hi Paul,

On Thu, Dec 18, 2014 at 12:54 AM, Paul E. McKenney
<paul...@linux.vnet.ibm.com> wrote:
> On Tue, Dec 16, 2014 at 11:00:20PM +0530, Arun KS wrote:
>> Hello,
>>
>> Adding some more info.
>>
>> Below is the rcu_data data structure corresponding to cpu4.
>
> This shows that RCU is idle.  What was the state of the system at the
> time you collected this data?

System initiated a suspend sequence and currently at disable_nonboot_cpus().
It has hotplugged 0, 1 and 2 successfully.  And even successful in hot
plugging cpu3.
But while calling the CPU_POST_DEAD notifier for cpu3, another driver tried to
unregister an atomic notifier. Which eventually calls syncronize_rcu()
which hangs the suspend task.

bt as follows,
PID: 202    TASK: edcd2a00  CPU: 4   COMMAND: "kworker/u16:4"
 #0 [<c0a1f8c0>] (__schedule) from [<c0a1d054>]
 #1 [<c0a1d054>] (schedule_timeout) from [<c0a1f018>]
 #2 [<c0a1f018>] (wait_for_common) from [<c013c570>]
 #3 [<c013c570>] (wait_rcu_gp) from [<c014407c>]
 #4 [<c014407c>] (atomic_notifier_chain_unregister) from [<c06be62c>]
 #5 [<c06be62c>] (cpufreq_interactive_disable_sched_input) from [<c06bee1c>]
 #6 [<c06bee1c>] (cpufreq_governor_interactive) from [<c06b7724>]
 #7 [<c06b7724>] (__cpufreq_governor) from [<c06b9f74>]
 #8 [<c06b9f74>] (__cpufreq_remove_dev_finish) from [<c06ba3a4>]
 #9 [<c06ba3a4>] (cpufreq_cpu_callback) from [<c0a22674>]
#10 [<c0a22674>] (notifier_call_chain) from [<c012284c>]
#11 [<c012284c>] (__cpu_notify) from [<c01229dc>]
#12 [<c01229dc>] (cpu_notify_nofail) from [<c0a0dd1c>]
#13 [<c0a0dd1c>] (_cpu_down) from [<c0122b48>]
#14 [<c0122b48>] (disable_nonboot_cpus) from [<c0168cd8>]
#15 [<c0168cd8>] (suspend_devices_and_enter) from [<c0169018>]
#16 [<c0169018>] (pm_suspend) from [<c01691e0>]
#17 [<c01691e0>] (try_to_suspend) from [<c01396f0>]
#18 [<c01396f0>] (process_one_work) from [<c0139db0>]
#19 [<c0139db0>] (worker_thread) from [<c013efa4>]
#20 [<c013efa4>] (kthread) from [<c01061f8>]

But the other cores 4-7 are active. I can see them going to idle tasks
coming out from idle because of interrupts, scheduling kworkers etc.
So when I took the data, all the online cores(4-7) were in idle as
shown below from runq data structures.

------start--------------
crash> runq
CPU 0 [OFFLINE]

CPU 1 [OFFLINE]

CPU 2 [OFFLINE]

CPU 3 [OFFLINE]

CPU 4 RUNQUEUE: c5439040
  CURRENT: PID: 0      TASK: f0c9d400  COMMAND: "swapper/4"
  RT PRIO_ARRAY: c5439130
     [no tasks queued]
  CFS RB_ROOT: c54390c0
     [no tasks queued]

CPU 5 RUNQUEUE: c5447040
  CURRENT: PID: 0      TASK: f0c9aa00  COMMAND: "swapper/5"
  RT PRIO_ARRAY: c5447130
     [no tasks queued]
  CFS RB_ROOT: c54470c0
     [no tasks queued]

CPU 6 RUNQUEUE: c5455040
  CURRENT: PID: 0      TASK: f0c9ce00  COMMAND: "swapper/6"
  RT PRIO_ARRAY: c5455130
     [no tasks queued]
  CFS RB_ROOT: c54550c0
     [no tasks queued]

CPU 7 RUNQUEUE: c5463040
  CURRENT: PID: 0      TASK: f0c9b000  COMMAND: "swapper/7"
  RT PRIO_ARRAY: c5463130
     [no tasks queued]
  CFS RB_ROOT: c54630c0
     [no tasks queued]
------end--------------

but one strange thing i can see is that rcu_read_lock_nesting for idle
tasks running on cpu 5 and cpu 6 are set to 1.

PID: 0      TASK: f0c9d400  CPU: 4   COMMAND: "swapper/4"
  rcu_read_lock_nesting = 0,

PID: 0      TASK: f0c9aa00  CPU: 5   COMMAND: "swapper/5"
  rcu_read_lock_nesting = 1,

PID: 0      TASK: f0c9ce00  CPU: 6   COMMAND: "swapper/6"
  rcu_read_lock_nesting = 1,

PID: 0      TASK: f0c9b000  CPU: 7   COMMAND: "swapper/7"
  rcu_read_lock_nesting = 0,

Does this means that the current grace period(suspend thread is
waiting on) is getting extended infinitely?
Also attaching the per_cpu rcu_data for online and offline cores.

Thanks,
Arun

>
>                                                         Thanx, Paul
>
>> struct rcu_data {
>>   completed = 5877,
>>   gpnum = 5877,
>>   passed_quiesce = true,
>>   qs_pending = false,
>>   beenonline = true,
>>   preemptible = true,
>>   mynode = 0xc117f340 <rcu_preempt_state>,
>>   grpmask = 16,
>>   nxtlist = 0xedaaec00,
>>   nxttail = {0xc54366c4, 0xe84d350c, 0xe84d350c, 0xe84d350c},
>>   nxtcompleted = {4294967035, 5878, 5878, 5878},
>>   qlen_lazy = 105,
>>   qlen = 415,
>>   qlen_last_fqs_check = 0,
>>   n_cbs_invoked = 86323,
>>   n_nocbs_invoked = 0,
>>   n_cbs_orphaned = 0,
>>   n_cbs_adopted = 139,
>>   n_force_qs_snap = 0,
>>   blimit = 10,
>>   dynticks = 0xc5436758,
>>   dynticks_snap = 7582140,
>>   dynticks_fqs = 41,
>>   offline_fqs = 0,
>>   n_rcu_pending = 59404,
>>   n_rp_qs_pending = 5,
>>   n_rp_report_qs = 4633,
>>   n_rp_cb_ready = 32,
>>   n_rp_cpu_needs_gp = 41088,
>>   n_rp_gp_completed = 2844,
>>   n_rp_gp_started = 1150,
>>   n_rp_need_nothing = 9657,
>>   barrier_head = {
>>     next = 0x0,
>>     func = 0x0
>>   },
>>   oom_head = {
>>     next = 0x0,
>>     func = 0x0
>>   },
>>   cpu = 4,
>>   rsp = 0xc117f340 <rcu_preempt_state>
>> }
>>
>>
>>
>> Also pasting complete rcu_preempt_state.
>>
>>
>>
>> rcu_preempt_state = $9 = {
>>   node = {{
>>       lock = {
>>         raw_lock = {
>>           {
>>             slock = 3129850509,
>>             tickets = {
>>               owner = 47757,
>>               next = 47757
>>             }
>>           }
>>         },
>>         magic = 3735899821,
>>         owner_cpu = 4294967295,
>>         owner = 0xffffffff
>>       },
>>       gpnum = 5877,
>>       completed = 5877,
>>       qsmask = 0,
>>       expmask = 0,
>>       qsmaskinit = 240,
>>       grpmask = 0,
>>       grplo = 0,
>>       grphi = 7,
>>       grpnum = 0 '\000',
>>       level = 0 '\000',
>>       parent = 0x0,
>>       blkd_tasks = {
>>         next = 0xc117f378 <rcu_preempt_state+56>,
>>         prev = 0xc117f378 <rcu_preempt_state+56>
>>       },
>>       gp_tasks = 0x0,
>>       exp_tasks = 0x0,
>>       need_future_gp = {1, 0},
>>       fqslock = {
>>         raw_lock = {
>>           {
>>             slock = 0,
>>             tickets = {
>>               owner = 0,
>>               next = 0
>>             }
>>           }
>>         },
>>         magic = 3735899821,
>>         owner_cpu = 4294967295,
>>         owner = 0xffffffff
>>       }
>>     }},
>>   level = {0xc117f340 <rcu_preempt_state>},
>>   levelcnt = {1, 0, 0, 0, 0},
>>   levelspread = "\b",
>>   rda = 0xc115e6b0 <rcu_preempt_data>,
>>   call = 0xc01975ac <call_rcu>,
>>   fqs_state = 0 '\000',
>>   boost = 0 '\000',
>>   gpnum = 5877,
>>   completed = 5877,
>>   gp_kthread = 0xf0c9e600,
>>   gp_wq = {
>>     lock = {
>>       {
>>         rlock = {
>>           raw_lock = {
>>             {
>>               slock = 2160230594,
>>               tickets = {
>>                 owner = 32962,
>>                 next = 32962
>>               }
>>             }
>>           },
>>           magic = 3735899821,
>>           owner_cpu = 4294967295,
>>           owner = 0xffffffff
>>         }
>>       }
>>     },
>>     task_list = {
>>       next = 0xf0cd1f20,
>>       prev = 0xf0cd1f20
>>     }
>>   },
>>   gp_flags = 1,
>>   orphan_lock = {
>>     raw_lock = {
>>       {
>>         slock = 327685,
>>         tickets = {
>>           owner = 5,
>>           next = 5
>>         }
>>       }
>>     },
>>     magic = 3735899821,
>>     owner_cpu = 4294967295,
>>     owner = 0xffffffff
>>   },
>>   orphan_nxtlist = 0x0,
>>   orphan_nxttail = 0xc117f490 <rcu_preempt_state+336>,
>>   orphan_donelist = 0x0,
>>   orphan_donetail = 0xc117f498 <rcu_preempt_state+344>,
>>   qlen_lazy = 0,
>>   qlen = 0,
>>   onoff_mutex = {
>>     count = {
>>       counter = 1
>>     },
>>     wait_lock = {
>>       {
>>         rlock = {
>>           raw_lock = {
>>             {
>>               slock = 811479134,
>>               tickets = {
>>                 owner = 12382,
>>                 next = 12382
>>               }
>>             }
>>           },
>>           magic = 3735899821,
>>           owner_cpu = 4294967295,
>>           owner = 0xffffffff
>>         }
>>       }
>>     },
>>     wait_list = {
>>       next = 0xc117f4bc <rcu_preempt_state+380>,
>>       prev = 0xc117f4bc <rcu_preempt_state+380>
>>     },
>>     owner = 0x0,
>>     name = 0x0,
>>     magic = 0xc117f4a8 <rcu_preempt_state+360>
>>   },
>>   barrier_mutex = {
>>     count = {
>>       counter = 1
>>     },
>>     wait_lock = {
>>       {
>>         rlock = {
>>           raw_lock = {
>>             {
>>               slock = 0,
>>               tickets = {
>>                 owner = 0,
>>                 next = 0
>>               }
>>             }
>>           },
>>           magic = 3735899821,
>>           owner_cpu = 4294967295,
>>           owner = 0xffffffff
>>         }
>>       }
>>     },
>>     wait_list = {
>>       next = 0xc117f4e4 <rcu_preempt_state+420>,
>>       prev = 0xc117f4e4 <rcu_preempt_state+420>
>>     },
>>     owner = 0x0,
>>     name = 0x0,
>>     magic = 0xc117f4d0 <rcu_preempt_state+400>
>>   },
>>   barrier_cpu_count = {
>>     counter = 0
>>   },
>>   barrier_completion = {
>>     done = 0,
>>     wait = {
>>       lock = {
>>         {
>>           rlock = {
>>             raw_lock = {
>>               {
>>                 slock = 0,
>>                 tickets = {
>>                   owner = 0,
>>                   next = 0
>>                 }
>>               }
>>             },
>>             magic = 0,
>>             owner_cpu = 0,
>>             owner = 0x0
>>           }
>>         }
>>       },
>>       task_list = {
>>         next = 0x0,
>>         prev = 0x0
>>       }
>>     }
>>   },
>>   n_barrier_done = 0,
>>   expedited_start = {
>>     counter = 0
>>   },
>>   expedited_done = {
>>     counter = 0
>>   },
>>   expedited_wrap = {
>>     counter = 0
>>   },
>>   expedited_tryfail = {
>>     counter = 0
>>   },
>>   expedited_workdone1 = {
>>     counter = 0
>>   },
>>   expedited_workdone2 = {
>>     counter = 0
>>   },
>>   expedited_normal = {
>>     counter = 0
>>   },
>>   expedited_stoppedcpus = {
>>     counter = 0
>>   },
>>   expedited_done_tries = {
>>     counter = 0
>>   },
>>   expedited_done_lost = {
>>     counter = 0
>>   },
>>   expedited_done_exit = {
>>     counter = 0
>>   },
>>   jiffies_force_qs = 4294963917,
>>   n_force_qs = 4028,
>>   n_force_qs_lh = 0,
>>   n_force_qs_ngp = 0,
>>   gp_start = 4294963911,
>>   jiffies_stall = 4294966011,
>>   gp_max = 17,
>>   name = 0xc0d833ab "rcu_preempt",
>>   abbr = 112 'p',
>>   flavors = {
>>     next = 0xc117f2ec <rcu_bh_state+556>,
>>     prev = 0xc117f300 <rcu_struct_flavors>
>>   },
>>   wakeup_work = {
>>     flags = 3,
>>     llnode = {
>>       next = 0x0
>>     },
>>     func = 0xc0195aa8 <rsp_wakeup>
>>   }
>> }
>>
>> Hope this helps.
>>
>> Thanks,
>> Arun
>>
>>
>> On Tue, Dec 16, 2014 at 11:59 AM, Arun KS <arunks.li...@gmail.com> wrote:
>> > Hello,
>> >
>> > I dig little deeper to understand the situation.
>> > All other cpus are in idle thread already.
>> > As per my understanding, for the grace period to end, at-least one of
>> > the following should happen on all online cpus,
>> >
>> > 1. a context switch.
>> > 2. user space switch.
>> > 3. switch to idle thread.
>> >
>> > In this situation, since all the other cores are already in idle,  non
>> > of the above are meet on all online cores.
>> > So grace period is getting extended and never finishes. Below is the
>> > state of runqueue when the hang happens.
>> > --------------start------------------------------------
>> > crash> runq
>> > CPU 0 [OFFLINE]
>> >
>> > CPU 1 [OFFLINE]
>> >
>> > CPU 2 [OFFLINE]
>> >
>> > CPU 3 [OFFLINE]
>> >
>> > CPU 4 RUNQUEUE: c3192e40
>> >   CURRENT: PID: 0      TASK: f0874440  COMMAND: "swapper/4"
>> >   RT PRIO_ARRAY: c3192f20
>> >      [no tasks queued]
>> >   CFS RB_ROOT: c3192eb0
>> >      [no tasks queued]
>> >
>> > CPU 5 RUNQUEUE: c31a0e40
>> >   CURRENT: PID: 0      TASK: f0874980  COMMAND: "swapper/5"
>> >   RT PRIO_ARRAY: c31a0f20
>> >      [no tasks queued]
>> >   CFS RB_ROOT: c31a0eb0
>> >      [no tasks queued]
>> >
>> > CPU 6 RUNQUEUE: c31aee40
>> >   CURRENT: PID: 0      TASK: f0874ec0  COMMAND: "swapper/6"
>> >   RT PRIO_ARRAY: c31aef20
>> >      [no tasks queued]
>> >   CFS RB_ROOT: c31aeeb0
>> >      [no tasks queued]
>> >
>> > CPU 7 RUNQUEUE: c31bce40
>> >   CURRENT: PID: 0      TASK: f0875400  COMMAND: "swapper/7"
>> >   RT PRIO_ARRAY: c31bcf20
>> >      [no tasks queued]
>> >   CFS RB_ROOT: c31bceb0
>> >      [no tasks queued]
>> > --------------end------------------------------------
>> >
>> > If my understanding is correct the below patch should help, because it
>> > will expedite grace periods during suspend,
>> > https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git/commit/?id=d1d74d14e98a6be740a6f12456c7d9ad47be9c9c
>> >
>> > But I wonder why it was not taken to stable trees. Can we take it?
>> > Appreciate your help.
>> >
>> > Thanks,
>> > Arun
>> >
>> > On Mon, Dec 15, 2014 at 10:34 PM, Arun KS <arunks.li...@gmail.com> wrote:
>> >> Hi,
>> >>
>> >> Here is the backtrace of the process hanging in wait_rcu_gp,
>> >>
>> >> PID: 247    TASK: e16e7380  CPU: 4   COMMAND: "kworker/u16:5"
>> >>  #0 [<c09fead0>] (__schedule) from [<c09fcab0>]
>> >>  #1 [<c09fcab0>] (schedule_timeout) from [<c09fe050>]
>> >>  #2 [<c09fe050>] (wait_for_common) from [<c013b2b4>]
>> >>  #3 [<c013b2b4>] (wait_rcu_gp) from [<c0142f50>]
>> >>  #4 [<c0142f50>] (atomic_notifier_chain_unregister) from [<c06b2ab8>]
>> >>  #5 [<c06b2ab8>] (cpufreq_interactive_disable_sched_input) from 
>> >> [<c06b32a8>]
>> >>  #6 [<c06b32a8>] (cpufreq_governor_interactive) from [<c06abbf8>]
>> >>  #7 [<c06abbf8>] (__cpufreq_governor) from [<c06ae474>]
>> >>  #8 [<c06ae474>] (__cpufreq_remove_dev_finish) from [<c06ae8c0>]
>> >>  #9 [<c06ae8c0>] (cpufreq_cpu_callback) from [<c0a0185c>]
>> >> #10 [<c0a0185c>] (notifier_call_chain) from [<c0121888>]
>> >> #11 [<c0121888>] (__cpu_notify) from [<c0121a04>]
>> >> #12 [<c0121a04>] (cpu_notify_nofail) from [<c09ee7f0>]
>> >> #13 [<c09ee7f0>] (_cpu_down) from [<c0121b70>]
>> >> #14 [<c0121b70>] (disable_nonboot_cpus) from [<c016788c>]
>> >> #15 [<c016788c>] (suspend_devices_and_enter) from [<c0167bcc>]
>> >> #16 [<c0167bcc>] (pm_suspend) from [<c0167d94>]
>> >> #17 [<c0167d94>] (try_to_suspend) from [<c0138460>]
>> >> #18 [<c0138460>] (process_one_work) from [<c0138b18>]
>> >> #19 [<c0138b18>] (worker_thread) from [<c013dc58>]
>> >> #20 [<c013dc58>] (kthread) from [<c01061b8>]
>> >>
>> >> Will this patch helps here,
>> >> https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git/commit/?id=d1d74d14e98a6be740a6f12456c7d9ad47be9c9c
>> >>
>> >> I couldn't really understand why it got struck in  synchronize_rcu().
>> >> Please give some pointers to debug this further.
>> >>
>> >> Below are the configs enable related to RCU.
>> >>
>> >> CONFIG_TREE_PREEMPT_RCU=y
>> >> CONFIG_PREEMPT_RCU=y
>> >> CONFIG_RCU_STALL_COMMON=y
>> >> CONFIG_RCU_FANOUT=32
>> >> CONFIG_RCU_FANOUT_LEAF=16
>> >> CONFIG_RCU_FAST_NO_HZ=y
>> >> CONFIG_RCU_CPU_STALL_TIMEOUT=21
>> >> CONFIG_RCU_CPU_STALL_VERBOSE=y
>> >>
>> >> Kernel version is 3.10.28
>> >> Architecture is ARM
>> >>
>> >> Thanks,
>> >> Arun
>>
>
crash> struct rcu_data C54286B0
struct rcu_data {
  completed = 2833,
  gpnum = 2833,
  passed_quiesce = false,
  qs_pending = false,
  beenonline = true,
  preemptible = true,
  mynode = 0xc117f340 <rcu_preempt_state>,
  grpmask = 8,
  nxtlist = 0x0,
  nxttail = {0xc54286c4, 0xc54286c4, 0xc54286c4, 0x0},
  nxtcompleted = {0, 4294967136, 4294967137, 4294967137},
  qlen_lazy = 0,
  qlen = 0,
  qlen_last_fqs_check = 0,
  n_cbs_invoked = 609,
  n_nocbs_invoked = 0,
  n_cbs_orphaned = 13,
  n_cbs_adopted = 0,
  n_force_qs_snap = 1428,
  blimit = 10,
  dynticks = 0xc5428758,
  dynticks_snap = 13206053,
  dynticks_fqs = 16,
  offline_fqs = 0,
  n_rcu_pending = 181,
  n_rp_qs_pending = 1,
  n_rp_report_qs = 21,
  n_rp_cb_ready = 0,
  n_rp_cpu_needs_gp = 0,
  n_rp_gp_completed = 22,
  n_rp_gp_started = 8,
  n_rp_need_nothing = 130,
  barrier_head = {
    next = 0x0,
    func = 0x0
  },
  oom_head = {
    next = 0x0,
    func = 0x0
  },
  cpu = 3,
  rsp = 0xc117f340 <rcu_preempt_state>
}

crash> struct rcu_data C541A6B0
struct rcu_data {
  completed = 5877,
  gpnum = 5877,
  passed_quiesce = true,
  qs_pending = false,
  beenonline = true,
  preemptible = true,
  mynode = 0xc117f340 <rcu_preempt_state>,
  grpmask = 4,
  nxtlist = 0x0,
  nxttail = {0xc541a6c4, 0xc541a6c4, 0xc541a6c4, 0x0},
  nxtcompleted = {0, 5877, 5878, 5878},
  qlen_lazy = 0,
  qlen = 0,
  qlen_last_fqs_check = 0,
  n_cbs_invoked = 61565,
  n_nocbs_invoked = 0,
  n_cbs_orphaned = 139,
  n_cbs_adopted = 100,
  n_force_qs_snap = 0,
  blimit = 10,
  dynticks = 0xc541a758,
  dynticks_snap = 13901017,
  dynticks_fqs = 75,
  offline_fqs = 0,
  n_rcu_pending = 16546,
  n_rp_qs_pending = 3,
  n_rp_report_qs = 4539,
  n_rp_cb_ready = 69,
  n_rp_cpu_needs_gp = 782,
  n_rp_gp_completed = 4196,
  n_rp_gp_started = 1739,
  n_rp_need_nothing = 5221,
  barrier_head = {
    next = 0x0,
    func = 0x0
  },
  oom_head = {
    next = 0x0,
    func = 0x0
  },
  cpu = 2,
  rsp = 0xc117f340 <rcu_preempt_state>
}

crash> struct rcu_data C540C6B0
struct rcu_data {
  completed = 5877,
  gpnum = 5877,
  passed_quiesce = true,
  qs_pending = false,
  beenonline = true,
  preemptible = true,
  mynode = 0xc117f340 <rcu_preempt_state>,
  grpmask = 2,
  nxtlist = 0x0,
  nxttail = {0xc540c6c4, 0xc540c6c4, 0xc540c6c4, 0x0},
  nxtcompleted = {4294967030, 5878, 5878, 5878},
  qlen_lazy = 0,
  qlen = 0,
  qlen_last_fqs_check = 0,
  n_cbs_invoked = 74292,
  n_nocbs_invoked = 0,
  n_cbs_orphaned = 100,
  n_cbs_adopted = 65,
  n_force_qs_snap = 0,
  blimit = 10,
  dynticks = 0xc540c758,
  dynticks_snap = 10753433,
  dynticks_fqs = 69,
  offline_fqs = 0,
  n_rcu_pending = 18350,
  n_rp_qs_pending = 6,
  n_rp_report_qs = 5009,
  n_rp_cb_ready = 50,
  n_rp_cpu_needs_gp = 915,
  n_rp_gp_completed = 4423,
  n_rp_gp_started = 1826,
  n_rp_need_nothing = 6127,
  barrier_head = {
    next = 0x0,
    func = 0x0
  },
  oom_head = {
    next = 0x0,
    func = 0x0
  },
  cpu = 1,
  rsp = 0xc117f340 <rcu_preempt_state>
}
crash> struct rcu_data C53FE6B0
struct rcu_data {
  completed = 5877,
  gpnum = 5877,
  passed_quiesce = true,
  qs_pending = false,
  beenonline = true,
  preemptible = true,
  mynode = 0xc117f340 <rcu_preempt_state>,
  grpmask = 1,
  nxtlist = 0x0,
  nxttail = {0xc53fe6c4, 0xc53fe6c4, 0xc53fe6c4, 0x0},
  nxtcompleted = {4294966997, 5875, 5876, 5876},
  qlen_lazy = 0,
  qlen = 0,
  qlen_last_fqs_check = 0,
  n_cbs_invoked = 123175,
  n_nocbs_invoked = 0,
  n_cbs_orphaned = 52,
  n_cbs_adopted = 0,
  n_force_qs_snap = 0,
  blimit = 10,
  dynticks = 0xc53fe758,
  dynticks_snap = 6330446,
  dynticks_fqs = 46,
  offline_fqs = 0,
  n_rcu_pending = 22529,
  n_rp_qs_pending = 3,
  n_rp_report_qs = 5290,
  n_rp_cb_ready = 279,
  n_rp_cpu_needs_gp = 740,
  n_rp_gp_completed = 2707,
  n_rp_gp_started = 1208,
  n_rp_need_nothing = 12305,
  barrier_head = {
    next = 0x0,
    func = 0x0
  },
  oom_head = {
    next = 0x0,
    func = 0x0
  },
  cpu = 0,
  rsp = 0xc117f340 <rcu_preempt_state>
}

crash> struct rcu_data c54366b0
struct rcu_data {
  completed = 5877,
  gpnum = 5877,
  passed_quiesce = true,
  qs_pending = false,
  beenonline = true,
  preemptible = true,
  mynode = 0xc117f340 <rcu_preempt_state>,
  grpmask = 16,
  nxtlist = 0xedaaec00,
  nxttail = {0xc54366c4, 0xe84d350c, 0xe84d350c, 0xe84d350c},
  nxtcompleted = {4294967035, 5878, 5878, 5878},
  qlen_lazy = 105,
  qlen = 415,
  qlen_last_fqs_check = 0,
  n_cbs_invoked = 86323,
  n_nocbs_invoked = 0,
  n_cbs_orphaned = 0,
  n_cbs_adopted = 139,
  n_force_qs_snap = 0,
  blimit = 10,
  dynticks = 0xc5436758,
  dynticks_snap = 7582140,
  dynticks_fqs = 41,
  offline_fqs = 0,
  n_rcu_pending = 59404,
  n_rp_qs_pending = 5,
  n_rp_report_qs = 4633,
  n_rp_cb_ready = 32,
  n_rp_cpu_needs_gp = 41088,
  n_rp_gp_completed = 2844,
  n_rp_gp_started = 1150,
  n_rp_need_nothing = 9657,
  barrier_head = {
    next = 0x0,
    func = 0x0
  },
  oom_head = {
    next = 0x0,
    func = 0x0
  },
  cpu = 4,
  rsp = 0xc117f340 <rcu_preempt_state>
}

crash> struct rcu_data c54446b0
struct rcu_data {
  completed = 5877,
  gpnum = 5877,
  passed_quiesce = true,
  qs_pending = false,
  beenonline = true,
  preemptible = true,
  mynode = 0xc117f340 <rcu_preempt_state>,
  grpmask = 32,
  nxtlist = 0xcf9e856c,
  nxttail = {0xc54446c4, 0xcfb3050c, 0xcfb3050c, 0xcfb3050c},
  nxtcompleted = {0, 5878, 5878, 5878},
  qlen_lazy = 0,
  qlen = 117,
  qlen_last_fqs_check = 0,
  n_cbs_invoked = 36951,
  n_nocbs_invoked = 0,
  n_cbs_orphaned = 0,
  n_cbs_adopted = 0,
  n_force_qs_snap = 1428,
  blimit = 10,
  dynticks = 0xc5444758,
  dynticks_snap = 86034,
  dynticks_fqs = 46,
  offline_fqs = 0,
  n_rcu_pending = 49104,
  n_rp_qs_pending = 3,
  n_rp_report_qs = 2360,
  n_rp_cb_ready = 18,
  n_rp_cpu_needs_gp = 40106,
  n_rp_gp_completed = 1334,
  n_rp_gp_started = 791,
  n_rp_need_nothing = 4495,
  barrier_head = {
    next = 0x0,
    func = 0x0
  },
  oom_head = {
    next = 0x0,
    func = 0x0
  },
  cpu = 5,
  rsp = 0xc117f340 <rcu_preempt_state>
}

crash> struct rcu_data c54526b0
struct rcu_data {
  completed = 5877,
  gpnum = 5877,
  passed_quiesce = true,
  qs_pending = false,
  beenonline = true,
  preemptible = true,
  mynode = 0xc117f340 <rcu_preempt_state>,
  grpmask = 64,
  nxtlist = 0xe613d200,
  nxttail = {0xc54526c4, 0xe6fc9d0c, 0xe6fc9d0c, 0xe6fc9d0c},
  nxtcompleted = {0, 5878, 5878, 5878},
  qlen_lazy = 2,
  qlen = 35,
  qlen_last_fqs_check = 0,
  n_cbs_invoked = 34459,
  n_nocbs_invoked = 0,
  n_cbs_orphaned = 0,
  n_cbs_adopted = 0,
  n_force_qs_snap = 1428,
  blimit = 10,
  dynticks = 0xc5452758,
  dynticks_snap = 116840,
  dynticks_fqs = 47,
  offline_fqs = 0,
  n_rcu_pending = 48486,
  n_rp_qs_pending = 3,
  n_rp_report_qs = 2223,
  n_rp_cb_ready = 24,
  n_rp_cpu_needs_gp = 40101,
  n_rp_gp_completed = 1226,
  n_rp_gp_started = 789,
  n_rp_need_nothing = 4123,
  barrier_head = {
    next = 0x0,
    func = 0x0
  },
  oom_head = {
    next = 0x0,
    func = 0x0
  },
  cpu = 6,
  rsp = 0xc117f340 <rcu_preempt_state>
}

crash> struct rcu_data c54606b0
struct rcu_data {
  completed = 5877,
  gpnum = 5877,
  passed_quiesce = true,
  qs_pending = false,
  beenonline = true,
  preemptible = true,
  mynode = 0xc117f340 <rcu_preempt_state>,
  grpmask = 128,
  nxtlist = 0xdec32a6c,
  nxttail = {0xc54606c4, 0xe6fcf10c, 0xe6fcf10c, 0xe6fcf10c},
  nxtcompleted = {0, 5878, 5878, 5878},
  qlen_lazy = 1,
  qlen = 30,
  qlen_last_fqs_check = 0,
  n_cbs_invoked = 31998,
  n_nocbs_invoked = 0,
  n_cbs_orphaned = 0,
  n_cbs_adopted = 0,
  n_force_qs_snap = 1428,
  blimit = 10,
  dynticks = 0xc5460758,
  dynticks_snap = 57846,
  dynticks_fqs = 54,
  offline_fqs = 0,
  n_rcu_pending = 47502,
  n_rp_qs_pending = 2,
  n_rp_report_qs = 2142,
  n_rp_cb_ready = 37,
  n_rp_cpu_needs_gp = 40049,
  n_rp_gp_completed = 1223,
  n_rp_gp_started = 661,
  n_rp_need_nothing = 3390,
  barrier_head = {
    next = 0x0,
    func = 0x0
  },
  oom_head = {
    next = 0x0,
    func = 0x0
  },
  cpu = 7,
  rsp = 0xc117f340 <rcu_preempt_state>
}
rcu_preempt_state = $9 = {
  node = {{
      lock = {
        raw_lock = {
          {
            slock = 3129850509,
            tickets = {
              owner = 47757,
              next = 47757
            }
          }
        },
        magic = 3735899821,
        owner_cpu = 4294967295,
        owner = 0xffffffff
      },
      gpnum = 5877,
      completed = 5877,
      qsmask = 0,
      expmask = 0,
      qsmaskinit = 240,
      grpmask = 0,
      grplo = 0,
      grphi = 7,
      grpnum = 0 '\000',
      level = 0 '\000',
      parent = 0x0,
      blkd_tasks = {
        next = 0xc117f378 <rcu_preempt_state+56>,
        prev = 0xc117f378 <rcu_preempt_state+56>
      },
      gp_tasks = 0x0,
      exp_tasks = 0x0,
      need_future_gp = {1, 0},
      fqslock = {
        raw_lock = {
          {
            slock = 0,
            tickets = {
              owner = 0,
              next = 0
            }
          }
        },
        magic = 3735899821,
        owner_cpu = 4294967295,
        owner = 0xffffffff
      }
    }},
  level = {0xc117f340 <rcu_preempt_state>},
  levelcnt = {1, 0, 0, 0, 0},
  levelspread = "\b",
  rda = 0xc115e6b0 <rcu_preempt_data>,
  call = 0xc01975ac <call_rcu>,
  fqs_state = 0 '\000',
  boost = 0 '\000',
  gpnum = 5877,
  completed = 5877,
  gp_kthread = 0xf0c9e600,
  gp_wq = {
    lock = {
      {
        rlock = {
          raw_lock = {
            {
              slock = 2160230594,
              tickets = {
                owner = 32962,
                next = 32962
              }
            }
          },
          magic = 3735899821,
          owner_cpu = 4294967295,
          owner = 0xffffffff
        }
      }
    },
    task_list = {
      next = 0xf0cd1f20,
      prev = 0xf0cd1f20
    }
  },
  gp_flags = 1,
  orphan_lock = {
    raw_lock = {
      {
        slock = 327685,
        tickets = {
          owner = 5,
          next = 5
        }
      }
    },
    magic = 3735899821,
    owner_cpu = 4294967295,
    owner = 0xffffffff
  },
  orphan_nxtlist = 0x0,
  orphan_nxttail = 0xc117f490 <rcu_preempt_state+336>,
  orphan_donelist = 0x0,
  orphan_donetail = 0xc117f498 <rcu_preempt_state+344>,
  qlen_lazy = 0,
  qlen = 0,
  onoff_mutex = {
    count = {
      counter = 1
    },
    wait_lock = {
      {
        rlock = {
          raw_lock = {
            {
              slock = 811479134,
              tickets = {
                owner = 12382,
                next = 12382
              }
            }
          },
          magic = 3735899821,
          owner_cpu = 4294967295,
          owner = 0xffffffff
        }
      }
    },
    wait_list = {
      next = 0xc117f4bc <rcu_preempt_state+380>,
      prev = 0xc117f4bc <rcu_preempt_state+380>
    },
    owner = 0x0,
    name = 0x0,
    magic = 0xc117f4a8 <rcu_preempt_state+360>
  },
  barrier_mutex = {
    count = {
      counter = 1
    },
    wait_lock = {
      {
        rlock = {
          raw_lock = {
            {
              slock = 0,
              tickets = {
                owner = 0,
                next = 0
              }
            }
          },
          magic = 3735899821,
          owner_cpu = 4294967295,
          owner = 0xffffffff
        }
      }
    },
    wait_list = {
      next = 0xc117f4e4 <rcu_preempt_state+420>,
      prev = 0xc117f4e4 <rcu_preempt_state+420>
    },
    owner = 0x0,
    name = 0x0,
    magic = 0xc117f4d0 <rcu_preempt_state+400>
  },
  barrier_cpu_count = {
    counter = 0
  },
  barrier_completion = {
    done = 0,
    wait = {
      lock = {
        {
          rlock = {
            raw_lock = {
              {
                slock = 0,
                tickets = {
                  owner = 0,
                  next = 0
                }
              }
            },
            magic = 0,
            owner_cpu = 0,
            owner = 0x0
          }
        }
      },
      task_list = {
        next = 0x0,
        prev = 0x0
      }
    }
  },
  n_barrier_done = 0,
  expedited_start = {
    counter = 0
  },
  expedited_done = {
    counter = 0
  },
  expedited_wrap = {
    counter = 0
  },
  expedited_tryfail = {
    counter = 0
  },
  expedited_workdone1 = {
    counter = 0
  },
  expedited_workdone2 = {
    counter = 0
  },
  expedited_normal = {
    counter = 0
  },
  expedited_stoppedcpus = {
    counter = 0
  },
  expedited_done_tries = {
    counter = 0
  },
  expedited_done_lost = {
    counter = 0
  },
  expedited_done_exit = {
    counter = 0
  },
  jiffies_force_qs = 4294963917,
  n_force_qs = 4028,
  n_force_qs_lh = 0,
  n_force_qs_ngp = 0,
  gp_start = 4294963911,
  jiffies_stall = 4294966011,
  gp_max = 17,
  name = 0xc0d833ab "rcu_preempt",
  abbr = 112 'p',
  flavors = {
    next = 0xc117f2ec <rcu_bh_state+556>,
    prev = 0xc117f300 <rcu_struct_flavors>
  },
  wakeup_work = {
    flags = 3,
    llnode = {
      next = 0x0
    },
    func = 0xc0195aa8 <rsp_wakeup>
  }


Reply via email to