Use the new pr_cont_t mechanism. Signed-off-by: John Ogness <john.ogn...@linutronix.de> --- kernel/locking/lockdep.c | 463 +++++++++++++++++++++++---------------- 1 file changed, 275 insertions(+), 188 deletions(-)
diff --git a/kernel/locking/lockdep.c b/kernel/locking/lockdep.c index 2fad21d345b0..3210f244cb94 100644 --- a/kernel/locking/lockdep.c +++ b/kernel/locking/lockdep.c @@ -638,7 +638,7 @@ void get_usage_chars(struct lock_class *class, char usage[LOCK_USAGE_CHARS]) usage[i] = '\0'; } -static void __print_lock_name(struct lock_class *class) +static void __print_lock_name(pr_cont_t *c, struct lock_class *class) { char str[KSYM_NAME_LEN]; const char *name; @@ -646,30 +646,30 @@ static void __print_lock_name(struct lock_class *class) name = class->name; if (!name) { name = __get_key_name(class->key, str); - printk(KERN_CONT "%s", name); + pr_cont_append(c, "%s", name); } else { - printk(KERN_CONT "%s", name); + pr_cont_append(c, "%s", name); if (class->name_version > 1) - printk(KERN_CONT "#%d", class->name_version); + pr_cont_append(c, "#%d", class->name_version); if (class->subclass) - printk(KERN_CONT "/%d", class->subclass); + pr_cont_append(c, "/%d", class->subclass); } } -static void print_lock_name(struct lock_class *class) +static void _print_lock_name(pr_cont_t *c, struct lock_class *class) { char usage[LOCK_USAGE_CHARS]; get_usage_chars(class, usage); - printk(KERN_CONT " ("); - __print_lock_name(class); - printk(KERN_CONT "){%s}-{%hd:%hd}", usage, + pr_cont_append(c, " ("); + __print_lock_name(c, class); + pr_cont_append(c, "){%s}-{%hd:%hd}", usage, class->wait_type_outer ?: class->wait_type_inner, class->wait_type_inner); } -static void print_lockdep_cache(struct lockdep_map *lock) +static void _print_lockdep_cache(pr_cont_t *c, struct lockdep_map *lock) { const char *name; char str[KSYM_NAME_LEN]; @@ -678,10 +678,10 @@ static void print_lockdep_cache(struct lockdep_map *lock) if (!name) name = __get_key_name(lock->key->subkeys, str); - printk(KERN_CONT "%s", name); + pr_cont_append(c, "%s", name); } -static void print_lock(struct held_lock *hlock) +static void _print_lock(pr_cont_t *c, struct held_lock *hlock) { /* * We can be called locklessly through debug_show_all_locks() so be @@ -696,16 +696,16 @@ static void print_lock(struct held_lock *hlock) struct lock_class *lock = hlock_class(hlock); if (!lock) { - printk(KERN_CONT "<RELEASED>\n"); + pr_cont_append(c, "<RELEASED>"); return; } - printk(KERN_CONT "%px", hlock->instance); - print_lock_name(lock); - printk(KERN_CONT ", at: %pS\n", (void *)hlock->acquire_ip); + pr_cont_append(c, "%px", hlock->instance); + _print_lock_name(c, lock); + pr_cont_append(c, ", at: %pS\n", (void *)hlock->acquire_ip); } -static void lockdep_print_held_locks(struct task_struct *p) +static void lockdep_print_held_locks(pr_cont_t *c, struct task_struct *p) { int i, depth = READ_ONCE(p->lockdep_depth); @@ -721,8 +721,9 @@ static void lockdep_print_held_locks(struct task_struct *p) if (p->state == TASK_RUNNING && p != current) return; for (i = 0; i < depth; i++) { - printk(" #%d: ", i); - print_lock(p->held_locks + i); + pr_cont_begin(c, " #%d: ", i); + _print_lock(c, p->held_locks + i); + pr_cont_end(c); } } @@ -937,7 +938,7 @@ static bool class_lock_list_valid(struct lock_class *c, struct list_head *h) list_for_each_entry(e, h, entry) { if (e->links_to != c) { - printk(KERN_INFO "class %s: mismatch for lock entry %ld; class %s <> %s", + printk(KERN_INFO "class %s: mismatch for lock entry %ld; class %s <> %s\n", c->name ? : "(?)", (unsigned long)(e - list_entries), e->links_to && e->links_to->name ? @@ -1195,6 +1196,7 @@ register_lock_class(struct lockdep_map *lock, unsigned int subclass, int force) struct lockdep_subclass_key *key; struct hlist_head *hash_head; struct lock_class *class; + pr_cont_t c; DEBUG_LOCKS_WARN_ON(!irqs_disabled()); @@ -1263,10 +1265,10 @@ register_lock_class(struct lockdep_map *lock, unsigned int subclass, int force) if (verbose(class)) { graph_unlock(); - printk("\nnew class %px: %s", class->key, class->name); + pr_cont_begin(&c, "\nnew class %px: %s", class->key, class->name); if (class->name_version > 1) - printk(KERN_CONT "#%d", class->name_version); - printk(KERN_CONT "\n"); + pr_cont_append(&c, "#%d", class->name_version); + pr_cont_end(&c); dump_stack(); if (!graph_lock()) { @@ -1567,18 +1569,20 @@ static void print_lock_trace(const struct lock_trace *trace, * has been detected): */ static noinline void -print_circular_bug_entry(struct lock_list *target, int depth) +print_circular_bug_entry(pr_cont_t *c, struct lock_list *target, int depth) { if (debug_locks_silent) return; - printk("\n-> #%u", depth); - print_lock_name(target->class); - printk(KERN_CONT ":\n"); + pr_cont_begin(c, "\n-> #%u", depth); + _print_lock_name(c, target->class); + pr_cont_append(c, ":"); + pr_cont_end(c); print_lock_trace(target->trace, 6); } static void -print_circular_lock_scenario(struct held_lock *src, +print_circular_lock_scenario(pr_cont_t *c, + struct held_lock *src, struct held_lock *tgt, struct lock_list *prt) { @@ -1600,30 +1604,35 @@ print_circular_lock_scenario(struct held_lock *src, * from the safe_class lock to the unsafe_class lock. */ if (parent != source) { - printk("Chain exists of:\n "); - __print_lock_name(source); - printk(KERN_CONT " --> "); - __print_lock_name(parent); - printk(KERN_CONT " --> "); - __print_lock_name(target); - printk(KERN_CONT "\n\n"); + pr_cont_begin(c, "Chain exists of:\n "); + __print_lock_name(c, source); + pr_cont_append(c, " --> "); + __print_lock_name(c, parent); + pr_cont_append(c, " --> "); + __print_lock_name(c, target); + pr_cont_append(c, "\n"); + pr_cont_end(c); } printk(" Possible unsafe locking scenario:\n\n"); printk(" CPU0 CPU1\n"); printk(" ---- ----\n"); - printk(" lock("); - __print_lock_name(target); - printk(KERN_CONT ");\n"); - printk(" lock("); - __print_lock_name(parent); - printk(KERN_CONT ");\n"); - printk(" lock("); - __print_lock_name(target); - printk(KERN_CONT ");\n"); - printk(" lock("); - __print_lock_name(source); - printk(KERN_CONT ");\n"); + pr_cont_begin(c, " lock("); + __print_lock_name(c, target); + pr_cont_append(c, ");"); + pr_cont_end(c); + pr_cont_begin(c, " lock("); + __print_lock_name(c, parent); + pr_cont_append(c, ");"); + pr_cont_end(c); + pr_cont_begin(c, " lock("); + __print_lock_name(c, target); + pr_cont_append(c, ");"); + pr_cont_end(c); + pr_cont_begin(c, " lock("); + __print_lock_name(c, source); + pr_cont_append(c, ");"); + pr_cont_end(c); printk("\n *** DEADLOCK ***\n\n"); } @@ -1632,7 +1641,7 @@ print_circular_lock_scenario(struct held_lock *src, * header first: */ static noinline void -print_circular_bug_header(struct lock_list *entry, unsigned int depth, +print_circular_bug_header(pr_cont_t *c, struct lock_list *entry, unsigned int depth, struct held_lock *check_src, struct held_lock *check_tgt) { @@ -1648,15 +1657,19 @@ print_circular_bug_header(struct lock_list *entry, unsigned int depth, pr_warn("------------------------------------------------------\n"); pr_warn("%s/%d is trying to acquire lock:\n", curr->comm, task_pid_nr(curr)); - print_lock(check_src); + pr_cont_begin(c, KERN_WARNING); + _print_lock(c, check_src); + pr_cont_end(c); pr_warn("\nbut task is already holding lock:\n"); - print_lock(check_tgt); + pr_cont_begin(c, KERN_WARNING); + _print_lock(c, check_tgt); + pr_cont_end(c); pr_warn("\nwhich lock already depends on the new lock.\n\n"); pr_warn("\nthe existing dependency chain (in reverse order) is:\n"); - print_circular_bug_entry(entry, depth); + print_circular_bug_entry(c, entry, depth); } static inline int class_equal(struct lock_list *entry, void *data) @@ -1664,7 +1677,7 @@ static inline int class_equal(struct lock_list *entry, void *data) return entry->class == data; } -static noinline void print_circular_bug(struct lock_list *this, +static noinline void print_circular_bug(pr_cont_t *c, struct lock_list *this, struct lock_list *target, struct held_lock *check_src, struct held_lock *check_tgt) @@ -1683,21 +1696,21 @@ static noinline void print_circular_bug(struct lock_list *this, depth = get_lock_depth(target); - print_circular_bug_header(target, depth, check_src, check_tgt); + print_circular_bug_header(c, target, depth, check_src, check_tgt); parent = get_lock_parent(target); first_parent = parent; while (parent) { - print_circular_bug_entry(parent, --depth); + print_circular_bug_entry(c, parent, --depth); parent = get_lock_parent(parent); } printk("\nother info that might help us debug this:\n\n"); - print_circular_lock_scenario(check_src, check_tgt, + print_circular_lock_scenario(c, check_src, check_tgt, first_parent); - lockdep_print_held_locks(curr); + lockdep_print_held_locks(c, curr); printk("\nstack backtrace:\n"); dump_stack(); @@ -1803,6 +1816,7 @@ static noinline int check_noncircular(struct held_lock *src, struct held_lock *target, struct lock_trace **const trace) { + pr_cont_t c; int ret; struct lock_list *target_entry; struct lock_list src_entry = { @@ -1824,7 +1838,7 @@ check_noncircular(struct held_lock *src, struct held_lock *target, *trace = save_trace(); } - print_circular_bug(&src_entry, target_entry, src, target); + print_circular_bug(&c, &src_entry, target_entry, src, target); } return ret; @@ -1928,23 +1942,25 @@ find_usage_backwards(struct lock_list *root, unsigned long usage_mask, return result; } -static void print_lock_class_header(struct lock_class *class, int depth) +static void print_lock_class_header(pr_cont_t *c, struct lock_class *class, int depth) { int bit; - printk("%*s->", depth, ""); - print_lock_name(class); + pr_cont_begin(c, "%*s->", depth, ""); + _print_lock_name(c, class); #ifdef CONFIG_DEBUG_LOCKDEP - printk(KERN_CONT " ops: %lu", debug_class_ops_read(class)); + pr_cont_append(c, " ops: %lu", debug_class_ops_read(class)); #endif - printk(KERN_CONT " {\n"); + pr_cont_append(c, " {"); + pr_cont_end(c); for (bit = 0; bit < LOCK_USAGE_STATES; bit++) { if (class->usage_mask & (1 << bit)) { int len = depth; - len += printk("%*s %s", depth, "", usage_str[bit]); - len += printk(KERN_CONT " at:\n"); + len += pr_cont_begin(c, "%*s %s", depth, "", usage_str[bit]); + len += pr_cont_append(c, " at:"); + pr_cont_end(c); print_lock_trace(class->usage_traces[bit], len); } } @@ -1958,7 +1974,7 @@ static void print_lock_class_header(struct lock_class *class, int depth) * printk the shortest lock dependencies from @start to @end in reverse order: */ static void __used -print_shortest_lock_dependencies(struct lock_list *leaf, +print_shortest_lock_dependencies(pr_cont_t *c, struct lock_list *leaf, struct lock_list *root) { struct lock_list *entry = leaf; @@ -1968,7 +1984,7 @@ print_shortest_lock_dependencies(struct lock_list *leaf, depth = get_lock_depth(leaf); do { - print_lock_class_header(entry->class, depth); + print_lock_class_header(c, entry->class, depth); printk("%*s ... acquired at:\n", depth, ""); print_lock_trace(entry->trace, 2); printk("\n"); @@ -1984,7 +2000,8 @@ print_shortest_lock_dependencies(struct lock_list *leaf, } static void -print_irq_lock_scenario(struct lock_list *safe_entry, +print_irq_lock_scenario(pr_cont_t *c, + struct lock_list *safe_entry, struct lock_list *unsafe_entry, struct lock_class *prev_class, struct lock_class *next_class) @@ -2010,37 +2027,43 @@ print_irq_lock_scenario(struct lock_list *safe_entry, * from the safe_class lock to the unsafe_class lock. */ if (middle_class != unsafe_class) { - printk("Chain exists of:\n "); - __print_lock_name(safe_class); - printk(KERN_CONT " --> "); - __print_lock_name(middle_class); - printk(KERN_CONT " --> "); - __print_lock_name(unsafe_class); - printk(KERN_CONT "\n\n"); + pr_cont_begin(c, "Chain exists of:\n "); + __print_lock_name(c, safe_class); + pr_cont_append(c, " --> "); + __print_lock_name(c, middle_class); + pr_cont_append(c, " --> "); + __print_lock_name(c, unsafe_class); + pr_cont_append(c, "\n"); + pr_cont_end(c); } printk(" Possible interrupt unsafe locking scenario:\n\n"); printk(" CPU0 CPU1\n"); printk(" ---- ----\n"); - printk(" lock("); - __print_lock_name(unsafe_class); - printk(KERN_CONT ");\n"); + pr_cont_begin(c, " lock("); + __print_lock_name(c, unsafe_class); + pr_cont_append(c, ");"); + pr_cont_end(c); printk(" local_irq_disable();\n"); - printk(" lock("); - __print_lock_name(safe_class); - printk(KERN_CONT ");\n"); - printk(" lock("); - __print_lock_name(middle_class); - printk(KERN_CONT ");\n"); + pr_cont_begin(c, " lock("); + __print_lock_name(c, safe_class); + pr_cont_append(c, ");"); + pr_cont_end(c); + pr_cont_begin(c, " lock("); + __print_lock_name(c, middle_class); + pr_cont_append(c, ");"); + pr_cont_end(c); printk(" <Interrupt>\n"); - printk(" lock("); - __print_lock_name(safe_class); - printk(KERN_CONT ");\n"); + pr_cont_begin(c, " lock("); + __print_lock_name(c, safe_class); + pr_cont_append(c, ");"); + pr_cont_end(c); printk("\n *** DEADLOCK ***\n\n"); } static void -print_bad_irq_dependency(struct task_struct *curr, +print_bad_irq_dependency(pr_cont_t *c, + struct task_struct *curr, struct lock_list *prev_root, struct lock_list *next_root, struct lock_list *backwards_entry, @@ -2066,48 +2089,57 @@ print_bad_irq_dependency(struct task_struct *curr, curr->softirq_context, softirq_count() >> SOFTIRQ_SHIFT, lockdep_hardirqs_enabled(), curr->softirqs_enabled); - print_lock(next); + pr_cont_begin(c, KERN_WARNING); + _print_lock(c, next); + pr_cont_end(c); pr_warn("\nand this task is already holding:\n"); - print_lock(prev); + pr_cont_begin(c, KERN_WARNING); + _print_lock(c, prev); + pr_cont_end(c); pr_warn("which would create a new lock dependency:\n"); - print_lock_name(hlock_class(prev)); - pr_cont(" ->"); - print_lock_name(hlock_class(next)); - pr_cont("\n"); + pr_cont_begin(c, KERN_WARNING); + _print_lock_name(c, hlock_class(prev)); + pr_cont_append(c, " ->"); + _print_lock_name(c, hlock_class(next)); + pr_cont_end(c); pr_warn("\nbut this new dependency connects a %s-irq-safe lock:\n", irqclass); - print_lock_name(backwards_entry->class); + pr_cont_begin(c, KERN_WARNING); + _print_lock_name(c, backwards_entry->class); + pr_cont_end(c); pr_warn("\n... which became %s-irq-safe at:\n", irqclass); print_lock_trace(backwards_entry->class->usage_traces[bit1], 1); pr_warn("\nto a %s-irq-unsafe lock:\n", irqclass); - print_lock_name(forwards_entry->class); + pr_cont_begin(c, KERN_WARNING); + _print_lock_name(c, forwards_entry->class); + pr_cont_end(c); pr_warn("\n... which became %s-irq-unsafe at:\n", irqclass); - pr_warn("..."); + pr_warn("...\n"); print_lock_trace(forwards_entry->class->usage_traces[bit2], 1); pr_warn("\nother info that might help us debug this:\n\n"); - print_irq_lock_scenario(backwards_entry, forwards_entry, + print_irq_lock_scenario(c, backwards_entry, forwards_entry, hlock_class(prev), hlock_class(next)); - lockdep_print_held_locks(curr); + lockdep_print_held_locks(c, curr); pr_warn("\nthe dependencies between %s-irq-safe lock and the holding lock:\n", irqclass); prev_root->trace = save_trace(); if (!prev_root->trace) return; - print_shortest_lock_dependencies(backwards_entry, prev_root); + print_shortest_lock_dependencies(c, backwards_entry, prev_root); pr_warn("\nthe dependencies between the lock to be acquired"); pr_warn(" and %s-irq-unsafe lock:\n", irqclass); next_root->trace = save_trace(); if (!next_root->trace) return; - print_shortest_lock_dependencies(forwards_entry, next_root); + print_shortest_lock_dependencies(c, forwards_entry, next_root); pr_warn("\nstack backtrace:\n"); dump_stack(); @@ -2247,6 +2279,7 @@ static int check_irq_usage(struct task_struct *curr, struct held_lock *prev, struct lock_list *target_entry1; struct lock_list *target_entry; struct lock_list this, that; + pr_cont_t c; int ret; /* @@ -2308,7 +2341,7 @@ static int check_irq_usage(struct task_struct *curr, struct held_lock *prev, if (DEBUG_LOCKS_WARN_ON(ret == -1)) return 1; - print_bad_irq_dependency(curr, &this, &that, + print_bad_irq_dependency(&c, curr, &this, &that, target_entry, target_entry1, prev, next, backward_bit, forward_bit, @@ -2347,7 +2380,7 @@ static void dec_chains(int irq_context) } static void -print_deadlock_scenario(struct held_lock *nxt, struct held_lock *prv) +print_deadlock_scenario(pr_cont_t *c, struct held_lock *nxt, struct held_lock *prv) { struct lock_class *next = hlock_class(nxt); struct lock_class *prev = hlock_class(prv); @@ -2355,18 +2388,20 @@ print_deadlock_scenario(struct held_lock *nxt, struct held_lock *prv) printk(" Possible unsafe locking scenario:\n\n"); printk(" CPU0\n"); printk(" ----\n"); - printk(" lock("); - __print_lock_name(prev); - printk(KERN_CONT ");\n"); - printk(" lock("); - __print_lock_name(next); - printk(KERN_CONT ");\n"); + pr_cont_begin(c, " lock("); + __print_lock_name(c, prev); + pr_cont_append(c, ");"); + pr_cont_end(c); + pr_cont_begin(c, " lock("); + __print_lock_name(c, next); + pr_cont_append(c, ");"); + pr_cont_end(c); printk("\n *** DEADLOCK ***\n\n"); printk(" May be due to missing lock nesting notation\n\n"); } static void -print_deadlock_bug(struct task_struct *curr, struct held_lock *prev, +print_deadlock_bug(pr_cont_t *c, struct task_struct *curr, struct held_lock *prev, struct held_lock *next) { if (!debug_locks_off_graph_unlock() || debug_locks_silent) @@ -2379,13 +2414,17 @@ print_deadlock_bug(struct task_struct *curr, struct held_lock *prev, pr_warn("--------------------------------------------\n"); pr_warn("%s/%d is trying to acquire lock:\n", curr->comm, task_pid_nr(curr)); - print_lock(next); + pr_cont_begin(c, KERN_WARNING); + _print_lock(c, next); + pr_cont_end(c); pr_warn("\nbut task is already holding lock:\n"); - print_lock(prev); + pr_cont_begin(c, KERN_WARNING); + _print_lock(c, prev); + pr_cont_end(c); pr_warn("\nother info that might help us debug this:\n"); - print_deadlock_scenario(next, prev); - lockdep_print_held_locks(curr); + print_deadlock_scenario(c, next, prev); + lockdep_print_held_locks(c, curr); pr_warn("\nstack backtrace:\n"); dump_stack(); @@ -2404,6 +2443,7 @@ check_deadlock(struct task_struct *curr, struct held_lock *next) { struct held_lock *prev; struct held_lock *nest = NULL; + pr_cont_t c; int i; for (i = 0; i < curr->lockdep_depth; i++) { @@ -2429,7 +2469,7 @@ check_deadlock(struct task_struct *curr, struct held_lock *next) if (nest) return 2; - print_deadlock_bug(curr, prev, next); + print_deadlock_bug(&c, curr, prev, next); return 0; } return 1; @@ -2901,18 +2941,18 @@ static inline int get_first_held_lock(struct task_struct *curr, /* * Returns the next chain_key iteration */ -static u64 print_chain_key_iteration(int class_idx, u64 chain_key) +static u64 _print_chain_key_iteration(pr_cont_t *c, int class_idx, u64 chain_key) { u64 new_chain_key = iterate_chain_key(chain_key, class_idx); - printk(" class_idx:%d -> chain_key:%016Lx", + pr_cont_append(c, " class_idx:%d -> chain_key:%016Lx", class_idx, (unsigned long long)new_chain_key); return new_chain_key; } static void -print_chain_keys_held_locks(struct task_struct *curr, struct held_lock *hlock_next) +print_chain_keys_held_locks(pr_cont_t *c, struct task_struct *curr, struct held_lock *hlock_next) { struct held_lock *hlock; u64 chain_key = INITIAL_CHAIN_KEY; @@ -2923,16 +2963,20 @@ print_chain_keys_held_locks(struct task_struct *curr, struct held_lock *hlock_ne hlock_next->irq_context); for (; i < depth; i++) { hlock = curr->held_locks + i; - chain_key = print_chain_key_iteration(hlock->class_idx, chain_key); - print_lock(hlock); + pr_cont_begin(c, ""); + chain_key = _print_chain_key_iteration(c, hlock->class_idx, chain_key); + _print_lock(c, hlock); + pr_cont_end(c); } - print_chain_key_iteration(hlock_next->class_idx, chain_key); - print_lock(hlock_next); + pr_cont_begin(c, ""); + _print_chain_key_iteration(c, hlock_next->class_idx, chain_key); + _print_lock(c, hlock_next); + pr_cont_end(c); } -static void print_chain_keys_chain(struct lock_chain *chain) +static void print_chain_keys_chain(pr_cont_t *c, struct lock_chain *chain) { int i; u64 chain_key = INITIAL_CHAIN_KEY; @@ -2941,14 +2985,15 @@ static void print_chain_keys_chain(struct lock_chain *chain) printk("depth: %u\n", chain->depth); for (i = 0; i < chain->depth; i++) { class_id = chain_hlocks[chain->base + i]; - chain_key = print_chain_key_iteration(class_id, chain_key); - print_lock_name(lock_classes + class_id); - printk("\n"); + pr_cont_begin(c, ""); + chain_key = _print_chain_key_iteration(c, class_id, chain_key); + _print_lock_name(c, lock_classes + class_id); + pr_cont_end(c); } } -static void print_collision(struct task_struct *curr, +static void print_collision(pr_cont_t *c, struct task_struct *curr, struct held_lock *hlock_next, struct lock_chain *chain) { @@ -2961,10 +3006,10 @@ static void print_collision(struct task_struct *curr, pr_warn("Hash chain already cached but the contents don't match!\n"); pr_warn("Held locks:"); - print_chain_keys_held_locks(curr, hlock_next); + print_chain_keys_held_locks(c, curr, hlock_next); pr_warn("Locks in cached chain:"); - print_chain_keys_chain(chain); + print_chain_keys_chain(c, chain); pr_warn("\nstack backtrace:\n"); dump_stack(); @@ -2983,11 +3028,12 @@ static int check_no_collision(struct task_struct *curr, { #ifdef CONFIG_DEBUG_LOCKDEP int i, j, id; + pr_cont_t c; i = get_first_held_lock(curr, hlock); if (DEBUG_LOCKS_WARN_ON(chain->depth != curr->lockdep_depth - (i - 1))) { - print_collision(curr, hlock, chain); + print_collision(&c, curr, hlock, chain); return 0; } @@ -2995,7 +3041,7 @@ static int check_no_collision(struct task_struct *curr, id = curr->held_locks[i].class_idx; if (DEBUG_LOCKS_WARN_ON(chain_hlocks[chain->base + j] != id)) { - print_collision(curr, hlock, chain); + print_collision(&c, curr, hlock, chain); return 0; } } @@ -3296,25 +3342,27 @@ static void check_chain_key(struct task_struct *curr) static int mark_lock(struct task_struct *curr, struct held_lock *this, enum lock_usage_bit new_bit); -static void print_usage_bug_scenario(struct held_lock *lock) +static void print_usage_bug_scenario(pr_cont_t *c, struct held_lock *lock) { struct lock_class *class = hlock_class(lock); printk(" Possible unsafe locking scenario:\n\n"); printk(" CPU0\n"); printk(" ----\n"); - printk(" lock("); - __print_lock_name(class); - printk(KERN_CONT ");\n"); + pr_cont_begin(c, " lock("); + __print_lock_name(c, class); + pr_cont_append(c, ");"); + pr_cont_end(c); printk(" <Interrupt>\n"); - printk(" lock("); - __print_lock_name(class); - printk(KERN_CONT ");\n"); + pr_cont_begin(c, " lock("); + __print_lock_name(c, class); + pr_cont_append(c, ");"); + pr_cont_end(c); printk("\n *** DEADLOCK ***\n\n"); } static void -print_usage_bug(struct task_struct *curr, struct held_lock *this, +print_usage_bug(pr_cont_t *c, struct task_struct *curr, struct held_lock *this, enum lock_usage_bit prev_bit, enum lock_usage_bit new_bit) { if (!debug_locks_off_graph_unlock() || debug_locks_silent) @@ -3335,16 +3383,18 @@ print_usage_bug(struct task_struct *curr, struct held_lock *this, lockdep_softirq_context(curr), softirq_count() >> SOFTIRQ_SHIFT, lockdep_hardirqs_enabled(), lockdep_softirqs_enabled(curr)); - print_lock(this); + pr_cont_begin(c, KERN_WARNING); + _print_lock(c, this); + pr_cont_end(c); pr_warn("{%s} state was registered at:\n", usage_str[prev_bit]); print_lock_trace(hlock_class(this)->usage_traces[prev_bit], 1); print_irqtrace_events(curr); pr_warn("\nother info that might help us debug this:\n"); - print_usage_bug_scenario(this); + print_usage_bug_scenario(c, this); - lockdep_print_held_locks(curr); + lockdep_print_held_locks(c, curr); pr_warn("\nstack backtrace:\n"); dump_stack(); @@ -3357,8 +3407,10 @@ static inline int valid_state(struct task_struct *curr, struct held_lock *this, enum lock_usage_bit new_bit, enum lock_usage_bit bad_bit) { + pr_cont_t c; + if (unlikely(hlock_class(this)->usage_mask & (1 << bad_bit))) { - print_usage_bug(curr, this, bad_bit, new_bit); + print_usage_bug(&c, curr, this, bad_bit, new_bit); return 0; } return 1; @@ -3369,7 +3421,7 @@ valid_state(struct task_struct *curr, struct held_lock *this, * print irq inversion bug: */ static void -print_irq_inversion_bug(struct task_struct *curr, +print_irq_inversion_bug(pr_cont_t *c, struct task_struct *curr, struct lock_list *root, struct lock_list *other, struct held_lock *this, int forwards, const char *irqclass) @@ -3388,12 +3440,16 @@ print_irq_inversion_bug(struct task_struct *curr, pr_warn("--------------------------------------------------------\n"); pr_warn("%s/%d just changed the state of lock:\n", curr->comm, task_pid_nr(curr)); - print_lock(this); + pr_cont_begin(c, KERN_WARNING); + _print_lock(c, this); + pr_cont_end(c); if (forwards) pr_warn("but this lock took another, %s-unsafe lock in the past:\n", irqclass); else pr_warn("but this lock was taken by another, %s-safe lock in the past:\n", irqclass); - print_lock_name(other->class); + pr_cont_begin(c, KERN_WARNING); + _print_lock_name(c, other->class); + pr_cont_end(c); pr_warn("\n\nand interrupts could create inverse lock ordering between them.\n\n"); pr_warn("\nother info that might help us debug this:\n"); @@ -3410,19 +3466,19 @@ print_irq_inversion_bug(struct task_struct *curr, depth--; } while (entry && entry != root && (depth >= 0)); if (forwards) - print_irq_lock_scenario(root, other, + print_irq_lock_scenario(c, root, other, middle ? middle->class : root->class, other->class); else - print_irq_lock_scenario(other, root, + print_irq_lock_scenario(c, other, root, middle ? middle->class : other->class, root->class); - lockdep_print_held_locks(curr); + lockdep_print_held_locks(c, curr); pr_warn("\nthe shortest dependencies between 2nd lock and 1st lock:\n"); root->trace = save_trace(); if (!root->trace) return; - print_shortest_lock_dependencies(other, root); + print_shortest_lock_dependencies(c, other, root); pr_warn("\nstack backtrace:\n"); dump_stack(); @@ -3439,6 +3495,7 @@ check_usage_forwards(struct task_struct *curr, struct held_lock *this, int ret; struct lock_list root; struct lock_list *target_entry; + pr_cont_t c; root.parent = NULL; root.class = hlock_class(this); @@ -3450,7 +3507,7 @@ check_usage_forwards(struct task_struct *curr, struct held_lock *this, if (ret == 1) return ret; - print_irq_inversion_bug(curr, &root, target_entry, + print_irq_inversion_bug(&c, curr, &root, target_entry, this, 1, irqclass); return 0; } @@ -3466,6 +3523,7 @@ check_usage_backwards(struct task_struct *curr, struct held_lock *this, int ret; struct lock_list root; struct lock_list *target_entry; + pr_cont_t c; root.parent = NULL; root.class = hlock_class(this); @@ -3477,7 +3535,7 @@ check_usage_backwards(struct task_struct *curr, struct held_lock *this, if (ret == 1) return ret; - print_irq_inversion_bug(curr, &root, target_entry, + print_irq_inversion_bug(&c, curr, &root, target_entry, this, 0, irqclass); return 0; } @@ -3970,6 +4028,7 @@ static int mark_lock(struct task_struct *curr, struct held_lock *this, enum lock_usage_bit new_bit) { unsigned int new_mask = 1 << new_bit, ret = 1; + pr_cont_t c; if (new_bit >= LOCK_USAGE_STATES) { DEBUG_LOCKS_WARN_ON(1); @@ -4015,7 +4074,9 @@ static int mark_lock(struct task_struct *curr, struct held_lock *this, */ if (ret == 2) { printk("\nmarked lock as {%s}:\n", usage_str[new_bit]); - print_lock(this); + pr_cont_begin(&c, ""); + _print_lock(&c, this); + pr_cont_end(&c); print_irqtrace_events(curr); dump_stack(); } @@ -4048,7 +4109,7 @@ static inline short task_wait_context(struct task_struct *curr) } static int -print_lock_invalid_wait_context(struct task_struct *curr, +print_lock_invalid_wait_context(pr_cont_t *c, struct task_struct *curr, struct held_lock *hlock) { short curr_inner; @@ -4065,14 +4126,16 @@ print_lock_invalid_wait_context(struct task_struct *curr, pr_warn("-----------------------------\n"); pr_warn("%s/%d is trying to lock:\n", curr->comm, task_pid_nr(curr)); - print_lock(hlock); + pr_cont_begin(c, KERN_WARNING); + _print_lock(c, hlock); + pr_cont_end(c); pr_warn("other info that might help us debug this:\n"); curr_inner = task_wait_context(curr); pr_warn("context-{%d:%d}\n", curr_inner, curr_inner); - lockdep_print_held_locks(curr); + lockdep_print_held_locks(c, curr); pr_warn("stack backtrace:\n"); dump_stack(); @@ -4100,6 +4163,7 @@ static int check_wait_context(struct task_struct *curr, struct held_lock *next) short next_inner = hlock_class(next)->wait_type_inner; short next_outer = hlock_class(next)->wait_type_outer; short curr_inner; + pr_cont_t c; int depth; if (!curr->lockdep_depth || !next_inner || next->trylock) @@ -4136,7 +4200,7 @@ static int check_wait_context(struct task_struct *curr, struct held_lock *next) } if (next_outer > curr_inner) - return print_lock_invalid_wait_context(curr, next); + return print_lock_invalid_wait_context(&c, curr, next); return 0; } @@ -4236,7 +4300,7 @@ struct lock_class_key __lockdep_no_validate__; EXPORT_SYMBOL_GPL(__lockdep_no_validate__); static void -print_lock_nested_lock_not_held(struct task_struct *curr, +print_lock_nested_lock_not_held(pr_cont_t *c, struct task_struct *curr, struct held_lock *hlock, unsigned long ip) { @@ -4252,7 +4316,9 @@ print_lock_nested_lock_not_held(struct task_struct *curr, pr_warn("----------------------------------\n"); pr_warn("%s/%d is trying to lock:\n", curr->comm, task_pid_nr(curr)); - print_lock(hlock); + pr_cont_begin(c, KERN_WARNING); + _print_lock(c, hlock); + pr_cont_end(c); pr_warn("\nbut this task is not holding:\n"); pr_warn("%s\n", hlock->nest_lock->name); @@ -4261,7 +4327,7 @@ print_lock_nested_lock_not_held(struct task_struct *curr, dump_stack(); pr_warn("\nother info that might help us debug this:\n"); - lockdep_print_held_locks(curr); + lockdep_print_held_locks(c, curr); pr_warn("\nstack backtrace:\n"); dump_stack(); @@ -4289,6 +4355,7 @@ static int __lock_acquire(struct lockdep_map *lock, unsigned int subclass, int chain_head = 0; int class_idx; u64 chain_key; + pr_cont_t c; if (unlikely(!debug_locks)) return 0; @@ -4310,10 +4377,10 @@ static int __lock_acquire(struct lockdep_map *lock, unsigned int subclass, debug_class_ops_inc(class); if (very_verbose(class)) { - printk("\nacquire class [%px] %s", class->key, class->name); + pr_cont_begin(&c, "\nacquire class [%px] %s", class->key, class->name); if (class->name_version > 1) - printk(KERN_CONT "#%d", class->name_version); - printk(KERN_CONT "\n"); + pr_cont_append(&c, "#%d", class->name_version); + pr_cont_end(&c); dump_stack(); } @@ -4414,7 +4481,7 @@ static int __lock_acquire(struct lockdep_map *lock, unsigned int subclass, chain_key = iterate_chain_key(chain_key, class_idx); if (nest_lock && !__lock_is_held(nest_lock, -1)) { - print_lock_nested_lock_not_held(curr, hlock, ip); + print_lock_nested_lock_not_held(&c, curr, hlock, ip); return 0; } @@ -4439,7 +4506,7 @@ static int __lock_acquire(struct lockdep_map *lock, unsigned int subclass, printk(KERN_DEBUG "depth: %i max: %lu!\n", curr->lockdep_depth, MAX_LOCK_DEPTH); - lockdep_print_held_locks(current); + lockdep_print_held_locks(&c, current); debug_show_all_locks(); dump_stack(); @@ -4452,7 +4519,7 @@ static int __lock_acquire(struct lockdep_map *lock, unsigned int subclass, return 1; } -static void print_unlock_imbalance_bug(struct task_struct *curr, +static void print_unlock_imbalance_bug(pr_cont_t *c, struct task_struct *curr, struct lockdep_map *lock, unsigned long ip) { @@ -4468,12 +4535,14 @@ static void print_unlock_imbalance_bug(struct task_struct *curr, pr_warn("-------------------------------------\n"); pr_warn("%s/%d is trying to release lock (", curr->comm, task_pid_nr(curr)); - print_lockdep_cache(lock); - pr_cont(") at:\n"); + pr_cont_begin(c, KERN_WARNING); + _print_lockdep_cache(c, lock); + pr_cont_append(c, ") at:"); + pr_cont_end(c); print_ip_sym(KERN_WARNING, ip); pr_warn("but there are no more locks to release!\n"); pr_warn("\nother info that might help us debug this:\n"); - lockdep_print_held_locks(curr); + lockdep_print_held_locks(c, curr); pr_warn("\nstack backtrace:\n"); dump_stack(); @@ -4592,6 +4661,7 @@ __lock_set_class(struct lockdep_map *lock, const char *name, unsigned int depth, merged = 0; struct held_lock *hlock; struct lock_class *class; + pr_cont_t c; int i; if (unlikely(!debug_locks)) @@ -4607,7 +4677,7 @@ __lock_set_class(struct lockdep_map *lock, const char *name, hlock = find_held_lock(curr, lock, depth, &i); if (!hlock) { - print_unlock_imbalance_bug(curr, lock, ip); + print_unlock_imbalance_bug(&c, curr, lock, ip); return 0; } @@ -4637,6 +4707,7 @@ static int __lock_downgrade(struct lockdep_map *lock, unsigned long ip) struct task_struct *curr = current; unsigned int depth, merged = 0; struct held_lock *hlock; + pr_cont_t c; int i; if (unlikely(!debug_locks)) @@ -4652,7 +4723,7 @@ static int __lock_downgrade(struct lockdep_map *lock, unsigned long ip) hlock = find_held_lock(curr, lock, depth, &i); if (!hlock) { - print_unlock_imbalance_bug(curr, lock, ip); + print_unlock_imbalance_bug(&c, curr, lock, ip); return 0; } @@ -4691,6 +4762,7 @@ __lock_release(struct lockdep_map *lock, unsigned long ip) struct task_struct *curr = current; unsigned int depth, merged = 1; struct held_lock *hlock; + pr_cont_t c; int i; if (unlikely(!debug_locks)) @@ -4702,7 +4774,7 @@ __lock_release(struct lockdep_map *lock, unsigned long ip) * own any locks, you've been drinking again? */ if (depth <= 0) { - print_unlock_imbalance_bug(curr, lock, ip); + print_unlock_imbalance_bug(&c, curr, lock, ip); return 0; } @@ -4712,7 +4784,7 @@ __lock_release(struct lockdep_map *lock, unsigned long ip) */ hlock = find_held_lock(curr, lock, depth, &i); if (!hlock) { - print_unlock_imbalance_bug(curr, lock, ip); + print_unlock_imbalance_bug(&c, curr, lock, ip); return 0; } @@ -4942,6 +5014,7 @@ static void verify_lock_unused(struct lockdep_map *lock, struct held_lock *hlock { #ifdef CONFIG_PROVE_LOCKING struct lock_class *class = look_up_lock_class(lock, subclass); + pr_cont_t c; /* if it doesn't have a class (yet), it certainly hasn't been used yet */ if (!class) @@ -4952,7 +5025,7 @@ static void verify_lock_unused(struct lockdep_map *lock, struct held_lock *hlock hlock->class_idx = class - lock_classes; - print_usage_bug(current, hlock, LOCK_USED, LOCK_USAGE_STATES); + print_usage_bug(&c, current, hlock, LOCK_USED, LOCK_USAGE_STATES); #endif } @@ -5103,7 +5176,7 @@ void lock_unpin_lock(struct lockdep_map *lock, struct pin_cookie cookie) EXPORT_SYMBOL_GPL(lock_unpin_lock); #ifdef CONFIG_LOCK_STAT -static void print_lock_contention_bug(struct task_struct *curr, +static void print_lock_contention_bug(pr_cont_t *c, struct task_struct *curr, struct lockdep_map *lock, unsigned long ip) { @@ -5119,12 +5192,14 @@ static void print_lock_contention_bug(struct task_struct *curr, pr_warn("---------------------------------\n"); pr_warn("%s/%d is trying to contend lock (", curr->comm, task_pid_nr(curr)); - print_lockdep_cache(lock); - pr_cont(") at:\n"); + pr_cont_begin(c, KERN_WARNING); + _print_lockdep_cache(c, lock); + pr_cont_append(c, ") at:"); + pr_cont_end(c); print_ip_sym(KERN_WARNING, ip); pr_warn("but there are no locks held!\n"); pr_warn("\nother info that might help us debug this:\n"); - lockdep_print_held_locks(curr); + lockdep_print_held_locks(c, curr); pr_warn("\nstack backtrace:\n"); dump_stack(); @@ -5138,6 +5213,7 @@ __lock_contended(struct lockdep_map *lock, unsigned long ip) struct lock_class_stats *stats; unsigned int depth; int i, contention_point, contending_point; + pr_cont_t c; depth = curr->lockdep_depth; /* @@ -5149,7 +5225,7 @@ __lock_contended(struct lockdep_map *lock, unsigned long ip) hlock = find_held_lock(curr, lock, depth, &i); if (!hlock) { - print_lock_contention_bug(curr, lock, ip); + print_lock_contention_bug(&c, curr, lock, ip); return; } @@ -5179,6 +5255,7 @@ __lock_acquired(struct lockdep_map *lock, unsigned long ip) struct lock_class_stats *stats; unsigned int depth; u64 now, waittime = 0; + pr_cont_t c; int i, cpu; depth = curr->lockdep_depth; @@ -5191,7 +5268,7 @@ __lock_acquired(struct lockdep_map *lock, unsigned long ip) hlock = find_held_lock(curr, lock, depth, &i); if (!hlock) { - print_lock_contention_bug(curr, lock, _RET_IP_); + print_lock_contention_bug(&c, curr, lock, _RET_IP_); return; } @@ -5733,7 +5810,7 @@ void __init lockdep_init(void) } static void -print_freed_lock_bug(struct task_struct *curr, const void *mem_from, +print_freed_lock_bug(pr_cont_t *c, struct task_struct *curr, const void *mem_from, const void *mem_to, struct held_lock *hlock) { if (!debug_locks_off()) @@ -5748,8 +5825,10 @@ print_freed_lock_bug(struct task_struct *curr, const void *mem_from, pr_warn("-------------------------\n"); pr_warn("%s/%d is freeing memory %px-%px, with a lock still held there!\n", curr->comm, task_pid_nr(curr), mem_from, mem_to-1); - print_lock(hlock); - lockdep_print_held_locks(curr); + pr_cont_begin(c, KERN_WARNING); + _print_lock(c, hlock); + pr_cont_end(c); + lockdep_print_held_locks(c, curr); pr_warn("\nstack backtrace:\n"); dump_stack(); @@ -5772,6 +5851,7 @@ void debug_check_no_locks_freed(const void *mem_from, unsigned long mem_len) struct task_struct *curr = current; struct held_lock *hlock; unsigned long flags; + pr_cont_t c; int i; if (unlikely(!debug_locks)) @@ -5785,7 +5865,7 @@ void debug_check_no_locks_freed(const void *mem_from, unsigned long mem_len) sizeof(*hlock->instance))) continue; - print_freed_lock_bug(curr, mem_from, mem_from + mem_len, hlock); + print_freed_lock_bug(&c, curr, mem_from, mem_from + mem_len, hlock); break; } raw_local_irq_restore(flags); @@ -5794,6 +5874,8 @@ EXPORT_SYMBOL_GPL(debug_check_no_locks_freed); static void print_held_locks_bug(void) { + pr_cont_t c; + if (!debug_locks_off()) return; if (debug_locks_silent) @@ -5805,7 +5887,7 @@ static void print_held_locks_bug(void) current->comm, task_pid_nr(current)); print_kernel_ident(); pr_warn("------------------------------------\n"); - lockdep_print_held_locks(current); + lockdep_print_held_locks(&c, current); pr_warn("\nstack backtrace:\n"); dump_stack(); } @@ -5821,6 +5903,7 @@ EXPORT_SYMBOL_GPL(debug_check_no_locks_held); void debug_show_all_locks(void) { struct task_struct *g, *p; + pr_cont_t c; if (unlikely(!debug_locks)) { pr_warn("INFO: lockdep is turned off.\n"); @@ -5832,7 +5915,7 @@ void debug_show_all_locks(void) for_each_process_thread(g, p) { if (!p->lockdep_depth) continue; - lockdep_print_held_locks(p); + lockdep_print_held_locks(&c, p); touch_nmi_watchdog(); touch_all_softlockup_watchdogs(); } @@ -5850,17 +5933,20 @@ EXPORT_SYMBOL_GPL(debug_show_all_locks); */ void debug_show_held_locks(struct task_struct *task) { + pr_cont_t c; + if (unlikely(!debug_locks)) { printk("INFO: lockdep is turned off.\n"); return; } - lockdep_print_held_locks(task); + lockdep_print_held_locks(&c, task); } EXPORT_SYMBOL_GPL(debug_show_held_locks); asmlinkage __visible void lockdep_sys_exit(void) { struct task_struct *curr = current; + pr_cont_t c; if (unlikely(curr->lockdep_depth)) { if (!debug_locks_off()) @@ -5872,7 +5958,7 @@ asmlinkage __visible void lockdep_sys_exit(void) pr_warn("------------------------------------------------\n"); pr_warn("%s/%d is leaving the kernel with locks still held!\n", curr->comm, curr->pid); - lockdep_print_held_locks(curr); + lockdep_print_held_locks(&c, curr); } /* @@ -5885,6 +5971,7 @@ asmlinkage __visible void lockdep_sys_exit(void) void lockdep_rcu_suspicious(const char *file, const int line, const char *s) { struct task_struct *curr = current; + pr_cont_t c; /* Note: the following can be executed concurrently, so be careful. */ pr_warn("\n"); @@ -5921,7 +6008,7 @@ void lockdep_rcu_suspicious(const char *file, const int line, const char *s) if (!rcu_is_watching()) pr_warn("RCU used illegally from extended quiescent state!\n"); - lockdep_print_held_locks(curr); + lockdep_print_held_locks(&c, curr); pr_warn("\nstack backtrace:\n"); dump_stack(); } -- 2.20.1