On Thu, Nov 29, 2018 at 11:49:02AM +0100, Peter Zijlstra wrote:
> On Wed, Nov 28, 2018 at 03:43:20PM -0800, Bart Van Assche wrote:
> > Instead of abandoning elements of list_entries[] that are no longer in
> > use, make alloc_list_entry() reuse array elements that have been freed.
> 
> > diff --git a/include/linux/lockdep.h b/include/linux/lockdep.h
> > index 43327a1dd488..01e55fca7c2c 100644
> > --- a/include/linux/lockdep.h
> > +++ b/include/linux/lockdep.h
> > @@ -183,6 +183,11 @@ static inline void lockdep_copy_map(struct lockdep_map 
> > *to,
> >  struct lock_list {
> >     /* Entry in locks_after or locks_before. */
> >     struct list_head                lock_order_entry;
> > +   /*
> > +    * Entry in all_list_entries when in use and entry in
> > +    * free_list_entries when not in use.
> > +    */
> > +   struct list_head                alloc_entry;
> >     struct lock_class               *class;
> >     struct lock_class               *links_to;
> >     struct stack_trace              trace;
> 
> > +static LIST_HEAD(all_list_entries);
> > +static LIST_HEAD(free_list_entries);
> >  
> 
> > @@ -862,7 +867,10 @@ register_lock_class(struct lockdep_map *lock, unsigned 
> > int subclass, int force)
> >   */
> >  static struct lock_list *alloc_list_entry(void)
> >  {
> > -   if (nr_list_entries >= MAX_LOCKDEP_ENTRIES) {
> > +   struct lock_list *e = list_first_entry_or_null(&free_list_entries,
> > +                                                  typeof(*e), alloc_entry);
> > +
> > +   if (!e) {
> >             if (!debug_locks_off_graph_unlock())
> >                     return NULL;
> >  
> > @@ -870,7 +878,8 @@ static struct lock_list *alloc_list_entry(void)
> >             dump_stack();
> >             return NULL;
> >     }
> > -   return list_entries + nr_list_entries++;
> > +   list_move_tail(&e->alloc_entry, &all_list_entries);
> > +   return e;
> >  }
> 
> > @@ -4235,19 +4244,19 @@ static void zap_class(struct list_head 
> > *zapped_classes,
> >                   struct lock_class *class)
> >  {
> >     struct lock_class *links_to;
> > +   struct lock_list *entry, *tmp;
> >  
> >     /*
> >      * Remove all dependencies this lock is
> >      * involved in:
> >      */
> > +   list_for_each_entry_safe(entry, tmp, &all_list_entries, alloc_entry) {
> >             if (entry->class != class && entry->links_to != class)
> >                     continue;
> >             links_to = entry->links_to;
> >             WARN_ON_ONCE(entry->class == links_to);
> >             list_del_rcu(&entry->lock_order_entry);
> > +           list_move(&entry->alloc_entry, &free_list_entries);
> >             entry->class = NULL;
> >             entry->links_to = NULL;
> >             check_free_class(zapped_classes, class);
> 
> Hurm.. I'm confused here.
> 
> The reason you cannot re-use lock_order_entry for the free list is
> because list_del_rcu(), right? But if so, then what ensures the
> list_entry is not re-used before it's grace-period?

Also; if you have to grow lock_list by 16 bytes just to be able to free
it, a bitmap allocator is much cheaper, space wise.

Some people seem to really care about the static image size, and
lockdep's .data section does matter to them.

Reply via email to