This patch does not change any functionality but makes the next patch
in this series easier to read.

Signed-off-by: Bart Van Assche <bvanass...@acm.org>
---
 kernel/locking/lockdep.c | 35 +++++++++++++++++++++++------------
 1 file changed, 23 insertions(+), 12 deletions(-)

diff --git a/kernel/locking/lockdep.c b/kernel/locking/lockdep.c
index 141bb0662ff5..0e273731d028 100644
--- a/kernel/locking/lockdep.c
+++ b/kernel/locking/lockdep.c
@@ -4459,18 +4459,16 @@ static void free_zapped_classes(struct list_head 
*zapped_classes)
 }
 
 /*
- * Used in module.c to remove lock classes from memory that is going to be
- * freed; and possibly re-used by other modules.
- *
- * We will have had one sync_sched() before getting here, so we're guaranteed
- * nobody will look up these exact classes -- they're properly dead but still
- * allocated.
+ * Remove all lock classes from the class hash table and from the
+ * all_lock_classes list whose key or name is in the address range
+ * [start, start + size). Move these lock classes to the
+ * @zapped_classes list.
  */
-void lockdep_free_key_range(void *start, unsigned long size)
+static void __lockdep_free_key_range(struct list_head *zapped_classes,
+                                    void *start, unsigned long size)
 {
        struct lock_class *class;
        struct hlist_head *head;
-       LIST_HEAD(zapped_classes);
        unsigned long flags;
        int i;
        int locked;
@@ -4478,9 +4476,8 @@ void lockdep_free_key_range(void *start, unsigned long 
size)
        raw_local_irq_save(flags);
        locked = graph_lock();
 
-       /*
-        * Unhash all classes that were created by this module:
-        */
+       INIT_LIST_HEAD(zapped_classes);
+
        for (i = 0; i < CLASSHASH_SIZE; i++) {
                head = classhash_table + i;
                hlist_for_each_entry_rcu(class, head, hash_entry) {
@@ -4488,14 +4485,28 @@ void lockdep_free_key_range(void *start, unsigned long 
size)
                            (!within(class->key, start, size) &&
                             !within(class->name, start, size)))
                                continue;
-                       zap_class(&zapped_classes, class);
+                       zap_class(zapped_classes, class);
                }
        }
 
        if (locked)
                graph_unlock();
        raw_local_irq_restore(flags);
+}
+
+/*
+ * Used in module.c to remove lock classes from memory that is going to be
+ * freed; and possibly re-used by other modules.
+ *
+ * We will have had one sync_sched() before getting here, so we're guaranteed
+ * nobody will look up these exact classes -- they're properly dead but still
+ * allocated.
+ */
+void lockdep_free_key_range(void *start, unsigned long size)
+{
+       LIST_HEAD(zapped_classes);
 
+       __lockdep_free_key_range(&zapped_classes, start, size);
        free_zapped_classes(&zapped_classes);
 }
 
-- 
2.20.0.rc0.387.gc7a69e6b6c-goog

Reply via email to