For each cpu in cpumask, checks if it's running a lockless pagetable
walk. Then, run serialize_against_pte_lookup() only on these cpus.

serialize_agains_pte_lookup() can take a long while when there are a
lot of cpus in cpumask.

This method is intended to reduce this waiting, while not impacting
too much on the lockless pagetable walk.

Signed-off-by: Leonardo Bras <leona...@linux.ibm.com>
---
 arch/powerpc/mm/book3s64/pgtable.c | 16 +++++++++++++++-
 1 file changed, 15 insertions(+), 1 deletion(-)

diff --git a/arch/powerpc/mm/book3s64/pgtable.c 
b/arch/powerpc/mm/book3s64/pgtable.c
index bb138b628f86..4822ff1aac4b 100644
--- a/arch/powerpc/mm/book3s64/pgtable.c
+++ b/arch/powerpc/mm/book3s64/pgtable.c
@@ -96,8 +96,22 @@ static DEFINE_PER_CPU(int, lockless_pgtbl_walk_counter);
  */
 void serialize_against_pte_lookup(struct mm_struct *mm)
 {
+       int cpu;
+       struct cpumask cm;
+
        smp_mb();
-       smp_call_function_many(mm_cpumask(mm), do_nothing, NULL, 1);
+
+       /*
+        * Fills a new cpumask only with cpus that are currently doing a
+        * lockless pagetable walk. This reduces time spent in this function.
+        */
+       cpumask_clear(&cm);
+       for_each_cpu(cpu, mm_cpumask((mm))) {
+               if (per_cpu(lockless_pgtbl_walk_counter, cpu) > 0)
+                       cpumask_set_cpu(cpu, &cm);
+       }
+
+       smp_call_function_many(&cm, do_nothing, NULL, 1);
 }
 
 /* begin_lockless_pgtbl_walk: Must be inserted before a function call that does
-- 
2.24.1

Reply via email to