Hi Linus,

the patch below fixes a race in the slab allocator:
If one cpu increases the batch count (kmem_tune_cpucache, just before 
smp_call_function_all_cpus()) and another cpu allocates a new batch
of slab entries before the IPI arrives, then kmem_cache_alloc_batch
writes behind the end of the cpu local array.

The patch is against 2.4.0-test11.
--
  Manfred

<<<<<<<<<<<<<<
--- 2.4/mm/slab.c       Tue Oct  3 20:03:44 2000
+++ build-2.4/mm/slab.c Sat Dec  2 15:00:32 2000
@@ -188,7 +188,7 @@
        unsigned int            num;    /* # of objs per slab */
        spinlock_t              spinlock;
 #ifdef CONFIG_SMP
-       unsigned int            batchcount;
+       volatile unsigned int   batchcount;
 #endif
 
 /* 2) slab additions /removals */
@@ -1637,11 +1637,19 @@
                }
        }
        new.cachep = cachep;
-       spin_lock_irq(&cachep->spinlock);
-       cachep->batchcount = batchcount;
-       spin_unlock_irq(&cachep->spinlock);
+       if(batchcount < cachep->batchcount) {
+               spin_lock_irq(&cachep->spinlock);
+               cachep->batchcount = batchcount;
+               spin_unlock_irq(&cachep->spinlock);
+       }
 
        smp_call_function_all_cpus(do_ccupdate_local, (void *)&new);
+
+       if(batchcount > cachep->batchcount) {
+               spin_lock_irq(&cachep->spinlock);
+               cachep->batchcount = batchcount;
+               spin_unlock_irq(&cachep->spinlock);
+       }
 
        for (i = 0; i < smp_num_cpus; i++) {
                cpucache_t* ccold = new.new[cpu_logical_map(i)];
>>>>>>>>>>>>>>

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
Please read the FAQ at http://www.tux.org/lkml/

Reply via email to