Author: alc
Date: Sun May 20 18:25:45 2012
New Revision: 235695
URL: http://svn.freebsd.org/changeset/base/235695

Log:
  Replace all uses of the vm page queues lock by a r/w lock that is private
  to this pmap.c.  This new r/w lock is used primarily to synchronize access
  to the PV lists.  However, it will be used in a somewhat unconventional
  way.  As finer-grained PV list locking is added to each of the pmap
  functions that acquire this r/w lock, its acquisition will be changed from
  write to read, enabling concurrent execution of the pmap functions with
  finer-grained locking.
  
  Reviewed by:  kib
  X-MFC after:  r235598

Modified:
  head/sys/amd64/amd64/pmap.c

Modified: head/sys/amd64/amd64/pmap.c
==============================================================================
--- head/sys/amd64/amd64/pmap.c Sun May 20 18:07:35 2012        (r235694)
+++ head/sys/amd64/amd64/pmap.c Sun May 20 18:25:45 2012        (r235695)
@@ -117,6 +117,7 @@ __FBSDID("$FreeBSD$");
 #include <sys/mman.h>
 #include <sys/mutex.h>
 #include <sys/proc.h>
+#include <sys/rwlock.h>
 #include <sys/sx.h>
 #include <sys/vmmeter.h>
 #include <sys/sched.h>
@@ -204,6 +205,7 @@ static u_int64_t    DMPDPphys;      /* phys addr
 static TAILQ_HEAD(pch, pv_chunk) pv_chunks = TAILQ_HEAD_INITIALIZER(pv_chunks);
 static long pv_entry_count;
 static struct md_page *pv_table;
+static struct rwlock pvh_global_lock;
 
 /*
  * All those kernel PT submaps that BSD is so fond of
@@ -581,6 +583,11 @@ pmap_bootstrap(vm_paddr_t *firstaddr)
        CPU_FILL(&kernel_pmap->pm_active);      /* don't allow deactivation */
        TAILQ_INIT(&kernel_pmap->pm_pvchunk);
 
+       /*
+        * Initialize the global pv list lock.
+        */
+       rw_init(&pvh_global_lock, "pvh global");
+
        /*
         * Reserve some special page table entries/VA space for temporary
         * mapping of pages.
@@ -1651,9 +1658,9 @@ _pmap_allocpte(pmap_t pmap, vm_pindex_t 
            VM_ALLOC_WIRED | VM_ALLOC_ZERO)) == NULL) {
                if (flags & M_WAITOK) {
                        PMAP_UNLOCK(pmap);
-                       vm_page_unlock_queues();
+                       rw_wunlock(&pvh_global_lock);
                        VM_WAIT;
-                       vm_page_lock_queues();
+                       rw_wlock(&pvh_global_lock);
                        PMAP_LOCK(pmap);
                }
 
@@ -2055,7 +2062,7 @@ pmap_pv_reclaim(pmap_t locked_pmap)
        uint64_t inuse, freemask;
        int bit, field, freed;
        
-       mtx_assert(&vm_page_queue_mtx, MA_OWNED);
+       rw_assert(&pvh_global_lock, RA_WLOCKED);
        PMAP_LOCK_ASSERT(locked_pmap, MA_OWNED);
        pmap = NULL;
        free = m = NULL;
@@ -2171,7 +2178,7 @@ free_pv_entry(pmap_t pmap, pv_entry_t pv
        struct pv_chunk *pc;
        int idx, field, bit;
 
-       mtx_assert(&vm_page_queue_mtx, MA_OWNED);
+       rw_assert(&pvh_global_lock, RA_WLOCKED);
        PMAP_LOCK_ASSERT(pmap, MA_OWNED);
        PV_STAT(pv_entry_frees++);
        PV_STAT(pv_entry_spare++);
@@ -2214,8 +2221,8 @@ get_pv_entry(pmap_t pmap, boolean_t try)
        struct pv_chunk *pc;
        vm_page_t m;
 
+       rw_assert(&pvh_global_lock, RA_WLOCKED);
        PMAP_LOCK_ASSERT(pmap, MA_OWNED);
-       mtx_assert(&vm_page_queue_mtx, MA_OWNED);
        PV_STAT(pv_entry_allocs++);
 retry:
        pc = TAILQ_FIRST(&pmap->pm_pvchunk);
@@ -2284,7 +2291,7 @@ pmap_pvh_remove(struct md_page *pvh, pma
 {
        pv_entry_t pv;
 
-       mtx_assert(&vm_page_queue_mtx, MA_OWNED);
+       rw_assert(&pvh_global_lock, RA_WLOCKED);
        TAILQ_FOREACH(pv, &pvh->pv_list, pv_list) {
                if (pmap == PV_PMAP(pv) && va == pv->pv_va) {
                        TAILQ_REMOVE(&pvh->pv_list, pv, pv_list);
@@ -2307,7 +2314,7 @@ pmap_pv_demote_pde(pmap_t pmap, vm_offse
        vm_offset_t va_last;
        vm_page_t m;
 
-       mtx_assert(&vm_page_queue_mtx, MA_OWNED);
+       rw_assert(&pvh_global_lock, RA_WLOCKED);
        KASSERT((pa & PDRMASK) == 0,
            ("pmap_pv_demote_pde: pa is not 2mpage aligned"));
 
@@ -2345,7 +2352,7 @@ pmap_pv_promote_pde(pmap_t pmap, vm_offs
        vm_offset_t va_last;
        vm_page_t m;
 
-       mtx_assert(&vm_page_queue_mtx, MA_OWNED);
+       rw_assert(&pvh_global_lock, RA_WLOCKED);
        KASSERT((pa & PDRMASK) == 0,
            ("pmap_pv_promote_pde: pa is not 2mpage aligned"));
 
@@ -2391,7 +2398,7 @@ pmap_remove_entry(pmap_t pmap, vm_page_t
 {
        struct md_page *pvh;
 
-       mtx_assert(&vm_page_queue_mtx, MA_OWNED);
+       rw_assert(&pvh_global_lock, RA_WLOCKED);
        pmap_pvh_free(&m->md, pmap, va);
        if (TAILQ_EMPTY(&m->md.pv_list) && (m->flags & PG_FICTITIOUS) == 0) {
                pvh = pa_to_pvh(VM_PAGE_TO_PHYS(m));
@@ -2409,8 +2416,8 @@ pmap_insert_entry(pmap_t pmap, vm_offset
 {
        pv_entry_t pv;
 
+       rw_assert(&pvh_global_lock, RA_WLOCKED);
        PMAP_LOCK_ASSERT(pmap, MA_OWNED);
-       mtx_assert(&vm_page_queue_mtx, MA_OWNED);
        pv = get_pv_entry(pmap, FALSE);
        pv->pv_va = va;
        TAILQ_INSERT_TAIL(&m->md.pv_list, pv, pv_list);
@@ -2424,8 +2431,8 @@ pmap_try_insert_pv_entry(pmap_t pmap, vm
 {
        pv_entry_t pv;
 
+       rw_assert(&pvh_global_lock, RA_WLOCKED);
        PMAP_LOCK_ASSERT(pmap, MA_OWNED);
-       mtx_assert(&vm_page_queue_mtx, MA_OWNED);
        if ((pv = get_pv_entry(pmap, TRUE)) != NULL) {
                pv->pv_va = va;
                TAILQ_INSERT_TAIL(&m->md.pv_list, pv, pv_list);
@@ -2443,7 +2450,7 @@ pmap_pv_insert_pde(pmap_t pmap, vm_offse
        struct md_page *pvh;
        pv_entry_t pv;
 
-       mtx_assert(&vm_page_queue_mtx, MA_OWNED);
+       rw_assert(&pvh_global_lock, RA_WLOCKED);
        if ((pv = get_pv_entry(pmap, TRUE)) != NULL) {
                pv->pv_va = va;
                pvh = pa_to_pvh(pa);
@@ -2711,7 +2718,7 @@ pmap_remove(pmap_t pmap, vm_offset_t sva
 
        anyvalid = 0;
 
-       vm_page_lock_queues();
+       rw_wlock(&pvh_global_lock);
        PMAP_LOCK(pmap);
 
        /*
@@ -2821,7 +2828,7 @@ pmap_remove(pmap_t pmap, vm_offset_t sva
 out:
        if (anyvalid)
                pmap_invalidate_all(pmap);
-       vm_page_unlock_queues();        
+       rw_wunlock(&pvh_global_lock);   
        PMAP_UNLOCK(pmap);
        pmap_free_zero_pages(free);
 }
@@ -2853,7 +2860,7 @@ pmap_remove_all(vm_page_t m)
        KASSERT((m->oflags & VPO_UNMANAGED) == 0,
            ("pmap_remove_all: page %p is not managed", m));
        free = NULL;
-       vm_page_lock_queues();
+       rw_wlock(&pvh_global_lock);
        if ((m->flags & PG_FICTITIOUS) != 0)
                goto small_mappings;
        pvh = pa_to_pvh(VM_PAGE_TO_PHYS(m));
@@ -2892,7 +2899,7 @@ small_mappings:
                PMAP_UNLOCK(pmap);
        }
        vm_page_aflag_clear(m, PGA_WRITEABLE);
-       vm_page_unlock_queues();
+       rw_wunlock(&pvh_global_lock);
        pmap_free_zero_pages(free);
 }
 
@@ -3013,12 +3020,12 @@ resume:
                        } else {
                                if (!pv_lists_locked) {
                                        pv_lists_locked = TRUE;
-                                       if (!mtx_trylock(&vm_page_queue_mtx)) {
+                                       if (!rw_try_wlock(&pvh_global_lock)) {
                                                if (anychanged)
                                                        pmap_invalidate_all(
                                                            pmap);
                                                PMAP_UNLOCK(pmap);
-                                               vm_page_lock_queues();
+                                               rw_wlock(&pvh_global_lock);
                                                goto resume;
                                        }
                                }
@@ -3069,7 +3076,7 @@ retry:
        if (anychanged)
                pmap_invalidate_all(pmap);
        if (pv_lists_locked)
-               vm_page_unlock_queues();
+               rw_wunlock(&pvh_global_lock);
        PMAP_UNLOCK(pmap);
 }
 
@@ -3224,7 +3231,7 @@ pmap_enter(pmap_t pmap, vm_offset_t va, 
 
        mpte = NULL;
 
-       vm_page_lock_queues();
+       rw_wlock(&pvh_global_lock);
        PMAP_LOCK(pmap);
 
        /*
@@ -3380,7 +3387,7 @@ validate:
            vm_reserv_level_iffullpop(m) == 0)
                pmap_promote_pde(pmap, pde, va);
 
-       vm_page_unlock_queues();
+       rw_wunlock(&pvh_global_lock);
        PMAP_UNLOCK(pmap);
 }
 
@@ -3396,7 +3403,7 @@ pmap_enter_pde(pmap_t pmap, vm_offset_t 
        pd_entry_t *pde, newpde;
        vm_page_t free, mpde;
 
-       mtx_assert(&vm_page_queue_mtx, MA_OWNED);
+       rw_assert(&pvh_global_lock, RA_WLOCKED);
        PMAP_LOCK_ASSERT(pmap, MA_OWNED);
        if ((mpde = pmap_allocpde(pmap, va, M_NOWAIT)) == NULL) {
                CTR2(KTR_PMAP, "pmap_enter_pde: failure for va %#lx"
@@ -3477,7 +3484,7 @@ pmap_enter_object(pmap_t pmap, vm_offset
        psize = atop(end - start);
        mpte = NULL;
        m = m_start;
-       vm_page_lock_queues();
+       rw_wlock(&pvh_global_lock);
        PMAP_LOCK(pmap);
        while (m != NULL && (diff = m->pindex - m_start->pindex) < psize) {
                va = start + ptoa(diff);
@@ -3491,7 +3498,7 @@ pmap_enter_object(pmap_t pmap, vm_offset
                            mpte);
                m = TAILQ_NEXT(m, listq);
        }
-       vm_page_unlock_queues();
+       rw_wunlock(&pvh_global_lock);
        PMAP_UNLOCK(pmap);
 }
 
@@ -3508,10 +3515,10 @@ void
 pmap_enter_quick(pmap_t pmap, vm_offset_t va, vm_page_t m, vm_prot_t prot)
 {
 
-       vm_page_lock_queues();
+       rw_wlock(&pvh_global_lock);
        PMAP_LOCK(pmap);
        (void)pmap_enter_quick_locked(pmap, va, m, prot, NULL);
-       vm_page_unlock_queues();
+       rw_wunlock(&pvh_global_lock);
        PMAP_UNLOCK(pmap);
 }
 
@@ -3526,7 +3533,7 @@ pmap_enter_quick_locked(pmap_t pmap, vm_
        KASSERT(va < kmi.clean_sva || va >= kmi.clean_eva ||
            (m->oflags & VPO_UNMANAGED) != 0,
            ("pmap_enter_quick_locked: managed mapping within the clean 
submap"));
-       mtx_assert(&vm_page_queue_mtx, MA_OWNED);
+       rw_assert(&pvh_global_lock, RA_WLOCKED);
        PMAP_LOCK_ASSERT(pmap, MA_OWNED);
 
        /*
@@ -3745,9 +3752,9 @@ retry:
                if (!wired != ((*pde & PG_W) == 0)) {
                        if (!are_queues_locked) {
                                are_queues_locked = TRUE;
-                               if (!mtx_trylock(&vm_page_queue_mtx)) {
+                               if (!rw_try_wlock(&pvh_global_lock)) {
                                        PMAP_UNLOCK(pmap);
-                                       vm_page_lock_queues();
+                                       rw_wlock(&pvh_global_lock);
                                        goto retry;
                                }
                        }
@@ -3766,7 +3773,7 @@ retry:
        }
 out:
        if (are_queues_locked)
-               vm_page_unlock_queues();
+               rw_wunlock(&pvh_global_lock);
        PMAP_UNLOCK(pmap);
 }
 
@@ -3790,7 +3797,7 @@ pmap_copy(pmap_t dst_pmap, pmap_t src_pm
        if (dst_addr != src_addr)
                return;
 
-       vm_page_lock_queues();
+       rw_wlock(&pvh_global_lock);
        if (dst_pmap < src_pmap) {
                PMAP_LOCK(dst_pmap);
                PMAP_LOCK(src_pmap);
@@ -3906,7 +3913,7 @@ pmap_copy(pmap_t dst_pmap, pmap_t src_pm
                }
        }
 out:
-       vm_page_unlock_queues();
+       rw_wunlock(&pvh_global_lock);
        PMAP_UNLOCK(src_pmap);
        PMAP_UNLOCK(dst_pmap);
 }      
@@ -3987,7 +3994,7 @@ pmap_page_exists_quick(pmap_t pmap, vm_p
        KASSERT((m->oflags & VPO_UNMANAGED) == 0,
            ("pmap_page_exists_quick: page %p is not managed", m));
        rv = FALSE;
-       vm_page_lock_queues();
+       rw_wlock(&pvh_global_lock);
        TAILQ_FOREACH(pv, &m->md.pv_list, pv_list) {
                if (PV_PMAP(pv) == pmap) {
                        rv = TRUE;
@@ -4009,7 +4016,7 @@ pmap_page_exists_quick(pmap_t pmap, vm_p
                                break;
                }
        }
-       vm_page_unlock_queues();
+       rw_wunlock(&pvh_global_lock);
        return (rv);
 }
 
@@ -4027,13 +4034,13 @@ pmap_page_wired_mappings(vm_page_t m)
        count = 0;
        if ((m->oflags & VPO_UNMANAGED) != 0)
                return (count);
-       vm_page_lock_queues();
+       rw_wlock(&pvh_global_lock);
        count = pmap_pvh_wired_mappings(&m->md, count);
        if ((m->flags & PG_FICTITIOUS) == 0) {
            count = pmap_pvh_wired_mappings(pa_to_pvh(VM_PAGE_TO_PHYS(m)),
                count);
        }
-       vm_page_unlock_queues();
+       rw_wunlock(&pvh_global_lock);
        return (count);
 }
 
@@ -4049,7 +4056,7 @@ pmap_pvh_wired_mappings(struct md_page *
        pt_entry_t *pte;
        pv_entry_t pv;
 
-       mtx_assert(&vm_page_queue_mtx, MA_OWNED);
+       rw_assert(&pvh_global_lock, RA_WLOCKED);
        TAILQ_FOREACH(pv, &pvh->pv_list, pv_list) {
                pmap = PV_PMAP(pv);
                PMAP_LOCK(pmap);
@@ -4072,11 +4079,11 @@ pmap_page_is_mapped(vm_page_t m)
 
        if ((m->oflags & VPO_UNMANAGED) != 0)
                return (FALSE);
-       vm_page_lock_queues();
+       rw_wlock(&pvh_global_lock);
        rv = !TAILQ_EMPTY(&m->md.pv_list) ||
            ((m->flags & PG_FICTITIOUS) == 0 &&
            !TAILQ_EMPTY(&pa_to_pvh(VM_PAGE_TO_PHYS(m))->pv_list));
-       vm_page_unlock_queues();
+       rw_wunlock(&pvh_global_lock);
        return (rv);
 }
 
@@ -4107,7 +4114,7 @@ pmap_remove_pages(pmap_t pmap)
                printf("warning: pmap_remove_pages called with non-current 
pmap\n");
                return;
        }
-       vm_page_lock_queues();
+       rw_wlock(&pvh_global_lock);
        PMAP_LOCK(pmap);
        TAILQ_FOREACH_SAFE(pc, &pmap->pm_pvchunk, pc_list, npc) {
                allfree = 1;
@@ -4216,7 +4223,7 @@ pmap_remove_pages(pmap_t pmap)
                }
        }
        pmap_invalidate_all(pmap);
-       vm_page_unlock_queues();
+       rw_wunlock(&pvh_global_lock);
        PMAP_UNLOCK(pmap);
        pmap_free_zero_pages(free);
 }
@@ -4244,11 +4251,11 @@ pmap_is_modified(vm_page_t m)
        if ((m->oflags & VPO_BUSY) == 0 &&
            (m->aflags & PGA_WRITEABLE) == 0)
                return (FALSE);
-       vm_page_lock_queues();
+       rw_wlock(&pvh_global_lock);
        rv = pmap_is_modified_pvh(&m->md) ||
            ((m->flags & PG_FICTITIOUS) == 0 &&
            pmap_is_modified_pvh(pa_to_pvh(VM_PAGE_TO_PHYS(m))));
-       vm_page_unlock_queues();
+       rw_wunlock(&pvh_global_lock);
        return (rv);
 }
 
@@ -4265,7 +4272,7 @@ pmap_is_modified_pvh(struct md_page *pvh
        pmap_t pmap;
        boolean_t rv;
 
-       mtx_assert(&vm_page_queue_mtx, MA_OWNED);
+       rw_assert(&pvh_global_lock, RA_WLOCKED);
        rv = FALSE;
        TAILQ_FOREACH(pv, &pvh->pv_list, pv_list) {
                pmap = PV_PMAP(pv);
@@ -4316,11 +4323,11 @@ pmap_is_referenced(vm_page_t m)
 
        KASSERT((m->oflags & VPO_UNMANAGED) == 0,
            ("pmap_is_referenced: page %p is not managed", m));
-       vm_page_lock_queues();
+       rw_wlock(&pvh_global_lock);
        rv = pmap_is_referenced_pvh(&m->md) ||
            ((m->flags & PG_FICTITIOUS) == 0 &&
            pmap_is_referenced_pvh(pa_to_pvh(VM_PAGE_TO_PHYS(m))));
-       vm_page_unlock_queues();
+       rw_wunlock(&pvh_global_lock);
        return (rv);
 }
 
@@ -4336,7 +4343,7 @@ pmap_is_referenced_pvh(struct md_page *p
        pmap_t pmap;
        boolean_t rv;
 
-       mtx_assert(&vm_page_queue_mtx, MA_OWNED);
+       rw_assert(&pvh_global_lock, RA_WLOCKED);
        rv = FALSE;
        TAILQ_FOREACH(pv, &pvh->pv_list, pv_list) {
                pmap = PV_PMAP(pv);
@@ -4375,7 +4382,7 @@ pmap_remove_write(vm_page_t m)
        if ((m->oflags & VPO_BUSY) == 0 &&
            (m->aflags & PGA_WRITEABLE) == 0)
                return;
-       vm_page_lock_queues();
+       rw_wlock(&pvh_global_lock);
        if ((m->flags & PG_FICTITIOUS) != 0)
                goto small_mappings;
        pvh = pa_to_pvh(VM_PAGE_TO_PHYS(m));
@@ -4409,7 +4416,7 @@ retry:
                PMAP_UNLOCK(pmap);
        }
        vm_page_aflag_clear(m, PGA_WRITEABLE);
-       vm_page_unlock_queues();
+       rw_wunlock(&pvh_global_lock);
 }
 
 /*
@@ -4437,7 +4444,7 @@ pmap_ts_referenced(vm_page_t m)
 
        KASSERT((m->oflags & VPO_UNMANAGED) == 0,
            ("pmap_ts_referenced: page %p is not managed", m));
-       vm_page_lock_queues();
+       rw_wlock(&pvh_global_lock);
        if ((m->flags & PG_FICTITIOUS) != 0)
                goto small_mappings;
        pvh = pa_to_pvh(VM_PAGE_TO_PHYS(m));
@@ -4495,7 +4502,7 @@ small_mappings:
                } while ((pv = pvn) != NULL && pv != pvf);
        }
 out:
-       vm_page_unlock_queues();
+       rw_wunlock(&pvh_global_lock);
        return (rtval);
 }
 
@@ -4525,7 +4532,7 @@ pmap_clear_modify(vm_page_t m)
         */
        if ((m->aflags & PGA_WRITEABLE) == 0)
                return;
-       vm_page_lock_queues();
+       rw_wlock(&pvh_global_lock);
        if ((m->flags & PG_FICTITIOUS) != 0)
                goto small_mappings;
        pvh = pa_to_pvh(VM_PAGE_TO_PHYS(m));
@@ -4574,7 +4581,7 @@ small_mappings:
                }
                PMAP_UNLOCK(pmap);
        }
-       vm_page_unlock_queues();
+       rw_wunlock(&pvh_global_lock);
 }
 
 /*
@@ -4594,7 +4601,7 @@ pmap_clear_reference(vm_page_t m)
 
        KASSERT((m->oflags & VPO_UNMANAGED) == 0,
            ("pmap_clear_reference: page %p is not managed", m));
-       vm_page_lock_queues();
+       rw_wlock(&pvh_global_lock);
        if ((m->flags & PG_FICTITIOUS) != 0)
                goto small_mappings;
        pvh = pa_to_pvh(VM_PAGE_TO_PHYS(m));
@@ -4634,7 +4641,7 @@ small_mappings:
                }
                PMAP_UNLOCK(pmap);
        }
-       vm_page_unlock_queues();
+       rw_wunlock(&pvh_global_lock);
 }
 
 /*
_______________________________________________
svn-src-all@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/svn-src-all
To unsubscribe, send any mail to "svn-src-all-unsubscr...@freebsd.org"

Reply via email to