Module Name: src Committed By: maxv Date: Sat Oct 5 07:19:49 UTC 2019
Modified Files: src/sys/arch/x86/x86: pmap.c Log Message: Switch to the new PTE naming: PG_PVLIST -> PTE_PVLIST PG_W -> PTE_WIRED PG_FRAME -> PTE_FRAME No functional change. To generate a diff of this commit: cvs rdiff -u -r1.335 -r1.336 src/sys/arch/x86/x86/pmap.c Please note that diffs are not public domain; they are subject to the copyright notices on the relevant files.
Modified files: Index: src/sys/arch/x86/x86/pmap.c diff -u src/sys/arch/x86/x86/pmap.c:1.335 src/sys/arch/x86/x86/pmap.c:1.336 --- src/sys/arch/x86/x86/pmap.c:1.335 Wed Aug 7 06:23:48 2019 +++ src/sys/arch/x86/x86/pmap.c Sat Oct 5 07:19:49 2019 @@ -1,4 +1,4 @@ -/* $NetBSD: pmap.c,v 1.335 2019/08/07 06:23:48 maxv Exp $ */ +/* $NetBSD: pmap.c,v 1.336 2019/10/05 07:19:49 maxv Exp $ */ /* * Copyright (c) 2008, 2010, 2016, 2017 The NetBSD Foundation, Inc. @@ -130,7 +130,7 @@ */ #include <sys/cdefs.h> -__KERNEL_RCSID(0, "$NetBSD: pmap.c,v 1.335 2019/08/07 06:23:48 maxv Exp $"); +__KERNEL_RCSID(0, "$NetBSD: pmap.c,v 1.336 2019/10/05 07:19:49 maxv Exp $"); #include "opt_user_ldt.h" #include "opt_lockdebug.h" @@ -299,10 +299,10 @@ struct pmap *const kernel_pmap_ptr = &ke struct bootspace bootspace __read_mostly; struct slotspace slotspace __read_mostly; -/* Set to PG_NX if supported. */ +/* Set to PTE_NX if supported. */ pd_entry_t pmap_pg_nx __read_mostly = 0; -/* Set to PG_G if supported. */ +/* Set to PTE_G if supported. */ pd_entry_t pmap_pg_g __read_mostly = 0; /* Set to true if large pages are supported. */ @@ -512,10 +512,10 @@ static inline void pmap_stats_update_bypte(struct pmap *pmap, pt_entry_t npte, pt_entry_t opte) { int resid_diff = ((npte & PTE_P) ? 1 : 0) - ((opte & PTE_P) ? 1 : 0); - int wired_diff = ((npte & PG_W) ? 1 : 0) - ((opte & PG_W) ? 1 : 0); + int wired_diff = ((npte & PTE_WIRED) ? 1 : 0) - ((opte & PTE_WIRED) ? 1 : 0); - KASSERT((npte & (PTE_P | PG_W)) != PG_W); - KASSERT((opte & (PTE_P | PG_W)) != PG_W); + KASSERT((npte & (PTE_P | PTE_WIRED)) != PTE_WIRED); + KASSERT((opte & (PTE_P | PTE_WIRED)) != PTE_WIRED); pmap_stats_update(pmap, resid_diff, wired_diff); } @@ -973,7 +973,7 @@ pmap_kremove1(vaddr_t sva, vsize_t len, } KASSERTMSG((opte & PTE_PS) == 0, "va %#" PRIxVADDR " is a large page", va); - KASSERTMSG((opte & PG_PVLIST) == 0, + KASSERTMSG((opte & PTE_PVLIST) == 0, "va %#" PRIxVADDR " is a pv tracked page", va); } if (localonly) { @@ -3507,12 +3507,12 @@ pmap_remove_pte(struct pmap *pmap, struc /* * If we are not on a pv_head list - we are done. */ - if ((opte & PG_PVLIST) == 0) { + if ((opte & PTE_PVLIST) == 0) { #ifndef DOM0OPS KASSERTMSG((PHYS_TO_VM_PAGE(pmap_pte2pa(opte)) == NULL), - "managed page without PG_PVLIST for %#"PRIxVADDR, va); + "managed page without PTE_PVLIST for %#"PRIxVADDR, va); KASSERTMSG((pmap_pv_tracked(pmap_pte2pa(opte)) == NULL), - "pv-tracked page without PG_PVLIST for %#"PRIxVADDR, va); + "pv-tracked page without PTE_PVLIST for %#"PRIxVADDR, va); #endif return true; } @@ -3522,7 +3522,7 @@ pmap_remove_pte(struct pmap *pmap, struc pp = VM_PAGE_TO_PP(pg); } else if ((pp = pmap_pv_tracked(pmap_pte2pa(opte))) == NULL) { paddr_t pa = pmap_pte2pa(opte); - panic("%s: PG_PVLIST with pv-untracked page" + panic("%s: PTE_PVLIST with pv-untracked page" " va = %#"PRIxVADDR"pa = %#"PRIxPADDR" (%#"PRIxPADDR")", __func__, va, pa, atop(pa)); } @@ -4124,8 +4124,8 @@ pmap_unwire(struct pmap *pmap, vaddr_t v opte = *ptep; KASSERT(pmap_valid_entry(opte)); - if (opte & PG_W) { - pt_entry_t npte = opte & ~PG_W; + if (opte & PTE_WIRED) { + pt_entry_t npte = opte & ~PTE_WIRED; opte = pmap_pte_testset(ptep, npte); pmap_stats_update_bypte(pmap, npte, opte); @@ -4202,7 +4202,7 @@ pmap_enter_ma(struct pmap *pmap, vaddr_t npte = ma | protection_codes[prot] | PTE_P; npte |= pmap_pat_flags(flags); if (wired) - npte |= PG_W; + npte |= PTE_WIRED; if (va < VM_MAXUSER_ADDRESS) npte |= PTE_U; @@ -4224,11 +4224,11 @@ pmap_enter_ma(struct pmap *pmap, vaddr_t new_pg = PHYS_TO_VM_PAGE(pa); if (new_pg != NULL) { /* This is a managed page */ - npte |= PG_PVLIST; + npte |= PTE_PVLIST; new_pp = VM_PAGE_TO_PP(new_pg); } else if ((new_pp = pmap_pv_tracked(pa)) != NULL) { /* This is an unmanaged pv-tracked page */ - npte |= PG_PVLIST; + npte |= PTE_PVLIST; } else { new_pp = NULL; } @@ -4334,19 +4334,19 @@ pmap_enter_ma(struct pmap *pmap, vaddr_t * If the same page, we can skip pv_entry handling. */ if (((opte ^ npte) & (PTE_FRAME | PTE_P)) == 0) { - KASSERT(((opte ^ npte) & PG_PVLIST) == 0); + KASSERT(((opte ^ npte) & PTE_PVLIST) == 0); goto same_pa; } /* * If old page is pv-tracked, remove pv_entry from its list. */ - if ((~opte & (PTE_P | PG_PVLIST)) == 0) { + if ((~opte & (PTE_P | PTE_PVLIST)) == 0) { if ((old_pg = PHYS_TO_VM_PAGE(oldpa)) != NULL) { KASSERT(uvm_page_locked_p(old_pg)); old_pp = VM_PAGE_TO_PP(old_pg); } else if ((old_pp = pmap_pv_tracked(oldpa)) == NULL) { - panic("%s: PG_PVLIST with pv-untracked page" + panic("%s: PTE_PVLIST with pv-untracked page" " va = %#"PRIxVADDR " pa = %#" PRIxPADDR " (%#" PRIxPADDR ")", __func__, va, oldpa, atop(pa)); @@ -5217,7 +5217,7 @@ pmap_ept_enter(struct pmap *pmap, vaddr_ /* * if the same page, inherit PTE_A and PTE_D. */ - if (((opte ^ npte) & (PG_FRAME | EPT_R)) == 0) { + if (((opte ^ npte) & (PTE_FRAME | EPT_R)) == 0) { npte |= opte & (EPT_A | EPT_D); } } while (pmap_pte_cas(ptep, opte, npte) != opte); @@ -5234,7 +5234,7 @@ pmap_ept_enter(struct pmap *pmap, vaddr_ /* * If the same page, we can skip pv_entry handling. */ - if (((opte ^ npte) & (PG_FRAME | EPT_R)) == 0) { + if (((opte ^ npte) & (PTE_FRAME | EPT_R)) == 0) { KASSERT(((opte ^ npte) & EPT_PVLIST) == 0); goto same_pa; } @@ -5272,7 +5272,7 @@ same_pa: } else { accessed = (opte & EPT_R) != 0; } - if (accessed && ((opte ^ npte) & (PG_FRAME | EPT_W)) != 0) { + if (accessed && ((opte ^ npte) & (PTE_FRAME | EPT_W)) != 0) { pmap_tlb_shootdown(pmap, va, 0, TLBSHOOT_ENTER); } @@ -5541,7 +5541,7 @@ pmap_ept_sync_pv(struct vm_page *ptp, va KASSERT((opte & (EPT_D | EPT_A)) != EPT_D); KASSERT((opte & (EPT_A | EPT_R)) != EPT_A); KASSERT(opte == 0 || (opte & EPT_R) != 0); - if ((opte & (PG_FRAME | EPT_R)) != expect) { + if ((opte & (PTE_FRAME | EPT_R)) != expect) { /* * We lost a race with a V->P operation like * pmap_remove(). Wait for the competitor @@ -5630,8 +5630,8 @@ pmap_ept_write_protect(struct pmap *pmap if (!(prot & VM_PROT_WRITE)) bit_rem = EPT_W; - sva &= PG_FRAME; - eva &= PG_FRAME; + sva &= PTE_FRAME; + eva &= PTE_FRAME; /* Acquire pmap. */ kpreempt_disable();