Author: marcel
Date: Wed Jul  2 22:09:06 2014
New Revision: 268185
URL: http://svnweb.freebsd.org/changeset/base/268185

Log:
  Improve the KTR logs by naming the values.

Modified:
  head/sys/ia64/ia64/pmap.c

Modified: head/sys/ia64/ia64/pmap.c
==============================================================================
--- head/sys/ia64/ia64/pmap.c   Wed Jul  2 22:06:31 2014        (r268184)
+++ head/sys/ia64/ia64/pmap.c   Wed Jul  2 22:09:06 2014        (r268185)
@@ -484,7 +484,7 @@ void
 pmap_page_init(vm_page_t m)
 {
 
-       CTR2(KTR_PMAP, "%s(%p)", __func__, m);
+       CTR2(KTR_PMAP, "%s(m=%p)", __func__, m);
 
        TAILQ_INIT(&m->md.pv_list);
        m->md.memattr = VM_MEMATTR_DEFAULT;
@@ -623,7 +623,7 @@ void
 pmap_pinit0(pmap_t pmap)
 {
 
-       CTR2(KTR_PMAP, "%s(%p)", __func__, pmap);
+       CTR2(KTR_PMAP, "%s(pm=%p)", __func__, pmap);
 
        PMAP_LOCK_INIT(pmap);
        pmap_pinit_common(pmap);
@@ -637,7 +637,7 @@ int
 pmap_pinit(pmap_t pmap)
 {
 
-       CTR2(KTR_PMAP, "%s(%p)", __func__, pmap);
+       CTR2(KTR_PMAP, "%s(pm=%p)", __func__, pmap);
 
        pmap_pinit_common(pmap);
        return (1);
@@ -657,7 +657,7 @@ pmap_release(pmap_t pmap)
 {
        int i;
 
-       CTR2(KTR_PMAP, "%s(%p)", __func__, pmap);
+       CTR2(KTR_PMAP, "%s(pm=%p)", __func__, pmap);
 
        for (i = 0; i < IA64_VM_MINKERN_REGION; i++)
                if (pmap->pm_rid[i])
@@ -674,7 +674,7 @@ pmap_growkernel(vm_offset_t addr)
        struct ia64_lpte *leaf;
        vm_page_t nkpg;
 
-       CTR2(KTR_PMAP, "%s(%#x)", __func__, addr);
+       CTR2(KTR_PMAP, "%s(va=%#lx)", __func__, addr);
 
        while (kernel_vm_end <= addr) {
                if (nkpt == PAGE_SIZE/8 + PAGE_SIZE*PAGE_SIZE/64)
@@ -1170,7 +1170,7 @@ pmap_extract(pmap_t pmap, vm_offset_t va
        pmap_t oldpmap;
        vm_paddr_t pa;
 
-       CTR3(KTR_PMAP, "%s(%p, %#x)", __func__, pmap, va);
+       CTR3(KTR_PMAP, "%s(pm=%p, va=%#lx)", __func__, pmap, va);
 
        pa = 0;
        PMAP_LOCK(pmap);
@@ -1198,7 +1198,8 @@ pmap_extract_and_hold(pmap_t pmap, vm_of
        vm_page_t m;
        vm_paddr_t pa;
 
-       CTR4(KTR_PMAP, "%s(%p, %#x, %#x)", __func__, pmap, va, prot);
+       CTR4(KTR_PMAP, "%s(pm=%p, va=%#lx, prot=%#x)", __func__, pmap, va,
+           prot);
 
        pa = 0;
        m = NULL;
@@ -1381,7 +1382,7 @@ pmap_kextract(vm_offset_t va)
        vm_paddr_t pa;
        u_int idx;
 
-       CTR2(KTR_PMAP, "%s(%#x)", __func__, va);
+       CTR2(KTR_PMAP, "%s(va=%#lx)", __func__, va);
 
        KASSERT(va >= VM_MAXUSER_ADDRESS, ("Must be kernel VA"));
 
@@ -1443,7 +1444,7 @@ pmap_qenter(vm_offset_t va, vm_page_t *m
        struct ia64_lpte *pte;
        int i;
 
-       CTR4(KTR_PMAP, "%s(%#x, %p, %d)", __func__, va, m, count);
+       CTR4(KTR_PMAP, "%s(va=%#lx, m_p=%p, cnt=%d)", __func__, va, m, count);
 
        for (i = 0; i < count; i++) {
                pte = pmap_find_kpte(va);
@@ -1468,7 +1469,7 @@ pmap_qremove(vm_offset_t va, int count)
        struct ia64_lpte *pte;
        int i;
 
-       CTR3(KTR_PMAP, "%s(%#x, %d)", __func__, va, count);
+       CTR3(KTR_PMAP, "%s(va=%#lx, cnt=%d)", __func__, va, count);
 
        for (i = 0; i < count; i++) {
                pte = pmap_find_kpte(va);
@@ -1490,7 +1491,7 @@ pmap_kenter(vm_offset_t va, vm_paddr_t p
 {
        struct ia64_lpte *pte;
 
-       CTR3(KTR_PMAP, "%s(%#x, %#x)", __func__, va, pa);
+       CTR3(KTR_PMAP, "%s(va=%#lx, pa=%#lx)", __func__, va, pa);
 
        pte = pmap_find_kpte(va);
        if (pmap_present(pte))
@@ -1510,7 +1511,7 @@ pmap_kremove(vm_offset_t va)
 {
        struct ia64_lpte *pte;
 
-       CTR2(KTR_PMAP, "%s(%#x)", __func__, va);
+       CTR2(KTR_PMAP, "%s(va=%#lx)", __func__, va);
 
        pte = pmap_find_kpte(va);
        if (pmap_present(pte)) {
@@ -1536,8 +1537,8 @@ vm_offset_t
 pmap_map(vm_offset_t *virt, vm_offset_t start, vm_offset_t end, int prot)
 {
 
-       CTR5(KTR_PMAP, "%s(%p, %#x, %#x, %#x)", __func__, virt, start, end,
-           prot);
+       CTR5(KTR_PMAP, "%s(va_p=%p, sva=%#lx, eva=%#lx, prot=%#x)", __func__,
+           virt, start, end, prot);
 
        return IA64_PHYS_TO_RR7(start);
 }
@@ -1558,7 +1559,8 @@ pmap_remove(pmap_t pmap, vm_offset_t sva
        vm_offset_t va;
        struct ia64_lpte *pte;
 
-       CTR4(KTR_PMAP, "%s(%p, %#x, %#x)", __func__, pmap, sva, eva);
+       CTR4(KTR_PMAP, "%s(pm=%p, sva=%#lx, eva=%#lx)", __func__, pmap, sva,
+           eva);
 
        /*
         * Perform an unsynchronized read.  This is, however, safe.
@@ -1597,7 +1599,7 @@ pmap_remove_all(vm_page_t m)
        pmap_t oldpmap;
        pv_entry_t pv;
 
-       CTR2(KTR_PMAP, "%s(%p)", __func__, m);
+       CTR2(KTR_PMAP, "%s(m=%p)", __func__, m);
 
        KASSERT((m->oflags & VPO_UNMANAGED) == 0,
            ("pmap_remove_all: page %p is not managed", m));
@@ -1631,8 +1633,8 @@ pmap_protect(pmap_t pmap, vm_offset_t sv
        pmap_t oldpmap;
        struct ia64_lpte *pte;
 
-       CTR5(KTR_PMAP, "%s(%p, %#x, %#x, %#x)", __func__, pmap, sva, eva,
-           prot);
+       CTR5(KTR_PMAP, "%s(pm=%p, sva=%#lx, eva=%#lx, prot=%#x)", __func__,
+           pmap, sva, eva, prot);
 
        if ((prot & VM_PROT_READ) == VM_PROT_NONE) {
                pmap_remove(pmap, sva, eva);
@@ -1700,8 +1702,8 @@ pmap_enter(pmap_t pmap, vm_offset_t va, 
        struct ia64_lpte *pte;
        boolean_t icache_inval, managed;
 
-       CTR6(KTR_PMAP, "pmap_enter(%p, %#x, %#x, %p, %#x, %u)", pmap, va,
-           access, m, prot, wired);
+       CTR6(KTR_PMAP, "pmap_enter(pm=%p, va=%#lx, acc=%#x, m=%p, prot=%#x, "
+           "wired=%u)", pmap, va, access, m, prot, wired);
 
        rw_wlock(&pvh_global_lock);
        PMAP_LOCK(pmap);
@@ -1834,8 +1836,8 @@ pmap_enter_object(pmap_t pmap, vm_offset
        vm_page_t m;
        vm_pindex_t diff, psize;
 
-       CTR6(KTR_PMAP, "%s(%p, %#x, %#x, %p, %#x)", __func__, pmap, start,
-           end, m_start, prot);
+       CTR6(KTR_PMAP, "%s(pm=%p, sva=%#lx, eva=%#lx, m=%p, prot=%#x)",
+           __func__, pmap, start, end, m_start, prot);
 
        VM_OBJECT_ASSERT_LOCKED(m_start->object);
 
@@ -1866,7 +1868,8 @@ pmap_enter_quick(pmap_t pmap, vm_offset_
 {
        pmap_t oldpmap;
 
-       CTR5(KTR_PMAP, "%s(%p, %#x, %p, %#x)", __func__, pmap, va, m, prot);
+       CTR5(KTR_PMAP, "%s(pm=%p, va=%#lx, m=%p, prot=%#x)", __func__, pmap,
+           va, m, prot);
 
        rw_wlock(&pvh_global_lock);
        PMAP_LOCK(pmap);
@@ -1929,8 +1932,8 @@ pmap_object_init_pt(pmap_t pmap, vm_offs
     vm_pindex_t pindex, vm_size_t size)
 {
 
-       CTR6(KTR_PMAP, "%s(%p, %#x, %p, %u, %#x)", __func__, pmap, addr,
-           object, pindex, size);
+       CTR6(KTR_PMAP, "%s(pm=%p, va=%#lx, obj=%p, idx=%lu, sz=%#lx)",
+           __func__, pmap, addr, object, pindex, size);
 
        VM_OBJECT_ASSERT_WLOCKED(object);
        KASSERT(object->type == OBJT_DEVICE || object->type == OBJT_SG,
@@ -1950,7 +1953,8 @@ pmap_change_wiring(pmap_t pmap, vm_offse
        pmap_t oldpmap;
        struct ia64_lpte *pte;
 
-       CTR4(KTR_PMAP, "%s(%p, %#x, %u)", __func__, pmap, va, wired);
+       CTR4(KTR_PMAP, "%s(pm=%p, va=%#lx, wired=%u)", __func__, pmap, va,
+           wired);
 
        PMAP_LOCK(pmap);
        oldpmap = pmap_switch(pmap);
@@ -1981,8 +1985,8 @@ pmap_copy(pmap_t dst_pmap, pmap_t src_pm
     vm_offset_t src_va)
 {
 
-       CTR6(KTR_PMAP, "%s(%p, %p, %#x, %#x, %#x)", __func__, dst_pmap,
-           src_pmap, dst_va, len, src_va);
+       CTR6(KTR_PMAP, "%s(dpm=%p, spm=%p, dva=%#lx, sz=%#lx, sva=%#lx)",
+           __func__, dst_pmap, src_pmap, dst_va, len, src_va);
 }
 
 /*
@@ -1995,7 +1999,7 @@ pmap_zero_page(vm_page_t m)
 {
        void *p;
 
-       CTR2(KTR_PMAP, "%s(%p)", __func__, m);
+       CTR2(KTR_PMAP, "%s(m=%p)", __func__, m);
 
        p = (void *)pmap_page_to_va(m);
        bzero(p, PAGE_SIZE);
@@ -2013,7 +2017,7 @@ pmap_zero_page_area(vm_page_t m, int off
 {
        char *p;
 
-       CTR4(KTR_PMAP, "%s(%p, %d, %d)", __func__, m, off, size);
+       CTR4(KTR_PMAP, "%s(m=%p, ofs=%d, len=%d)", __func__, m, off, size);
 
        p = (void *)pmap_page_to_va(m);
        bzero(p + off, size);
@@ -2029,7 +2033,7 @@ pmap_zero_page_idle(vm_page_t m)
 {
        void *p;
 
-       CTR2(KTR_PMAP, "%s(%p)", __func__, m);
+       CTR2(KTR_PMAP, "%s(m=%p)", __func__, m);
 
        p = (void *)pmap_page_to_va(m);
        bzero(p, PAGE_SIZE);
@@ -2046,7 +2050,7 @@ pmap_copy_page(vm_page_t msrc, vm_page_t
 {
        void *dst, *src;
 
-       CTR3(KTR_PMAP, "%s(%p, %p)", __func__, msrc, mdst);
+       CTR3(KTR_PMAP, "%s(sm=%p, dm=%p)", __func__, msrc, mdst);
 
        src = (void *)pmap_page_to_va(msrc);
        dst = (void *)pmap_page_to_va(mdst);
@@ -2061,8 +2065,8 @@ pmap_copy_pages(vm_page_t ma[], vm_offse
        vm_offset_t a_pg_offset, b_pg_offset;
        int cnt;
 
-       CTR6(KTR_PMAP, "%s(%p, %#x, %p, %#x, %#x)", __func__, ma,
-           a_offset, mb, b_offset, xfersize);
+       CTR6(KTR_PMAP, "%s(m0=%p, va0=%#lx, m1=%p, va1=%#lx, sz=%#x)",
+           __func__, ma, a_offset, mb, b_offset, xfersize);
 
        while (xfersize > 0) {
                a_pg_offset = a_offset & PAGE_MASK;
@@ -2094,7 +2098,7 @@ pmap_page_exists_quick(pmap_t pmap, vm_p
        int loops = 0;
        boolean_t rv;
 
-       CTR3(KTR_PMAP, "%s(%p, %p)", __func__, pmap, m);
+       CTR3(KTR_PMAP, "%s(pm=%p, m=%p)", __func__, pmap, m);
 
        KASSERT((m->oflags & VPO_UNMANAGED) == 0,
            ("pmap_page_exists_quick: page %p is not managed", m));
@@ -2127,7 +2131,7 @@ pmap_page_wired_mappings(vm_page_t m)
        pv_entry_t pv;
        int count;
 
-       CTR2(KTR_PMAP, "%s(%p)", __func__, m);
+       CTR2(KTR_PMAP, "%s(m=%p)", __func__, m);
 
        count = 0;
        if ((m->oflags & VPO_UNMANAGED) != 0)
@@ -2168,7 +2172,7 @@ pmap_remove_pages(pmap_t pmap)
        u_long inuse, bitmask;
        int allfree, bit, field, idx;
 
-       CTR2(KTR_PMAP, "%s(%p)", __func__, pmap);
+       CTR2(KTR_PMAP, "%s(pm=%p)", __func__, pmap);
 
        rw_wlock(&pvh_global_lock);
        PMAP_LOCK(pmap);
@@ -2237,7 +2241,7 @@ pmap_ts_referenced(vm_page_t m)
        pv_entry_t pv;
        int count = 0;
 
-       CTR2(KTR_PMAP, "%s(%p)", __func__, m);
+       CTR2(KTR_PMAP, "%s(m=%p)", __func__, m);
 
        KASSERT((m->oflags & VPO_UNMANAGED) == 0,
            ("pmap_ts_referenced: page %p is not managed", m));
@@ -2274,7 +2278,7 @@ pmap_is_modified(vm_page_t m)
        pv_entry_t pv;
        boolean_t rv;
 
-       CTR2(KTR_PMAP, "%s(%p)", __func__, m);
+       CTR2(KTR_PMAP, "%s(m=%p)", __func__, m);
 
        KASSERT((m->oflags & VPO_UNMANAGED) == 0,
            ("pmap_is_modified: page %p is not managed", m));
@@ -2316,7 +2320,7 @@ pmap_is_prefaultable(pmap_t pmap, vm_off
 {
        struct ia64_lpte *pte;
 
-       CTR3(KTR_PMAP, "%s(%p, %#x)", __func__, pmap, addr);
+       CTR3(KTR_PMAP, "%s(pm=%p, va=%#lx)", __func__, pmap, addr);
 
        pte = pmap_find_vhpt(addr);
        if (pte != NULL && pmap_present(pte))
@@ -2338,7 +2342,7 @@ pmap_is_referenced(vm_page_t m)
        pv_entry_t pv;
        boolean_t rv;
 
-       CTR2(KTR_PMAP, "%s(%p)", __func__, m);
+       CTR2(KTR_PMAP, "%s(m=%p)", __func__, m);
 
        KASSERT((m->oflags & VPO_UNMANAGED) == 0,
            ("pmap_is_referenced: page %p is not managed", m));
@@ -2372,8 +2376,8 @@ pmap_advise(pmap_t pmap, vm_offset_t sva
        pmap_t oldpmap;
        vm_page_t m;
 
-       CTR5(KTR_PMAP, "%s(%p, %#x, %#x, %d)", __func__, pmap, sva, eva,
-           advice);
+       CTR5(KTR_PMAP, "%s(pm=%p, sva=%#lx, eva=%#lx, adv=%d)", __func__,
+           pmap, sva, eva, advice);
 
        PMAP_LOCK(pmap);
        oldpmap = pmap_switch(pmap);
@@ -2417,7 +2421,7 @@ pmap_clear_modify(vm_page_t m)
        pmap_t oldpmap, pmap;
        pv_entry_t pv;
 
-       CTR2(KTR_PMAP, "%s(%p)", __func__, m);
+       CTR2(KTR_PMAP, "%s(m=%p)", __func__, m);
 
        KASSERT((m->oflags & VPO_UNMANAGED) == 0,
            ("pmap_clear_modify: page %p is not managed", m));
@@ -2460,7 +2464,7 @@ pmap_remove_write(vm_page_t m)
        pv_entry_t pv;
        vm_prot_t prot;
 
-       CTR2(KTR_PMAP, "%s(%p)", __func__, m);
+       CTR2(KTR_PMAP, "%s(m=%p)", __func__, m);
 
        KASSERT((m->oflags & VPO_UNMANAGED) == 0,
            ("pmap_remove_write: page %p is not managed", m));
@@ -2540,7 +2544,8 @@ pmap_mapdev_attr(vm_paddr_t pa, vm_size_
 {
        vm_offset_t va;
 
-       CTR4(KTR_PMAP, "%s(%#x, %#x, %#x)", __func__, pa, sz, attr);
+       CTR4(KTR_PMAP, "%s(pa=%#lx, sz=%#lx, attr=%#x)", __func__, pa, sz,
+           attr);
 
        va = pmap_mapdev_priv(pa, sz, attr);
        return ((void *)(uintptr_t)va);
@@ -2553,7 +2558,7 @@ void
 pmap_unmapdev(vm_offset_t va, vm_size_t size)
 {
 
-       CTR3(KTR_PMAP, "%s(%#x, %#x)", __func__, va, size);
+       CTR3(KTR_PMAP, "%s(va=%#lx, sz=%#lx)", __func__, va, size);
 }
 
 /*
@@ -2579,7 +2584,7 @@ pmap_page_set_memattr(vm_page_t m, vm_me
        pv_entry_t pv;
        void *va;
 
-       CTR3(KTR_PMAP, "%s(%p, %#x)", __func__, m, ma);
+       CTR3(KTR_PMAP, "%s(m=%p, attr=%#x)", __func__, m, ma);
 
        rw_wlock(&pvh_global_lock);
        m->md.memattr = ma;
@@ -2627,7 +2632,8 @@ pmap_mincore(pmap_t pmap, vm_offset_t ad
        vm_paddr_t pa;
        int val;
 
-       CTR4(KTR_PMAP, "%s(%p, %#x, %p)", __func__, pmap, addr, locked_pa);
+       CTR4(KTR_PMAP, "%s(pm=%p, va=%#lx, pa_p=%p)", __func__, pmap, addr,
+           locked_pa);
 
        PMAP_LOCK(pmap);
 retry:
@@ -2668,7 +2674,7 @@ void
 pmap_activate(struct thread *td)
 {
 
-       CTR2(KTR_PMAP, "%s(%p)", __func__, td);
+       CTR2(KTR_PMAP, "%s(td=%p)", __func__, td);
 
        pmap_switch(vmspace_pmap(td->td_proc->p_vmspace));
 }
@@ -2713,7 +2719,7 @@ pmap_sync_icache(pmap_t pm, vm_offset_t 
        vm_offset_t lim;
        vm_size_t len;
 
-       CTR4(KTR_PMAP, "%s(%p, %#x, %#x)", __func__, pm, va, sz);
+       CTR4(KTR_PMAP, "%s(pm=%p, va=%#lx, sz=%#lx)", __func__, pm, va, sz);
 
        sz += va & 31;
        va &= ~31;
@@ -2743,8 +2749,8 @@ pmap_align_superpage(vm_object_t object,
     vm_offset_t *addr, vm_size_t size)
 {
 
-       CTR5(KTR_PMAP, "%s(%p, %#x, %p, %#x)", __func__, object, offset, addr,
-           size);
+       CTR5(KTR_PMAP, "%s(obj=%p, ofs=%#lx, va_p=%p, sz=%#lx)", __func__,
+           object, offset, addr, size);
 }
 
 #include "opt_ddb.h"
_______________________________________________
svn-src-all@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/svn-src-all
To unsubscribe, send any mail to "svn-src-all-unsubscr...@freebsd.org"

Reply via email to