From owner-svn-src-all@FreeBSD.ORG Wed Jul 2 22:09:07 2014 Return-Path: Delivered-To: svn-src-all@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:1900:2254:206a::19:1]) (using TLSv1 with cipher ADH-AES256-SHA (256/256 bits)) (No client certificate requested) by hub.freebsd.org (Postfix) with ESMTPS id 0DC5839D; Wed, 2 Jul 2014 22:09:07 +0000 (UTC) Received: from svn.freebsd.org (svn.freebsd.org [IPv6:2001:1900:2254:2068::e6a:0]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (Client did not present a certificate) by mx1.freebsd.org (Postfix) with ESMTPS id EE7AA2784; Wed, 2 Jul 2014 22:09:06 +0000 (UTC) Received: from svn.freebsd.org ([127.0.1.70]) by svn.freebsd.org (8.14.8/8.14.8) with ESMTP id s62M96Z2022713; Wed, 2 Jul 2014 22:09:06 GMT (envelope-from marcel@svn.freebsd.org) Received: (from marcel@localhost) by svn.freebsd.org (8.14.8/8.14.8/Submit) id s62M96QR022712; Wed, 2 Jul 2014 22:09:06 GMT (envelope-from marcel@svn.freebsd.org) Message-Id: <201407022209.s62M96QR022712@svn.freebsd.org> From: Marcel Moolenaar Date: Wed, 2 Jul 2014 22:09:06 +0000 (UTC) To: src-committers@freebsd.org, svn-src-all@freebsd.org, svn-src-head@freebsd.org Subject: svn commit: r268185 - head/sys/ia64/ia64 X-SVN-Group: head MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit X-BeenThere: svn-src-all@freebsd.org X-Mailman-Version: 2.1.18 Precedence: list List-Id: "SVN commit messages for the entire src tree \(except for " user" and " projects" \)" List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Wed, 02 Jul 2014 22:09:07 -0000 Author: marcel Date: Wed Jul 2 22:09:06 2014 New Revision: 268185 URL: http://svnweb.freebsd.org/changeset/base/268185 Log: Improve the KTR logs by naming the values. Modified: head/sys/ia64/ia64/pmap.c Modified: head/sys/ia64/ia64/pmap.c ============================================================================== --- head/sys/ia64/ia64/pmap.c Wed Jul 2 22:06:31 2014 (r268184) +++ head/sys/ia64/ia64/pmap.c Wed Jul 2 22:09:06 2014 (r268185) @@ -484,7 +484,7 @@ void pmap_page_init(vm_page_t m) { - CTR2(KTR_PMAP, "%s(%p)", __func__, m); + CTR2(KTR_PMAP, "%s(m=%p)", __func__, m); TAILQ_INIT(&m->md.pv_list); m->md.memattr = VM_MEMATTR_DEFAULT; @@ -623,7 +623,7 @@ void pmap_pinit0(pmap_t pmap) { - CTR2(KTR_PMAP, "%s(%p)", __func__, pmap); + CTR2(KTR_PMAP, "%s(pm=%p)", __func__, pmap); PMAP_LOCK_INIT(pmap); pmap_pinit_common(pmap); @@ -637,7 +637,7 @@ int pmap_pinit(pmap_t pmap) { - CTR2(KTR_PMAP, "%s(%p)", __func__, pmap); + CTR2(KTR_PMAP, "%s(pm=%p)", __func__, pmap); pmap_pinit_common(pmap); return (1); @@ -657,7 +657,7 @@ pmap_release(pmap_t pmap) { int i; - CTR2(KTR_PMAP, "%s(%p)", __func__, pmap); + CTR2(KTR_PMAP, "%s(pm=%p)", __func__, pmap); for (i = 0; i < IA64_VM_MINKERN_REGION; i++) if (pmap->pm_rid[i]) @@ -674,7 +674,7 @@ pmap_growkernel(vm_offset_t addr) struct ia64_lpte *leaf; vm_page_t nkpg; - CTR2(KTR_PMAP, "%s(%#x)", __func__, addr); + CTR2(KTR_PMAP, "%s(va=%#lx)", __func__, addr); while (kernel_vm_end <= addr) { if (nkpt == PAGE_SIZE/8 + PAGE_SIZE*PAGE_SIZE/64) @@ -1170,7 +1170,7 @@ pmap_extract(pmap_t pmap, vm_offset_t va pmap_t oldpmap; vm_paddr_t pa; - CTR3(KTR_PMAP, "%s(%p, %#x)", __func__, pmap, va); + CTR3(KTR_PMAP, "%s(pm=%p, va=%#lx)", __func__, pmap, va); pa = 0; PMAP_LOCK(pmap); @@ -1198,7 +1198,8 @@ pmap_extract_and_hold(pmap_t pmap, vm_of vm_page_t m; vm_paddr_t pa; - CTR4(KTR_PMAP, "%s(%p, %#x, %#x)", __func__, pmap, va, prot); + CTR4(KTR_PMAP, "%s(pm=%p, va=%#lx, prot=%#x)", __func__, pmap, va, + prot); pa = 0; m = NULL; @@ -1381,7 +1382,7 @@ pmap_kextract(vm_offset_t va) vm_paddr_t pa; u_int idx; - CTR2(KTR_PMAP, "%s(%#x)", __func__, va); + CTR2(KTR_PMAP, "%s(va=%#lx)", __func__, va); KASSERT(va >= VM_MAXUSER_ADDRESS, ("Must be kernel VA")); @@ -1443,7 +1444,7 @@ pmap_qenter(vm_offset_t va, vm_page_t *m struct ia64_lpte *pte; int i; - CTR4(KTR_PMAP, "%s(%#x, %p, %d)", __func__, va, m, count); + CTR4(KTR_PMAP, "%s(va=%#lx, m_p=%p, cnt=%d)", __func__, va, m, count); for (i = 0; i < count; i++) { pte = pmap_find_kpte(va); @@ -1468,7 +1469,7 @@ pmap_qremove(vm_offset_t va, int count) struct ia64_lpte *pte; int i; - CTR3(KTR_PMAP, "%s(%#x, %d)", __func__, va, count); + CTR3(KTR_PMAP, "%s(va=%#lx, cnt=%d)", __func__, va, count); for (i = 0; i < count; i++) { pte = pmap_find_kpte(va); @@ -1490,7 +1491,7 @@ pmap_kenter(vm_offset_t va, vm_paddr_t p { struct ia64_lpte *pte; - CTR3(KTR_PMAP, "%s(%#x, %#x)", __func__, va, pa); + CTR3(KTR_PMAP, "%s(va=%#lx, pa=%#lx)", __func__, va, pa); pte = pmap_find_kpte(va); if (pmap_present(pte)) @@ -1510,7 +1511,7 @@ pmap_kremove(vm_offset_t va) { struct ia64_lpte *pte; - CTR2(KTR_PMAP, "%s(%#x)", __func__, va); + CTR2(KTR_PMAP, "%s(va=%#lx)", __func__, va); pte = pmap_find_kpte(va); if (pmap_present(pte)) { @@ -1536,8 +1537,8 @@ vm_offset_t pmap_map(vm_offset_t *virt, vm_offset_t start, vm_offset_t end, int prot) { - CTR5(KTR_PMAP, "%s(%p, %#x, %#x, %#x)", __func__, virt, start, end, - prot); + CTR5(KTR_PMAP, "%s(va_p=%p, sva=%#lx, eva=%#lx, prot=%#x)", __func__, + virt, start, end, prot); return IA64_PHYS_TO_RR7(start); } @@ -1558,7 +1559,8 @@ pmap_remove(pmap_t pmap, vm_offset_t sva vm_offset_t va; struct ia64_lpte *pte; - CTR4(KTR_PMAP, "%s(%p, %#x, %#x)", __func__, pmap, sva, eva); + CTR4(KTR_PMAP, "%s(pm=%p, sva=%#lx, eva=%#lx)", __func__, pmap, sva, + eva); /* * Perform an unsynchronized read. This is, however, safe. @@ -1597,7 +1599,7 @@ pmap_remove_all(vm_page_t m) pmap_t oldpmap; pv_entry_t pv; - CTR2(KTR_PMAP, "%s(%p)", __func__, m); + CTR2(KTR_PMAP, "%s(m=%p)", __func__, m); KASSERT((m->oflags & VPO_UNMANAGED) == 0, ("pmap_remove_all: page %p is not managed", m)); @@ -1631,8 +1633,8 @@ pmap_protect(pmap_t pmap, vm_offset_t sv pmap_t oldpmap; struct ia64_lpte *pte; - CTR5(KTR_PMAP, "%s(%p, %#x, %#x, %#x)", __func__, pmap, sva, eva, - prot); + CTR5(KTR_PMAP, "%s(pm=%p, sva=%#lx, eva=%#lx, prot=%#x)", __func__, + pmap, sva, eva, prot); if ((prot & VM_PROT_READ) == VM_PROT_NONE) { pmap_remove(pmap, sva, eva); @@ -1700,8 +1702,8 @@ pmap_enter(pmap_t pmap, vm_offset_t va, struct ia64_lpte *pte; boolean_t icache_inval, managed; - CTR6(KTR_PMAP, "pmap_enter(%p, %#x, %#x, %p, %#x, %u)", pmap, va, - access, m, prot, wired); + CTR6(KTR_PMAP, "pmap_enter(pm=%p, va=%#lx, acc=%#x, m=%p, prot=%#x, " + "wired=%u)", pmap, va, access, m, prot, wired); rw_wlock(&pvh_global_lock); PMAP_LOCK(pmap); @@ -1834,8 +1836,8 @@ pmap_enter_object(pmap_t pmap, vm_offset vm_page_t m; vm_pindex_t diff, psize; - CTR6(KTR_PMAP, "%s(%p, %#x, %#x, %p, %#x)", __func__, pmap, start, - end, m_start, prot); + CTR6(KTR_PMAP, "%s(pm=%p, sva=%#lx, eva=%#lx, m=%p, prot=%#x)", + __func__, pmap, start, end, m_start, prot); VM_OBJECT_ASSERT_LOCKED(m_start->object); @@ -1866,7 +1868,8 @@ pmap_enter_quick(pmap_t pmap, vm_offset_ { pmap_t oldpmap; - CTR5(KTR_PMAP, "%s(%p, %#x, %p, %#x)", __func__, pmap, va, m, prot); + CTR5(KTR_PMAP, "%s(pm=%p, va=%#lx, m=%p, prot=%#x)", __func__, pmap, + va, m, prot); rw_wlock(&pvh_global_lock); PMAP_LOCK(pmap); @@ -1929,8 +1932,8 @@ pmap_object_init_pt(pmap_t pmap, vm_offs vm_pindex_t pindex, vm_size_t size) { - CTR6(KTR_PMAP, "%s(%p, %#x, %p, %u, %#x)", __func__, pmap, addr, - object, pindex, size); + CTR6(KTR_PMAP, "%s(pm=%p, va=%#lx, obj=%p, idx=%lu, sz=%#lx)", + __func__, pmap, addr, object, pindex, size); VM_OBJECT_ASSERT_WLOCKED(object); KASSERT(object->type == OBJT_DEVICE || object->type == OBJT_SG, @@ -1950,7 +1953,8 @@ pmap_change_wiring(pmap_t pmap, vm_offse pmap_t oldpmap; struct ia64_lpte *pte; - CTR4(KTR_PMAP, "%s(%p, %#x, %u)", __func__, pmap, va, wired); + CTR4(KTR_PMAP, "%s(pm=%p, va=%#lx, wired=%u)", __func__, pmap, va, + wired); PMAP_LOCK(pmap); oldpmap = pmap_switch(pmap); @@ -1981,8 +1985,8 @@ pmap_copy(pmap_t dst_pmap, pmap_t src_pm vm_offset_t src_va) { - CTR6(KTR_PMAP, "%s(%p, %p, %#x, %#x, %#x)", __func__, dst_pmap, - src_pmap, dst_va, len, src_va); + CTR6(KTR_PMAP, "%s(dpm=%p, spm=%p, dva=%#lx, sz=%#lx, sva=%#lx)", + __func__, dst_pmap, src_pmap, dst_va, len, src_va); } /* @@ -1995,7 +1999,7 @@ pmap_zero_page(vm_page_t m) { void *p; - CTR2(KTR_PMAP, "%s(%p)", __func__, m); + CTR2(KTR_PMAP, "%s(m=%p)", __func__, m); p = (void *)pmap_page_to_va(m); bzero(p, PAGE_SIZE); @@ -2013,7 +2017,7 @@ pmap_zero_page_area(vm_page_t m, int off { char *p; - CTR4(KTR_PMAP, "%s(%p, %d, %d)", __func__, m, off, size); + CTR4(KTR_PMAP, "%s(m=%p, ofs=%d, len=%d)", __func__, m, off, size); p = (void *)pmap_page_to_va(m); bzero(p + off, size); @@ -2029,7 +2033,7 @@ pmap_zero_page_idle(vm_page_t m) { void *p; - CTR2(KTR_PMAP, "%s(%p)", __func__, m); + CTR2(KTR_PMAP, "%s(m=%p)", __func__, m); p = (void *)pmap_page_to_va(m); bzero(p, PAGE_SIZE); @@ -2046,7 +2050,7 @@ pmap_copy_page(vm_page_t msrc, vm_page_t { void *dst, *src; - CTR3(KTR_PMAP, "%s(%p, %p)", __func__, msrc, mdst); + CTR3(KTR_PMAP, "%s(sm=%p, dm=%p)", __func__, msrc, mdst); src = (void *)pmap_page_to_va(msrc); dst = (void *)pmap_page_to_va(mdst); @@ -2061,8 +2065,8 @@ pmap_copy_pages(vm_page_t ma[], vm_offse vm_offset_t a_pg_offset, b_pg_offset; int cnt; - CTR6(KTR_PMAP, "%s(%p, %#x, %p, %#x, %#x)", __func__, ma, - a_offset, mb, b_offset, xfersize); + CTR6(KTR_PMAP, "%s(m0=%p, va0=%#lx, m1=%p, va1=%#lx, sz=%#x)", + __func__, ma, a_offset, mb, b_offset, xfersize); while (xfersize > 0) { a_pg_offset = a_offset & PAGE_MASK; @@ -2094,7 +2098,7 @@ pmap_page_exists_quick(pmap_t pmap, vm_p int loops = 0; boolean_t rv; - CTR3(KTR_PMAP, "%s(%p, %p)", __func__, pmap, m); + CTR3(KTR_PMAP, "%s(pm=%p, m=%p)", __func__, pmap, m); KASSERT((m->oflags & VPO_UNMANAGED) == 0, ("pmap_page_exists_quick: page %p is not managed", m)); @@ -2127,7 +2131,7 @@ pmap_page_wired_mappings(vm_page_t m) pv_entry_t pv; int count; - CTR2(KTR_PMAP, "%s(%p)", __func__, m); + CTR2(KTR_PMAP, "%s(m=%p)", __func__, m); count = 0; if ((m->oflags & VPO_UNMANAGED) != 0) @@ -2168,7 +2172,7 @@ pmap_remove_pages(pmap_t pmap) u_long inuse, bitmask; int allfree, bit, field, idx; - CTR2(KTR_PMAP, "%s(%p)", __func__, pmap); + CTR2(KTR_PMAP, "%s(pm=%p)", __func__, pmap); rw_wlock(&pvh_global_lock); PMAP_LOCK(pmap); @@ -2237,7 +2241,7 @@ pmap_ts_referenced(vm_page_t m) pv_entry_t pv; int count = 0; - CTR2(KTR_PMAP, "%s(%p)", __func__, m); + CTR2(KTR_PMAP, "%s(m=%p)", __func__, m); KASSERT((m->oflags & VPO_UNMANAGED) == 0, ("pmap_ts_referenced: page %p is not managed", m)); @@ -2274,7 +2278,7 @@ pmap_is_modified(vm_page_t m) pv_entry_t pv; boolean_t rv; - CTR2(KTR_PMAP, "%s(%p)", __func__, m); + CTR2(KTR_PMAP, "%s(m=%p)", __func__, m); KASSERT((m->oflags & VPO_UNMANAGED) == 0, ("pmap_is_modified: page %p is not managed", m)); @@ -2316,7 +2320,7 @@ pmap_is_prefaultable(pmap_t pmap, vm_off { struct ia64_lpte *pte; - CTR3(KTR_PMAP, "%s(%p, %#x)", __func__, pmap, addr); + CTR3(KTR_PMAP, "%s(pm=%p, va=%#lx)", __func__, pmap, addr); pte = pmap_find_vhpt(addr); if (pte != NULL && pmap_present(pte)) @@ -2338,7 +2342,7 @@ pmap_is_referenced(vm_page_t m) pv_entry_t pv; boolean_t rv; - CTR2(KTR_PMAP, "%s(%p)", __func__, m); + CTR2(KTR_PMAP, "%s(m=%p)", __func__, m); KASSERT((m->oflags & VPO_UNMANAGED) == 0, ("pmap_is_referenced: page %p is not managed", m)); @@ -2372,8 +2376,8 @@ pmap_advise(pmap_t pmap, vm_offset_t sva pmap_t oldpmap; vm_page_t m; - CTR5(KTR_PMAP, "%s(%p, %#x, %#x, %d)", __func__, pmap, sva, eva, - advice); + CTR5(KTR_PMAP, "%s(pm=%p, sva=%#lx, eva=%#lx, adv=%d)", __func__, + pmap, sva, eva, advice); PMAP_LOCK(pmap); oldpmap = pmap_switch(pmap); @@ -2417,7 +2421,7 @@ pmap_clear_modify(vm_page_t m) pmap_t oldpmap, pmap; pv_entry_t pv; - CTR2(KTR_PMAP, "%s(%p)", __func__, m); + CTR2(KTR_PMAP, "%s(m=%p)", __func__, m); KASSERT((m->oflags & VPO_UNMANAGED) == 0, ("pmap_clear_modify: page %p is not managed", m)); @@ -2460,7 +2464,7 @@ pmap_remove_write(vm_page_t m) pv_entry_t pv; vm_prot_t prot; - CTR2(KTR_PMAP, "%s(%p)", __func__, m); + CTR2(KTR_PMAP, "%s(m=%p)", __func__, m); KASSERT((m->oflags & VPO_UNMANAGED) == 0, ("pmap_remove_write: page %p is not managed", m)); @@ -2540,7 +2544,8 @@ pmap_mapdev_attr(vm_paddr_t pa, vm_size_ { vm_offset_t va; - CTR4(KTR_PMAP, "%s(%#x, %#x, %#x)", __func__, pa, sz, attr); + CTR4(KTR_PMAP, "%s(pa=%#lx, sz=%#lx, attr=%#x)", __func__, pa, sz, + attr); va = pmap_mapdev_priv(pa, sz, attr); return ((void *)(uintptr_t)va); @@ -2553,7 +2558,7 @@ void pmap_unmapdev(vm_offset_t va, vm_size_t size) { - CTR3(KTR_PMAP, "%s(%#x, %#x)", __func__, va, size); + CTR3(KTR_PMAP, "%s(va=%#lx, sz=%#lx)", __func__, va, size); } /* @@ -2579,7 +2584,7 @@ pmap_page_set_memattr(vm_page_t m, vm_me pv_entry_t pv; void *va; - CTR3(KTR_PMAP, "%s(%p, %#x)", __func__, m, ma); + CTR3(KTR_PMAP, "%s(m=%p, attr=%#x)", __func__, m, ma); rw_wlock(&pvh_global_lock); m->md.memattr = ma; @@ -2627,7 +2632,8 @@ pmap_mincore(pmap_t pmap, vm_offset_t ad vm_paddr_t pa; int val; - CTR4(KTR_PMAP, "%s(%p, %#x, %p)", __func__, pmap, addr, locked_pa); + CTR4(KTR_PMAP, "%s(pm=%p, va=%#lx, pa_p=%p)", __func__, pmap, addr, + locked_pa); PMAP_LOCK(pmap); retry: @@ -2668,7 +2674,7 @@ void pmap_activate(struct thread *td) { - CTR2(KTR_PMAP, "%s(%p)", __func__, td); + CTR2(KTR_PMAP, "%s(td=%p)", __func__, td); pmap_switch(vmspace_pmap(td->td_proc->p_vmspace)); } @@ -2713,7 +2719,7 @@ pmap_sync_icache(pmap_t pm, vm_offset_t vm_offset_t lim; vm_size_t len; - CTR4(KTR_PMAP, "%s(%p, %#x, %#x)", __func__, pm, va, sz); + CTR4(KTR_PMAP, "%s(pm=%p, va=%#lx, sz=%#lx)", __func__, pm, va, sz); sz += va & 31; va &= ~31; @@ -2743,8 +2749,8 @@ pmap_align_superpage(vm_object_t object, vm_offset_t *addr, vm_size_t size) { - CTR5(KTR_PMAP, "%s(%p, %#x, %p, %#x)", __func__, object, offset, addr, - size); + CTR5(KTR_PMAP, "%s(obj=%p, ofs=%#lx, va_p=%p, sz=%#lx)", __func__, + object, offset, addr, size); } #include "opt_ddb.h"