From 880dbede4c64eccd6507f1fdcc1cb3337d4563c6 Mon Sep 17 00:00:00 2001 From: "iap10@freefall.cl.cam.ac.uk" Date: Sun, 14 Aug 2005 16:55:34 +0000 Subject: Now we have verbose=y in Xen by default, tidy up a few unhelpful printk's. Signed-off-by: ian@xensource.com --- xen/arch/x86/mm.c | 6 +++--- xen/arch/x86/shadow.c | 4 ++-- xen/arch/x86/shadow32.c | 4 ++-- xen/include/asm-x86/shadow.h | 10 +++++----- 4 files changed, 12 insertions(+), 12 deletions(-) diff --git a/xen/arch/x86/mm.c b/xen/arch/x86/mm.c index 9b52c80e64..db4d9e63b1 100644 --- a/xen/arch/x86/mm.c +++ b/xen/arch/x86/mm.c @@ -2841,7 +2841,7 @@ long do_update_descriptor(unsigned long pa, u64 desc) * Writable Pagetables */ -#ifdef VERBOSE +#ifdef VVERBOSE int ptwr_debug = 0x0; #define PTWR_PRINTK(_f, _a...) \ do { if ( unlikely(ptwr_debug) ) printk( _f , ## _a ); } while ( 0 ) @@ -3054,7 +3054,7 @@ void ptwr_flush(struct domain *d, const int which) */ BUG(); } - PTWR_PRINTK("[%c] disconnected_l1va at %p is %lx\n", + PTWR_PRINTK("[%c] disconnected_l1va at %p is %" PRIpte "\n", PTWR_PRINT_WHICH, ptep, pte.l1); l1e_remove_flags(pte, _PAGE_RW); @@ -3072,7 +3072,7 @@ void ptwr_flush(struct domain *d, const int which) /* Ensure that there are no stale writable mappings in any TLB. */ /* NB. INVLPG is a serialising instruction: flushes pending updates. */ flush_tlb_one_mask(d->cpumask, l1va); - PTWR_PRINTK("[%c] disconnected_l1va at %p now %lx\n", + PTWR_PRINTK("[%c] disconnected_l1va at %p now %" PRIpte "\n", PTWR_PRINT_WHICH, ptep, pte.l1); /* diff --git a/xen/arch/x86/shadow.c b/xen/arch/x86/shadow.c index e4ee2a75a3..e6dd5fe0c1 100644 --- a/xen/arch/x86/shadow.c +++ b/xen/arch/x86/shadow.c @@ -1578,7 +1578,7 @@ static inline int l1pte_write_fault( if ( unlikely(!VALID_MFN(gmfn)) ) { - SH_LOG("l1pte_write_fault: invalid gpfn=%lx", gpfn); + SH_VLOG("l1pte_write_fault: invalid gpfn=%lx", gpfn); *spte_p = l1e_empty(); return 0; } @@ -1612,7 +1612,7 @@ static inline int l1pte_read_fault( if ( unlikely(!VALID_MFN(mfn)) ) { - SH_LOG("l1pte_read_fault: invalid gpfn=%lx", pfn); + SH_VLOG("l1pte_read_fault: invalid gpfn=%lx", pfn); *spte_p = l1e_empty(); return 0; } diff --git a/xen/arch/x86/shadow32.c b/xen/arch/x86/shadow32.c index de7656e674..8f7414a8a7 100644 --- a/xen/arch/x86/shadow32.c +++ b/xen/arch/x86/shadow32.c @@ -665,7 +665,7 @@ static void free_shadow_pages(struct domain *d) shadow_audit(d, 0); - SH_LOG("Free shadow table."); + SH_VLOG("Free shadow table."); } void shadow_mode_init(void) @@ -1137,7 +1137,7 @@ static void free_shadow_ht_entries(struct domain *d) d->arch.shadow_ht_free = NULL; ASSERT(d->arch.shadow_extras_count == 0); - SH_LOG("freed extras, now %d", d->arch.shadow_extras_count); + SH_VLOG("freed extras, now %d", d->arch.shadow_extras_count); if ( d->arch.shadow_dirty_bitmap != NULL ) { diff --git a/xen/include/asm-x86/shadow.h b/xen/include/asm-x86/shadow.h index 955d323d05..fe0bf1c198 100644 --- a/xen/include/asm-x86/shadow.h +++ b/xen/include/asm-x86/shadow.h @@ -483,9 +483,9 @@ static inline int __mark_dirty(struct domain *d, unsigned int mfn) #ifndef NDEBUG else if ( mfn < max_page ) { - SH_LOG("mark_dirty OOR! mfn=%x pfn=%lx max=%x (dom %p)", + SH_VLOG("mark_dirty OOR! mfn=%x pfn=%lx max=%x (dom %p)", mfn, pfn, d->arch.shadow_dirty_bitmap_size, d); - SH_LOG("dom=%p caf=%08x taf=%" PRtype_info, + SH_VLOG("dom=%p caf=%08x taf=%" PRtype_info, page_get_owner(&frame_table[mfn]), frame_table[mfn].count_info, frame_table[mfn].u.inuse.type_info ); @@ -736,7 +736,7 @@ static inline int l1pte_write_fault( if ( unlikely(!VALID_MFN(gmfn)) ) { - SH_LOG("l1pte_write_fault: invalid gpfn=%lx", gpfn); + SH_VLOG("l1pte_write_fault: invalid gpfn=%lx", gpfn); *spte_p = l1e_empty(); return 0; } @@ -770,7 +770,7 @@ static inline int l1pte_read_fault( if ( unlikely(!VALID_MFN(mfn)) ) { - SH_LOG("l1pte_read_fault: invalid gpfn=%lx", pfn); + SH_VLOG("l1pte_read_fault: invalid gpfn=%lx", pfn); *spte_p = l1e_empty(); return 0; } @@ -1472,7 +1472,7 @@ static inline void set_shadow_status( /* We need to allocate a new node. Ensure the quicklist is non-empty. */ if ( unlikely(d->arch.shadow_ht_free == NULL) ) { - SH_LOG("Allocate more shadow hashtable blocks."); + SH_VLOG("Allocate more shadow hashtable blocks."); extra = xmalloc_bytes( sizeof(void *) + (shadow_ht_extra_size * sizeof(*x))); -- cgit v1.2.3