diff options
author | Keir Fraser <keir@xensource.com> | 2007-10-22 14:19:22 +0100 |
---|---|---|
committer | Keir Fraser <keir@xensource.com> | 2007-10-22 14:19:22 +0100 |
commit | fedeb68b583b488631f4e5b41ca81688a0b9d30e (patch) | |
tree | 3e1f65da7357014f214dd0befedcae0d1cbdfc17 /xen/arch/x86/hvm/mtrr.c | |
parent | a2308fa704a40f23916a176d9e06bbc0e3469caf (diff) | |
download | xen-fedeb68b583b488631f4e5b41ca81688a0b9d30e.tar.gz xen-fedeb68b583b488631f4e5b41ca81688a0b9d30e.tar.bz2 xen-fedeb68b583b488631f4e5b41ca81688a0b9d30e.zip |
hvm/x86: MTRR/PAT virtualisation.
Signed-off-by: Disheng Su <disheng.su@intel.com>
Diffstat (limited to 'xen/arch/x86/hvm/mtrr.c')
-rw-r--r-- | xen/arch/x86/hvm/mtrr.c | 687 |
1 files changed, 687 insertions, 0 deletions
diff --git a/xen/arch/x86/hvm/mtrr.c b/xen/arch/x86/hvm/mtrr.c new file mode 100644 index 0000000000..016652723d --- /dev/null +++ b/xen/arch/x86/hvm/mtrr.c @@ -0,0 +1,687 @@ +/* + * mtrr.c: MTRR/PAT virtualization + * + * Copyright (c) 2007, Intel Corporation. + * + * This program is free software; you can redistribute it and/or modify it + * under the terms and conditions of the GNU General Public License, + * version 2, as published by the Free Software Foundation. + * + * This program is distributed in the hope it will be useful, but WITHOUT + * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for + * more details. + * + * You should have received a copy of the GNU General Public License along with + * this program; if not, write to the Free Software Foundation, Inc., 59 Temple + * Place - Suite 330, Boston, MA 02111-1307 USA. + */ + +#include <public/hvm/e820.h> +#include <xen/types.h> +#include <asm/e820.h> +#include <asm/paging.h> +#include <asm/p2m.h> +#include <xen/domain_page.h> +#include <stdbool.h> +#include <asm/mtrr.h> +#include <asm/hvm/support.h> + +/* Xen holds the native MTRR MSRs */ +extern struct mtrr_state mtrr_state; + +static u64 phys_base_msr_mask; +static u64 phys_mask_msr_mask; +static u32 size_or_mask; +static u32 size_and_mask; + +static void init_pat_entry_tbl(u64 pat); +static void init_mtrr_epat_tbl(void); +static unsigned char get_mtrr_type(struct mtrr_state *m, paddr_t pa); +/* get page attribute fields (PAn) from PAT MSR */ +#define pat_cr_2_paf(pat_cr,n) ((((u64)pat_cr) >> ((n)<<3)) & 0xff) +/* pat entry to PTE flags (PAT, PCD, PWT bits) */ +static unsigned char pat_entry_2_pte_flags[8] = { + 0, _PAGE_PWT, + _PAGE_PCD, _PAGE_PCD | _PAGE_PWT, + _PAGE_PAT, _PAGE_PAT | _PAGE_PWT, + _PAGE_PAT | _PAGE_PCD, _PAGE_PAT | _PAGE_PCD | _PAGE_PWT }; + +/* effective mm type lookup table, according to MTRR and PAT */ +static u8 mm_type_tbl[MTRR_NUM_TYPES][PAT_TYPE_NUMS] = { +/********PAT(UC,WC,RS,RS,WT,WP,WB,UC-)*/ +/* RS means reserved type(2,3), and type is hardcoded here */ + /*MTRR(UC):(UC,WC,RS,RS,UC,UC,UC,UC)*/ + {0, 1, 2, 2, 0, 0, 0, 0}, + /*MTRR(WC):(UC,WC,RS,RS,UC,UC,WC,WC)*/ + {0, 1, 2, 2, 0, 0, 1, 1}, + /*MTRR(RS):(RS,RS,RS,RS,RS,RS,RS,RS)*/ + {2, 2, 2, 2, 2, 2, 2, 2}, + /*MTRR(RS):(RS,RS,RS,RS,RS,RS,RS,RS)*/ + {2, 2, 2, 2, 2, 2, 2, 2}, + /*MTRR(WT):(UC,WC,RS,RS,WT,WP,WT,UC)*/ + {0, 1, 2, 2, 4, 5, 4, 0}, + /*MTRR(WP):(UC,WC,RS,RS,WT,WP,WP,WC)*/ + {0, 1, 2, 2, 4, 5, 5, 1}, + /*MTRR(WB):(UC,WC,RS,RS,WT,WP,WB,UC)*/ + {0, 1, 2, 2, 4, 5, 6, 0} +}; + +/* reverse lookup table, to find a pat type according to MTRR and effective + * memory type. This table is dynamically generated + */ +static u8 mtrr_epat_tbl[MTRR_NUM_TYPES][MEMORY_NUM_TYPES]; + +/* lookup table for PAT entry of a given PAT value in host pat */ +static u8 pat_entry_tbl[PAT_TYPE_NUMS]; + +static void get_mtrr_range(uint64_t base_msr, uint64_t mask_msr, + uint64_t *base, uint64_t *end) +{ + uint32_t mask_lo = (uint32_t)mask_msr; + uint32_t mask_hi = (uint32_t)(mask_msr >> 32); + uint32_t base_lo = (uint32_t)base_msr; + uint32_t base_hi = (uint32_t)(base_msr >> 32); + uint32_t size; + + if ( (mask_lo & 0x800) == 0 ) + { + /* Invalid (i.e. free) range */ + *base = 0; + *end = 0; + return; + } + + /* Work out the shifted address mask. */ + mask_lo = (size_or_mask | (mask_hi << (32 - PAGE_SHIFT)) | + (mask_lo >> PAGE_SHIFT)); + + /* This works correctly if size is a power of two (a contiguous range). */ + size = -mask_lo; + *base = base_hi << (32 - PAGE_SHIFT) | base_lo >> PAGE_SHIFT; + *end = *base + size - 1; +} + +bool_t is_var_mtrr_overlapped(struct mtrr_state *m) +{ + int seg, i; + uint64_t phys_base, phys_mask, phys_base_pre, phys_mask_pre; + uint64_t base_pre, end_pre, base, end; + uint8_t num_var_ranges = (u8)m->mtrr_cap; + + for ( i = 0; i < num_var_ranges; i++ ) + { + phys_base_pre = ((u64*)m->var_ranges)[i*2]; + phys_mask_pre = ((u64*)m->var_ranges)[i*2 + 1]; + + get_mtrr_range(phys_base_pre, phys_mask_pre, + &base_pre, &end_pre); + + for ( seg = i + 1; seg < num_var_ranges; seg ++ ) + { + phys_base = ((u64*)m->var_ranges)[seg*2]; + phys_mask = ((u64*)m->var_ranges)[seg*2 + 1]; + + get_mtrr_range(phys_base, phys_mask, + &base, &end); + + if ( ((base_pre != end_pre) && (base != end)) + || ((base >= base_pre) && (base <= end_pre)) + || ((end >= base_pre) && (end <= end_pre)) + || ((base_pre >= base) && (base_pre <= end)) + || ((end_pre >= base) && (end_pre <= end)) ) + { + /* MTRR is overlapped. */ + return 1; + } + } + } + return 0; +} + +/* reserved mtrr for guest OS */ +#define RESERVED_MTRR 2 +#define MTRRphysBase_MSR(reg) (0x200 + 2 * (reg)) +#define MTRRphysMask_MSR(reg) (0x200 + 2 * (reg) + 1) +bool mtrr_var_range_msr_set(struct mtrr_state *m, u32 msr, u64 msr_content); +bool mtrr_def_type_msr_set(struct mtrr_state *m, u64 msr_content); +bool mtrr_fix_range_msr_set(struct mtrr_state *m, int row, u64 msr_content); +static void set_var_mtrr(unsigned int reg, struct mtrr_state *m, + unsigned int base, unsigned int size, + unsigned int type) +{ + struct mtrr_var_range *vr; + + vr = &m->var_ranges[reg]; + + if ( size == 0 ) + { + /* The invalid bit is kept in the mask, so we simply clear the + * relevant mask register to disable a range. + */ + mtrr_var_range_msr_set(m, MTRRphysMask_MSR(reg), 0); + } + else + { + vr->base_lo = base << PAGE_SHIFT | type; + vr->base_hi = (base & size_and_mask) >> (32 - PAGE_SHIFT); + vr->mask_lo = -size << PAGE_SHIFT | 0x800; + vr->mask_hi = (-size & size_and_mask) >> (32 - PAGE_SHIFT); + + mtrr_var_range_msr_set(m, MTRRphysBase_MSR(reg), *(unsigned long *)vr); + mtrr_var_range_msr_set(m, MTRRphysMask_MSR(reg), + *((unsigned long *)vr + 1)); + } +} +/* From Intel Vol. III Section 10.11.4, the Range Size and Base Alignment has + * some kind of requirement: + * 1. The range size must be 2^N byte for N >= 12 (i.e 4KB minimum). + * 2. The base address must be 2^N aligned, where the N here is equal to + * the N in previous requirement. So a 8K range must be 8K aligned not 4K aligned. + */ +static unsigned int range_to_mtrr(unsigned int reg, struct mtrr_state *m, + unsigned int range_startk, unsigned int range_sizek, unsigned char type) +{ + if ( !range_sizek || (reg >= ((m->mtrr_cap & 0xff) - RESERVED_MTRR)) ) + return reg; + + while ( range_sizek ) + { + unsigned int max_align, align, sizek; + + max_align = (range_startk == 0) ? 32 : ffs(range_startk); + align = min_t(unsigned int, fls(range_sizek), max_align); + sizek = 1 << (align - 1); + + set_var_mtrr(reg++, m, range_startk, sizek, type); + + range_startk += sizek; + range_sizek -= sizek; + + if ( reg >= ((m->mtrr_cap & 0xff) - RESERVED_MTRR) ) + break; + } + + return reg; +} + +static void setup_fixed_mtrrs(struct vcpu *v) +{ + uint64_t content; + int i; + struct mtrr_state *m = &v->arch.hvm_vcpu.mtrr; + + /* 1. Map (0~A0000) as WB */ + content = 0x0606060606060606ull; + mtrr_fix_range_msr_set(m, 0, content); + mtrr_fix_range_msr_set(m, 1, content); + /* 2. Map VRAM(A0000~C0000) as WC */ + content = 0x0101010101010101; + mtrr_fix_range_msr_set(m, 2, content); + /* 3. Map (C0000~100000) as UC */ + for ( i = 3; i < 11; i++) + mtrr_fix_range_msr_set(m, i, 0); +} + +static void setup_var_mtrrs(struct vcpu *v) +{ + p2m_type_t p2m; + unsigned long e820_mfn; + char *p = NULL; + unsigned char nr = 0; + int i; + unsigned int reg = 0; + unsigned long size = 0; + unsigned long addr = 0; + struct e820entry *e820_table; + + e820_mfn = mfn_x(gfn_to_mfn(v->domain, + HVM_E820_PAGE >> PAGE_SHIFT, &p2m)); + + p = (char *)map_domain_page(e820_mfn); + + nr = *(unsigned char*)(p + HVM_E820_NR_OFFSET); + e820_table = (struct e820entry*)(p + HVM_E820_OFFSET); + /* search E820 table, set MTRR for RAM */ + for ( i = 0; i < nr; i++) + { + if ( (e820_table[i].addr >= 0x100000) && + (e820_table[i].type == E820_RAM) ) + { + if ( e820_table[i].addr == 0x100000 ) + { + size = e820_table[i].size + 0x100000 + PAGE_SIZE * 3; + addr = 0; + } + else + { + /* Larger than 4G */ + size = e820_table[i].size; + addr = e820_table[i].addr; + } + + reg = range_to_mtrr(reg, &v->arch.hvm_vcpu.mtrr, + addr >> PAGE_SHIFT, size >> PAGE_SHIFT, + MTRR_TYPE_WRBACK); + } + } +} + +void init_mtrr_in_hyper(struct vcpu *v) +{ + /* TODO:MTRR should be initialized in BIOS or other places. + * workaround to do it in here + */ + if ( v->arch.hvm_vcpu.mtrr.is_initialized ) + return; + + setup_fixed_mtrrs(v); + setup_var_mtrrs(v); + /* enable mtrr */ + mtrr_def_type_msr_set(&v->arch.hvm_vcpu.mtrr, 0xc00); + + v->arch.hvm_vcpu.mtrr.is_initialized = 1; +} + +static int reset_mtrr(struct mtrr_state *m) +{ + m->var_ranges = xmalloc_array(struct mtrr_var_range, MTRR_VCNT); + if ( m->var_ranges == NULL ) + return -ENOMEM; + memset(m->var_ranges, 0, MTRR_VCNT * sizeof(struct mtrr_var_range)); + memset(m->fixed_ranges, 0, sizeof(m->fixed_ranges)); + m->enabled = 0; + m->def_type = 0;/*mtrr is disabled*/ + m->mtrr_cap = (0x5<<8)|MTRR_VCNT;/*wc,fix enabled, and vcnt=8*/ + m->overlapped = 0; + return 0; +} + +/* init global variables for MTRR and PAT */ +void global_init_mtrr_pat(void) +{ + extern u64 host_pat; + u32 phys_addr; + + init_mtrr_epat_tbl(); + init_pat_entry_tbl(host_pat); + /* Get max physical address, set some global variable */ + if ( cpuid_eax(0x80000000) < 0x80000008 ) + phys_addr = 36; + else + phys_addr = cpuid_eax(0x80000008); + + phys_base_msr_mask = ~((((u64)1) << phys_addr) - 1) | 0xf00UL; + phys_mask_msr_mask = ~((((u64)1) << phys_addr) - 1) | 0x7ffUL; + + size_or_mask = ~((1 << (phys_addr - PAGE_SHIFT)) - 1); + size_and_mask = ~size_or_mask & 0xfff00000; +} + +static void init_pat_entry_tbl(u64 pat) +{ + int i, j; + + memset(&pat_entry_tbl, INVALID_MEM_TYPE, + PAT_TYPE_NUMS * sizeof(pat_entry_tbl[0])); + + for ( i = 0; i < PAT_TYPE_NUMS; i++ ) + { + for ( j = 0; j < PAT_TYPE_NUMS; j++ ) + { + if ( pat_cr_2_paf(pat, j) == i ) + { + pat_entry_tbl[i] = j; + break; + } + } + } +} + +unsigned char pat_type_2_pte_flags(unsigned char pat_type) +{ + int pat_entry = pat_entry_tbl[pat_type]; + + /* INVALID_MEM_TYPE, means doesn't find the pat_entry in host pat for + * a given pat_type. If host pat covers all the pat types, + * it can't happen. + */ + if ( likely(pat_entry != INVALID_MEM_TYPE) ) + return pat_entry_2_pte_flags[pat_entry]; + + return pat_entry_2_pte_flags[pat_entry_tbl[PAT_TYPE_UNCACHABLE]]; +} + +int reset_vmsr(struct mtrr_state *m, u64 *pat_ptr) +{ + int rc; + + rc = reset_mtrr(m); + if ( rc != 0 ) + return rc; + + *pat_ptr = ( (u64)PAT_TYPE_WRBACK) | /* PAT0: WB */ + ( (u64)PAT_TYPE_WRTHROUGH << 8 ) | /* PAT1: WT */ + ( (u64)PAT_TYPE_UC_MINUS << 16 ) | /* PAT2: UC- */ + ( (u64)PAT_TYPE_UNCACHABLE << 24 ) | /* PAT3: UC */ + ( (u64)PAT_TYPE_WRBACK << 32 ) | /* PAT4: WB */ + ( (u64)PAT_TYPE_WRTHROUGH << 40 ) | /* PAT5: WT */ + ( (u64)PAT_TYPE_UC_MINUS << 48 ) | /* PAT6: UC- */ + ( (u64)PAT_TYPE_UNCACHABLE << 56 ); /* PAT7: UC */ + + return 0; +} + +/* + * Get MTRR memory type for physical address pa. + */ +static unsigned char get_mtrr_type(struct mtrr_state *m, paddr_t pa) +{ + int addr, seg, index; + u8 overlap_mtrr = 0; + u8 overlap_mtrr_pos = 0; + u64 phys_base; + u64 phys_mask; + u8 num_var_ranges = m->mtrr_cap & 0xff; + + if ( unlikely(!(m->enabled & 0x2)) ) + return MTRR_TYPE_UNCACHABLE; + + if ( (pa < 0x100000) && (m->enabled & 1) ) + { + /* Fixed range MTRR takes effective */ + addr = (unsigned int) pa; + if ( addr < 0x80000 ) + { + seg = (addr >> 16); + return m->fixed_ranges[seg]; + } + else if ( addr < 0xc0000 ) + { + seg = (addr - 0x80000) >> 14; + index = (seg >> 3) + 1; + seg &= 7; /* select 0-7 segments */ + return m->fixed_ranges[index*8 + seg]; + } + else + { + /* 0xC0000 --- 0x100000 */ + seg = (addr - 0xc0000) >> 12; + index = (seg >> 3) + 3; + seg &= 7; /* select 0-7 segments */ + return m->fixed_ranges[index*8 + seg]; + } + } + + /* Match with variable MTRRs. */ + for ( seg = 0; seg < num_var_ranges; seg++ ) + { + phys_base = ((u64*)m->var_ranges)[seg*2]; + phys_mask = ((u64*)m->var_ranges)[seg*2 + 1]; + if ( phys_mask & (1 << MTRR_PHYSMASK_VALID_BIT) ) + { + if ( ((u64) pa & phys_mask) >> MTRR_PHYSMASK_SHIFT == + (phys_base & phys_mask) >> MTRR_PHYSMASK_SHIFT ) + { + if ( unlikely(m->overlapped) ) + { + overlap_mtrr |= 1 << (phys_base & MTRR_PHYSBASE_TYPE_MASK); + overlap_mtrr_pos = phys_base & MTRR_PHYSBASE_TYPE_MASK; + } + else + { + /* If no overlap, return the found one */ + return (phys_base & MTRR_PHYSBASE_TYPE_MASK); + } + } + } + } + + /* Overlapped or not found. */ + if ( unlikely(overlap_mtrr == 0) ) + return m->def_type; + + if ( likely(!(overlap_mtrr & ~( ((u8)1) << overlap_mtrr_pos ))) ) + /* Covers both one variable memory range matches and + * two or more identical match. + */ + return overlap_mtrr_pos; + + if ( overlap_mtrr & 0x1 ) + /* Two or more match, one is UC. */ + return MTRR_TYPE_UNCACHABLE; + + if ( !(overlap_mtrr & 0xaf) ) + /* Two or more match, WT and WB. */ + return MTRR_TYPE_WRTHROUGH; + + /* Behaviour is undefined, but return the last overlapped type. */ + return overlap_mtrr_pos; +} + +/* + * return the memory type from PAT. + * NOTE: valid only when paging is enabled. + * Only 4K page PTE is supported now. + */ +static unsigned char page_pat_type(u64 pat_cr, unsigned long pte_flags) +{ + int pat_entry; + + /* PCD/PWT -> bit 1/0 of PAT entry */ + pat_entry = ( pte_flags >> 3 ) & 0x3; + /* PAT bits as bit 2 of PAT entry */ + if ( pte_flags & _PAGE_PAT ) + pat_entry |= 4; + + return (unsigned char)pat_cr_2_paf(pat_cr, pat_entry); +} + +/* + * Effective memory type for leaf page. + */ +static u8 effective_mm_type( + struct mtrr_state *m, + u64 pat, + paddr_t gpa, + unsigned long pte_flags) +{ + unsigned char mtrr_mtype, pat_value, effective; + + mtrr_mtype = get_mtrr_type(m, gpa); + + pat_value = page_pat_type(pat, pte_flags); + + effective = mm_type_tbl[mtrr_mtype][pat_value]; + + return effective; +} + +static void init_mtrr_epat_tbl(void) +{ + int i, j; + /* set default value to an invalid type, just for checking conflict */ + memset(&mtrr_epat_tbl, INVALID_MEM_TYPE, sizeof(mtrr_epat_tbl)); + + for ( i = 0; i < MTRR_NUM_TYPES; i++ ) + { + for ( j = 0; j < PAT_TYPE_NUMS; j++ ) + { + int tmp = mm_type_tbl[i][j]; + if ( (tmp >= 0) && (tmp < MEMORY_NUM_TYPES) ) + mtrr_epat_tbl[i][tmp] = j; + } + } +} + +u32 get_pat_flags(struct vcpu *v, + u32 gl1e_flags, + paddr_t gpaddr, + paddr_t spaddr) +{ + u8 guest_eff_mm_type; + u8 shadow_mtrr_type; + u8 pat_entry_value; + u64 pat = v->arch.hvm_vcpu.pat_cr; + struct mtrr_state *g = &v->arch.hvm_vcpu.mtrr; + + /* 1. Get the effective memory type of guest physical address, + * with the pair of guest MTRR and PAT + */ + guest_eff_mm_type = effective_mm_type(g, pat, gpaddr, gl1e_flags); + /* 2. Get the memory type of host physical address, with MTRR */ + shadow_mtrr_type = get_mtrr_type(&mtrr_state, spaddr); + + /* 3. Find the memory type in PAT, with host MTRR memory type + * and guest effective memory type. + */ + pat_entry_value = mtrr_epat_tbl[shadow_mtrr_type][guest_eff_mm_type]; + /* If conflit occurs(e.g host MTRR is UC, guest memory type is + * WB),set UC as effective memory. Here, returning PAT_TYPE_UNCACHABLE will + * always set effective memory as UC. + */ + if ( pat_entry_value == INVALID_MEM_TYPE ) + { + gdprintk(XENLOG_WARNING, + "Conflict occurs for a given guest l1e flags:%x " + "at %"PRIx64" (the effective mm type:%d), " + "because the host mtrr type is:%d\n", + gl1e_flags, (uint64_t)gpaddr, guest_eff_mm_type, + shadow_mtrr_type); + pat_entry_value = PAT_TYPE_UNCACHABLE; + } + /* 4. Get the pte flags */ + return pat_type_2_pte_flags(pat_entry_value); +} + +/* Helper funtions for seting mtrr/pat */ +bool pat_msr_set(u64 *pat, u64 msr_content) +{ + u8 *value = (u8*)&msr_content; + int i; + + if ( *pat != msr_content ) + { + for ( i = 0; i < 8; i++ ) + if ( unlikely(!(value[i] == 0 || value[i] == 1 || + value[i] == 4 || value[i] == 5 || + value[i] == 6 || value[i] == 7)) ) + return 0; + + *pat = msr_content; + } + + return 1; +} + +bool mtrr_def_type_msr_set(struct mtrr_state *m, u64 msr_content) +{ + u8 def_type = msr_content & 0xff; + u8 enabled = (msr_content >> 10) & 0x3; + + if ( unlikely(!(def_type == 0 || def_type == 1 || def_type == 4 || + def_type == 5 || def_type == 6)) ) + { + HVM_DBG_LOG(DBG_LEVEL_MSR, "invalid MTRR def type:%x\n", def_type); + return 0; + } + + if ( unlikely(msr_content && (msr_content & ~0xcffUL)) ) + { + HVM_DBG_LOG(DBG_LEVEL_MSR, "invalid msr content:%"PRIx64"\n", + msr_content); + return 0; + } + + m->enabled = enabled; + m->def_type = def_type; + + return 1; +} + +bool mtrr_fix_range_msr_set(struct mtrr_state *m, int row, u64 msr_content) +{ + u64 *fixed_range_base = (u64 *)m->fixed_ranges; + + if ( fixed_range_base[row] != msr_content ) + { + u8 *range = (u8*)&msr_content; + int i, type; + + for ( i = 0; i < 8; i++ ) + { + type = range[i]; + if ( unlikely(!(type == 0 || type == 1 || + type == 4 || type == 5 || type == 6)) ) + return 0; + } + + fixed_range_base[row] = msr_content; + } + + return 1; +} + +bool mtrr_var_range_msr_set(struct mtrr_state *m, u32 msr, u64 msr_content) +{ + u32 index; + u64 msr_mask; + u64 *var_range_base = (u64*)m->var_ranges; + + index = msr - MSR_IA32_MTRR_PHYSBASE0; + + if ( var_range_base[index] != msr_content ) + { + u32 type = msr_content & 0xff; + + msr_mask = (index & 1) ? phys_mask_msr_mask : phys_base_msr_mask; + + if ( unlikely(!(type == 0 || type == 1 || + type == 4 || type == 5 || type == 6)) ) + return 0; + + if ( unlikely(msr_content && (msr_content & msr_mask)) ) + { + HVM_DBG_LOG(DBG_LEVEL_MSR, "invalid msr content:%"PRIx64"\n", + msr_content); + return 0; + } + + var_range_base[index] = msr_content; + } + + m->overlapped = is_var_mtrr_overlapped(m); + + return 1; +} + +bool_t mtrr_pat_not_equal(struct vcpu *vd, struct vcpu *vs) +{ + struct mtrr_state *md = &vd->arch.hvm_vcpu.mtrr; + struct mtrr_state *ms = &vs->arch.hvm_vcpu.mtrr; + int res; + u8 num_var_ranges = (u8)md->mtrr_cap; + + /* Test fixed ranges. */ + res = memcmp(md->fixed_ranges, ms->fixed_ranges, + NUM_FIXED_RANGES*sizeof(mtrr_type)); + if ( res ) + return 1; + + /* Test var ranges. */ + res = memcmp(md->var_ranges, ms->var_ranges, + num_var_ranges*sizeof(struct mtrr_var_range)); + if ( res ) + return 1; + + /* Test default type MSR. */ + if ( (md->def_type != ms->def_type) + && (md->enabled != ms->enabled) ) + return 1; + + /* Test PAT. */ + if ( vd->arch.hvm_vcpu.pat_cr != vs->arch.hvm_vcpu.pat_cr ) + return 1; + + return 0; +} |