From d61d49705fc0b11a6f406c1954a7a25be0677d18 Mon Sep 17 00:00:00 2001 From: "vh249@arcadians.cl.cam.ac.uk" Date: Mon, 15 Aug 2005 12:26:20 +0000 Subject: Small change to remove difference between Xen and mainline deactivate_mm. While there, cleanup prepare_arch_switch as well. This generates identical code. Signed-off-by: Chris Wright --- linux-2.6-xen-sparse/include/asm-xen/asm-i386/mmu_context.h | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/linux-2.6-xen-sparse/include/asm-xen/asm-i386/mmu_context.h b/linux-2.6-xen-sparse/include/asm-xen/asm-i386/mmu_context.h index f54bb7bdd7..477eccf6f5 100644 --- a/linux-2.6-xen-sparse/include/asm-xen/asm-i386/mmu_context.h +++ b/linux-2.6-xen-sparse/include/asm-xen/asm-i386/mmu_context.h @@ -34,10 +34,10 @@ static inline void __prepare_arch_switch(void) * are always kernel segments while inside the kernel. Must * happen before reload of cr3/ldt (i.e., not in __switch_to). */ - __asm__ __volatile__ ( "mov %%fs,%0 ; mov %%gs,%1" + asm volatile ( "mov %%fs,%0 ; mov %%gs,%1" : "=m" (*(int *)¤t->thread.fs), "=m" (*(int *)¤t->thread.gs)); - __asm__ __volatile__ ( "mov %0,%%fs ; mov %0,%%gs" + asm volatile ( "mov %0,%%fs ; mov %0,%%gs" : : "r" (0) ); } @@ -100,7 +100,7 @@ static inline void switch_mm(struct mm_struct *prev, } #define deactivate_mm(tsk, mm) \ - asm("mov %0,%%fs ; mov %0,%%gs": :"r" (0)) + asm("movl %0,%%fs ; movl %0,%%gs": :"r" (0)) #define activate_mm(prev, next) \ switch_mm((prev),(next),NULL) -- cgit v1.2.3