1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
|
#ifndef _ASM_IA64_XENSYSTEM_H
#define _ASM_IA64_XENSYSTEM_H
/*
* xen specific context definition
*
* Copyright (C) 2005 Hewlett-Packard Co.
* Dan Magenheimer (dan.magenheimer@hp.com)
*
* Copyright (C) 2005 Intel Co.
* Kun Tian (Kevin Tian) <kevin.tian@intel.com>
*
*/
#include <asm/config.h>
#include <linux/kernel.h>
/* Define HV space hierarchy */
#define XEN_VIRT_SPACE_LOW 0xe800000000000000
#define XEN_VIRT_SPACE_HIGH 0xf800000000000000
/* This is address to mapping rr7 switch stub, in region 5 */
#ifdef CONFIG_VTI
#define XEN_RR7_SWITCH_STUB 0xb700000000000000
#endif // CONFIG_VTI
#define KERNEL_START 0xf000000004000000
#define PERCPU_ADDR 0xf100000000000000-PERCPU_PAGE_SIZE
#define SHAREDINFO_ADDR 0xf100000000000000
#define VHPT_ADDR 0xf200000000000000
#ifndef __ASSEMBLY__
#define IA64_HAS_EXTRA_STATE(t) 0
#ifdef CONFIG_VTI
extern struct task_struct *vmx_ia64_switch_to (void *next_task);
#define __switch_to(prev,next,last) do { \
if (VMX_DOMAIN(prev)) \
vmx_save_state(prev); \
else { \
if (IA64_HAS_EXTRA_STATE(prev)) \
ia64_save_extra(prev); \
} \
if (VMX_DOMAIN(next)) \
vmx_load_state(next); \
else { \
if (IA64_HAS_EXTRA_STATE(next)) \
ia64_save_extra(next); \
} \
ia64_psr(ia64_task_regs(next))->dfh = !ia64_is_local_fpu_owner(next); \
(last) = vmx_ia64_switch_to((next)); \
} while (0)
#else // CONFIG_VTI
#define __switch_to(prev,next,last) do { \
ia64_save_fpu(prev->arch._thread.fph); \
ia64_load_fpu(next->arch._thread.fph); \
if (IA64_HAS_EXTRA_STATE(prev)) \
ia64_save_extra(prev); \
if (IA64_HAS_EXTRA_STATE(next)) \
ia64_load_extra(next); \
ia64_psr(ia64_task_regs(next))->dfh = !ia64_is_local_fpu_owner(next); \
(last) = ia64_switch_to((next)); \
} while (0)
#endif // CONFIG_VTI
#define __cmpxchg_user(ptr, new, old, _size) \
({ \
register long __gu_r8 asm ("r8"); \
register long __gu_r9 asm ("r9"); \
asm volatile ("mov ar.ccv=%0;;" :: "rO"(old)); \
asm volatile ("mov %2=r0;;\n" \
"[1:]\tcmpxchg"_size".acq %0=[%3],%4,ar.ccv\n" \
"\t.xdata4 \"__ex_table\", 1b-., 1f-.+4\n" \
"[1:]" \
: "=r"(old), "=r"(__gu_r9), "=r"(__gu_r8) : \
"r"(ptr), "r"(new) : "memory"); \
(old) = __gu_r9; \
__gu_r8; \
})
// NOTE: Xen defines args as pointer,old,new whereas ia64 uses pointer,new,old
// so reverse them here
#define cmpxchg_user(_p,_o,_n) \
({ \
register long _rc; \
ia64_mf(); \
switch ( sizeof(*(_p)) ) { \
case 1: _rc = __cmpxchg_user(_p,_n,_o,"1"); break; \
case 2: _rc = __cmpxchg_user(_p,_n,_o,"2"); break; \
case 4: _rc = __cmpxchg_user(_p,_n,_o,"4"); break; \
case 8: _rc = __cmpxchg_user(_p,_n,_o,"8"); break; \
} \
ia64_mf(); \
_rc; \
})
#endif // __ASSEMBLY__
#endif // _ASM_IA64_XENSYSTEM_H
|