diff options
author | Keir Fraser <keir@xen.org> | 2012-01-15 22:02:35 +0000 |
---|---|---|
committer | Keir Fraser <keir@xen.org> | 2012-01-15 22:02:35 +0000 |
commit | a2c9c2f9f08b1cf12cb56644417bde6373095e80 (patch) | |
tree | 5764cb261c784a684503c5c65b1e1a95d4a69dae /xen/arch | |
parent | a54239be9af2a50e9e849e3f7da8193256ad39bc (diff) | |
download | xen-a2c9c2f9f08b1cf12cb56644417bde6373095e80.tar.gz xen-a2c9c2f9f08b1cf12cb56644417bde6373095e80.tar.bz2 xen-a2c9c2f9f08b1cf12cb56644417bde6373095e80.zip |
x86: Make asmlinkage explicitly a no-op, and avoid usage in arch/x86
Signed-off-by: Keir Fraser <keir@xen.org>
Diffstat (limited to 'xen/arch')
-rw-r--r-- | xen/arch/x86/acpi/power.c | 2 | ||||
-rw-r--r-- | xen/arch/x86/hvm/svm/asid.c | 2 | ||||
-rw-r--r-- | xen/arch/x86/hvm/svm/intr.c | 2 | ||||
-rw-r--r-- | xen/arch/x86/hvm/svm/nestedsvm.c | 2 | ||||
-rw-r--r-- | xen/arch/x86/hvm/svm/svm.c | 4 | ||||
-rw-r--r-- | xen/arch/x86/hvm/vmx/intr.c | 2 | ||||
-rw-r--r-- | xen/arch/x86/hvm/vmx/vmx.c | 6 | ||||
-rw-r--r-- | xen/arch/x86/hvm/vmx/vvmx.c | 2 | ||||
-rw-r--r-- | xen/arch/x86/irq.c | 2 | ||||
-rw-r--r-- | xen/arch/x86/trace.c | 2 | ||||
-rw-r--r-- | xen/arch/x86/traps.c | 30 | ||||
-rw-r--r-- | xen/arch/x86/x86_32/traps.c | 2 | ||||
-rw-r--r-- | xen/arch/x86/x86_64/traps.c | 4 |
13 files changed, 31 insertions, 31 deletions
diff --git a/xen/arch/x86/acpi/power.c b/xen/arch/x86/acpi/power.c index 7bcc381fe4..4b242043c8 100644 --- a/xen/arch/x86/acpi/power.c +++ b/xen/arch/x86/acpi/power.c @@ -311,7 +311,7 @@ static void tboot_sleep(u8 sleep_state) } /* System is really put into sleep state by this stub */ -acpi_status asmlinkage acpi_enter_sleep_state(u8 sleep_state) +acpi_status acpi_enter_sleep_state(u8 sleep_state) { acpi_status status; diff --git a/xen/arch/x86/hvm/svm/asid.c b/xen/arch/x86/hvm/svm/asid.c index ede2be6cc5..14e2abd919 100644 --- a/xen/arch/x86/hvm/svm/asid.c +++ b/xen/arch/x86/hvm/svm/asid.c @@ -39,7 +39,7 @@ void svm_asid_init(struct cpuinfo_x86 *c) * Called directly before VMRUN. Checks if the VCPU needs a new ASID, * assigns it, and if required, issues required TLB flushes. */ -asmlinkage void svm_asid_handle_vmrun(void) +void svm_asid_handle_vmrun(void) { struct vcpu *curr = current; struct vmcb_struct *vmcb = curr->arch.hvm_svm.vmcb; diff --git a/xen/arch/x86/hvm/svm/intr.c b/xen/arch/x86/hvm/svm/intr.c index 973373d7ab..78f6908f82 100644 --- a/xen/arch/x86/hvm/svm/intr.c +++ b/xen/arch/x86/hvm/svm/intr.c @@ -132,7 +132,7 @@ static void enable_intr_window(struct vcpu *v, struct hvm_intack intack) vmcb, general1_intercepts | GENERAL1_INTERCEPT_VINTR); } -asmlinkage void svm_intr_assist(void) +void svm_intr_assist(void) { struct vcpu *v = current; struct vmcb_struct *vmcb = v->arch.hvm_svm.vmcb; diff --git a/xen/arch/x86/hvm/svm/nestedsvm.c b/xen/arch/x86/hvm/svm/nestedsvm.c index 40a68bf898..8714bb0d44 100644 --- a/xen/arch/x86/hvm/svm/nestedsvm.c +++ b/xen/arch/x86/hvm/svm/nestedsvm.c @@ -1404,7 +1404,7 @@ nestedsvm_vcpu_vmexit(struct vcpu *v, struct cpu_user_regs *regs, } /* VCPU switch */ -asmlinkage void nsvm_vcpu_switch(struct cpu_user_regs *regs) +void nsvm_vcpu_switch(struct cpu_user_regs *regs) { struct vcpu *v = current; struct nestedvcpu *nv; diff --git a/xen/arch/x86/hvm/svm/svm.c b/xen/arch/x86/hvm/svm/svm.c index 6986f4a97f..e4b41c4470 100644 --- a/xen/arch/x86/hvm/svm/svm.c +++ b/xen/arch/x86/hvm/svm/svm.c @@ -1843,7 +1843,7 @@ static struct hvm_function_table __read_mostly svm_function_table = { .nhvm_intr_blocked = nsvm_intr_blocked, }; -asmlinkage void svm_vmexit_handler(struct cpu_user_regs *regs) +void svm_vmexit_handler(struct cpu_user_regs *regs) { uint64_t exit_reason; struct vcpu *v = current; @@ -2249,7 +2249,7 @@ asmlinkage void svm_vmexit_handler(struct cpu_user_regs *regs) vmcb_set_vintr(vmcb, intr); } -asmlinkage void svm_trace_vmentry(void) +void svm_trace_vmentry(void) { struct vcpu *curr = current; HVMTRACE_ND(VMENTRY, diff --git a/xen/arch/x86/hvm/vmx/intr.c b/xen/arch/x86/hvm/vmx/intr.c index b44839163f..d675011042 100644 --- a/xen/arch/x86/hvm/vmx/intr.c +++ b/xen/arch/x86/hvm/vmx/intr.c @@ -200,7 +200,7 @@ static int nvmx_intr_intercept(struct vcpu *v, struct hvm_intack intack) return 0; } -asmlinkage void vmx_intr_assist(void) +void vmx_intr_assist(void) { struct hvm_intack intack; struct vcpu *v = current; diff --git a/xen/arch/x86/hvm/vmx/vmx.c b/xen/arch/x86/hvm/vmx/vmx.c index 414414f143..81e8b82ca5 100644 --- a/xen/arch/x86/hvm/vmx/vmx.c +++ b/xen/arch/x86/hvm/vmx/vmx.c @@ -2169,7 +2169,7 @@ static void vmx_failed_vmentry(unsigned int exit_reason, domain_crash(curr->domain); } -asmlinkage void vmx_enter_realmode(struct cpu_user_regs *regs) +void vmx_enter_realmode(struct cpu_user_regs *regs) { struct vcpu *v = current; @@ -2251,7 +2251,7 @@ static void vmx_idtv_reinject(unsigned long idtv_info) } } -asmlinkage void vmx_vmexit_handler(struct cpu_user_regs *regs) +void vmx_vmexit_handler(struct cpu_user_regs *regs) { unsigned int exit_reason, idtv_info, intr_info = 0, vector = 0; unsigned long exit_qualification, inst_len = 0; @@ -2718,7 +2718,7 @@ out: nvmx_idtv_handling(); } -asmlinkage void vmx_vmenter_helper(void) +void vmx_vmenter_helper(void) { struct vcpu *curr = current; u32 new_asid, old_asid; diff --git a/xen/arch/x86/hvm/vmx/vvmx.c b/xen/arch/x86/hvm/vmx/vvmx.c index 31682edd79..ae5c766d48 100644 --- a/xen/arch/x86/hvm/vmx/vvmx.c +++ b/xen/arch/x86/hvm/vmx/vvmx.c @@ -968,7 +968,7 @@ static void virtual_vmexit(struct cpu_user_regs *regs) vmreturn(regs, VMSUCCEED); } -asmlinkage void nvmx_switch_guest(void) +void nvmx_switch_guest(void) { struct vcpu *v = current; struct nestedvcpu *nvcpu = &vcpu_nestedhvm(v); diff --git a/xen/arch/x86/irq.c b/xen/arch/x86/irq.c index bfdfa2638e..058f89df14 100644 --- a/xen/arch/x86/irq.c +++ b/xen/arch/x86/irq.c @@ -757,7 +757,7 @@ void pirq_set_affinity(struct domain *d, int pirq, const cpumask_t *mask) DEFINE_PER_CPU(unsigned int, irq_count); -asmlinkage void do_IRQ(struct cpu_user_regs *regs) +void do_IRQ(struct cpu_user_regs *regs) { struct irqaction *action; uint32_t tsc_in; diff --git a/xen/arch/x86/trace.c b/xen/arch/x86/trace.c index 3ae3e2d40d..404f293188 100644 --- a/xen/arch/x86/trace.c +++ b/xen/arch/x86/trace.c @@ -11,7 +11,7 @@ #define TRC_64_FLAG 0 #endif -asmlinkage void trace_hypercall(void) +void trace_hypercall(void) { struct cpu_user_regs *regs = guest_cpu_user_regs(); diff --git a/xen/arch/x86/traps.c b/xen/arch/x86/traps.c index 2d585c8c43..b6f7b9ac40 100644 --- a/xen/arch/x86/traps.c +++ b/xen/arch/x86/traps.c @@ -425,7 +425,7 @@ static char *trapstr(int trapnr) * are disabled). In such situations we can't do much that is safe. We try to * print out some tracing and then we just spin. */ -asmlinkage void fatal_trap(int trapnr, struct cpu_user_regs *regs) +void fatal_trap(int trapnr, struct cpu_user_regs *regs) { static DEFINE_PER_CPU(char, depth); @@ -539,7 +539,7 @@ static unsigned int check_guest_io_breakpoint(struct vcpu *v, * Called from asm to set up the MCE trapbounce info. * Returns 0 if no callback is set up, else 1. */ -asmlinkage int set_guest_machinecheck_trapbounce(void) +int set_guest_machinecheck_trapbounce(void) { struct vcpu *v = current; struct trap_bounce *tb = &v->arch.pv_vcpu.trap_bounce; @@ -553,7 +553,7 @@ asmlinkage int set_guest_machinecheck_trapbounce(void) * Called from asm to set up the NMI trapbounce info. * Returns 0 if no callback is set up, else 1. */ -asmlinkage int set_guest_nmi_trapbounce(void) +int set_guest_nmi_trapbounce(void) { struct vcpu *v = current; struct trap_bounce *tb = &v->arch.pv_vcpu.trap_bounce; @@ -601,13 +601,13 @@ static inline void do_trap( } #define DO_ERROR_NOCODE(trapnr, name) \ -asmlinkage void do_##name(struct cpu_user_regs *regs) \ +void do_##name(struct cpu_user_regs *regs) \ { \ do_trap(trapnr, regs, 0); \ } #define DO_ERROR(trapnr, name) \ -asmlinkage void do_##name(struct cpu_user_regs *regs) \ +void do_##name(struct cpu_user_regs *regs) \ { \ do_trap(trapnr, regs, 1); \ } @@ -977,7 +977,7 @@ static int emulate_forced_invalid_op(struct cpu_user_regs *regs) return EXCRET_fault_fixed; } -asmlinkage void do_invalid_op(struct cpu_user_regs *regs) +void do_invalid_op(struct cpu_user_regs *regs) { struct bug_frame bug; struct bug_frame_str bug_str; @@ -1071,7 +1071,7 @@ asmlinkage void do_invalid_op(struct cpu_user_regs *regs) panic("FATAL TRAP: vector = %d (invalid opcode)\n", TRAP_invalid_op); } -asmlinkage void do_int3(struct cpu_user_regs *regs) +void do_int3(struct cpu_user_regs *regs) { DEBUGGER_trap_entry(TRAP_int3, regs); @@ -1084,7 +1084,7 @@ asmlinkage void do_int3(struct cpu_user_regs *regs) do_guest_trap(TRAP_int3, regs, 0); } -asmlinkage void do_machine_check(struct cpu_user_regs *regs) +void do_machine_check(struct cpu_user_regs *regs) { machine_check_vector(regs, regs->error_code); } @@ -1385,7 +1385,7 @@ static int fixup_page_fault(unsigned long addr, struct cpu_user_regs *regs) * Bit 3: Reserved bit violation * Bit 4: Instruction fetch */ -asmlinkage void do_page_fault(struct cpu_user_regs *regs) +void do_page_fault(struct cpu_user_regs *regs) { unsigned long addr, fixup; unsigned int error_code; @@ -1451,7 +1451,7 @@ asmlinkage void do_page_fault(struct cpu_user_regs *regs) * during early boot (an issue was seen once, but was most likely a hardware * problem). */ -asmlinkage void __init do_early_page_fault(struct cpu_user_regs *regs) +void __init do_early_page_fault(struct cpu_user_regs *regs) { static int stuck; static unsigned long prev_eip, prev_cr2; @@ -3049,7 +3049,7 @@ static void emulate_gate_op(struct cpu_user_regs *regs) #endif } -asmlinkage void do_general_protection(struct cpu_user_regs *regs) +void do_general_protection(struct cpu_user_regs *regs) { struct vcpu *v = current; unsigned long fixup; @@ -3307,7 +3307,7 @@ static int dummy_nmi_callback(struct cpu_user_regs *regs, int cpu) static nmi_callback_t nmi_callback = dummy_nmi_callback; -asmlinkage void do_nmi(struct cpu_user_regs *regs) +void do_nmi(struct cpu_user_regs *regs) { unsigned int cpu = smp_processor_id(); unsigned char reason; @@ -3343,7 +3343,7 @@ void unset_nmi_callback(void) nmi_callback = dummy_nmi_callback; } -asmlinkage void do_device_not_available(struct cpu_user_regs *regs) +void do_device_not_available(struct cpu_user_regs *regs) { struct vcpu *curr = current; @@ -3384,7 +3384,7 @@ static void ler_enable(void) wrmsrl(MSR_IA32_DEBUGCTLMSR, debugctl | 1); } -asmlinkage void do_debug(struct cpu_user_regs *regs) +void do_debug(struct cpu_user_regs *regs) { struct vcpu *v = current; @@ -3435,7 +3435,7 @@ asmlinkage void do_debug(struct cpu_user_regs *regs) return; } -asmlinkage void do_spurious_interrupt_bug(struct cpu_user_regs *regs) +void do_spurious_interrupt_bug(struct cpu_user_regs *regs) { } diff --git a/xen/arch/x86/x86_32/traps.c b/xen/arch/x86/x86_32/traps.c index ac88aefd97..8f68808da8 100644 --- a/xen/arch/x86/x86_32/traps.c +++ b/xen/arch/x86/x86_32/traps.c @@ -231,7 +231,7 @@ static struct notifier_block cpu_doublefault_tss_nfb = { .notifier_call = cpu_doublefault_tss_callback }; -asmlinkage void do_double_fault(void) +void do_double_fault(void) { struct tss_struct *tss; unsigned int cpu; diff --git a/xen/arch/x86/x86_64/traps.c b/xen/arch/x86/x86_64/traps.c index 38ce30300c..b6f1e61caf 100644 --- a/xen/arch/x86/x86_64/traps.c +++ b/xen/arch/x86/x86_64/traps.c @@ -216,8 +216,8 @@ void show_page_walk(unsigned long addr) l1_table_offset(addr), l1e_get_intpte(l1e), pfn); } -asmlinkage void double_fault(void); -asmlinkage void do_double_fault(struct cpu_user_regs *regs) +void double_fault(void); +void do_double_fault(struct cpu_user_regs *regs) { unsigned int cpu; |