@@ -1322,7 +1322,7 @@ int reprogram_timer(s_time_t timeout)
return apic_tmict || !timeout;
}
-static void cf_check apic_timer_interrupt(struct cpu_user_regs *regs)
+static void cf_check apic_timer_interrupt(void)
{
ack_APIC_irq();
perfc_incr(apic_timer);
@@ -1341,7 +1341,7 @@ void smp_send_state_dump(unsigned int cp
/*
* Spurious interrupts should _never_ happen with our APIC/SMP architecture.
*/
-static void cf_check spurious_interrupt(struct cpu_user_regs *regs)
+static void cf_check spurious_interrupt(void)
{
/*
* Check if this is a vectored interrupt (most likely, as this is probably
@@ -1355,7 +1355,7 @@ static void cf_check spurious_interrupt(
is_spurious = !nmi_check_continuation();
if (this_cpu(state_dump_pending)) {
this_cpu(state_dump_pending) = false;
- dump_execstate(regs);
+ dump_execstate(get_irq_regs());
is_spurious = false;
}
@@ -1372,7 +1372,7 @@ static void cf_check spurious_interrupt(
* This interrupt should never happen with our APIC/SMP architecture
*/
-static void cf_check error_interrupt(struct cpu_user_regs *regs)
+static void cf_check error_interrupt(void)
{
static const char *const esr_fields[] = {
", Send CS error",
@@ -1407,7 +1407,7 @@ static void cf_check error_interrupt(str
* This interrupt handles performance counters interrupt
*/
-static void cf_check pmu_interrupt(struct cpu_user_regs *regs)
+static void cf_check pmu_interrupt(void)
{
ack_APIC_irq();
vpmu_do_interrupt();
@@ -58,7 +58,7 @@ bool __read_mostly lmce_support;
#define MCE_RING 0x1
static DEFINE_PER_CPU(int, last_state);
-static void cf_check intel_thermal_interrupt(struct cpu_user_regs *regs)
+static void cf_check intel_thermal_interrupt(void)
{
uint64_t msr_content;
unsigned int cpu = smp_processor_id();
@@ -642,7 +642,7 @@ static void cpu_mcheck_disable(void)
clear_cmci();
}
-static void cf_check cmci_interrupt(struct cpu_user_regs *regs)
+static void cf_check cmci_interrupt(void)
{
mctelem_cookie_t mctc;
struct mca_summary bs;
@@ -158,7 +158,7 @@ static void __init init_memmap(void)
}
}
-static void cf_check xen_evtchn_upcall(struct cpu_user_regs *regs)
+static void cf_check xen_evtchn_upcall(void)
{
struct vcpu_info *vcpu_info = this_cpu(vcpu_info);
unsigned long pending;
@@ -2782,7 +2782,7 @@ static struct hvm_function_table __initd
};
/* Handle VT-d posted-interrupt when VCPU is blocked. */
-static void cf_check pi_wakeup_interrupt(struct cpu_user_regs *regs)
+static void cf_check pi_wakeup_interrupt(void)
{
struct vmx_vcpu *vmx, *tmp;
spinlock_t *lock = &per_cpu(vmx_pi_blocking, smp_processor_id()).lock;
@@ -2814,7 +2814,7 @@ static void cf_check pi_wakeup_interrupt
}
/* Handle VT-d posted-interrupt when VCPU is running. */
-static void cf_check pi_notification_interrupt(struct cpu_user_regs *regs)
+static void cf_check pi_notification_interrupt(void)
{
ack_APIC_irq();
this_cpu(irq_count)++;
@@ -72,17 +72,15 @@ extern int opt_irq_vector_map;
#define platform_legacy_irq(irq) ((irq) < 16)
-void cf_check event_check_interrupt(struct cpu_user_regs *regs);
-void cf_check invalidate_interrupt(struct cpu_user_regs *regs);
-void cf_check call_function_interrupt(struct cpu_user_regs *regs);
-void cf_check irq_move_cleanup_interrupt(struct cpu_user_regs *regs);
+void cf_check event_check_interrupt(void);
+void cf_check invalidate_interrupt(void);
+void cf_check call_function_interrupt(void);
+void cf_check irq_move_cleanup_interrupt(void);
uint8_t alloc_hipriority_vector(void);
-void set_direct_apic_vector(
- uint8_t vector, void (*handler)(struct cpu_user_regs *regs));
-void alloc_direct_apic_vector(
- uint8_t *vector, void (*handler)(struct cpu_user_regs *regs));
+void set_direct_apic_vector(uint8_t vector, void (*handler)(void));
+void alloc_direct_apic_vector(uint8_t *vector, void (*handler)(void));
void do_IRQ(struct cpu_user_regs *regs);
@@ -743,7 +743,7 @@ void move_native_irq(struct irq_desc *de
desc->handler->enable(desc);
}
-void cf_check irq_move_cleanup_interrupt(struct cpu_user_regs *regs)
+void cf_check irq_move_cleanup_interrupt(void)
{
unsigned vector, me;
@@ -913,16 +913,14 @@ uint8_t alloc_hipriority_vector(void)
return next++;
}
-static void (*direct_apic_vector[X86_NR_VECTORS])(struct cpu_user_regs *regs);
-void set_direct_apic_vector(
- uint8_t vector, void (*handler)(struct cpu_user_regs *regs))
+static void (*direct_apic_vector[X86_NR_VECTORS])(void);
+void set_direct_apic_vector(uint8_t vector, void (*handler)(void))
{
BUG_ON(direct_apic_vector[vector] != NULL);
direct_apic_vector[vector] = handler;
}
-void alloc_direct_apic_vector(
- uint8_t *vector, void (*handler)(struct cpu_user_regs *regs))
+void alloc_direct_apic_vector(uint8_t *vector, void (*handler)(void))
{
static DEFINE_SPINLOCK(lock);
@@ -1907,7 +1905,7 @@ void do_IRQ(struct cpu_user_regs *regs)
if ( irq < 0 )
{
if ( direct_apic_vector[vector] )
- direct_apic_vector[vector](regs);
+ direct_apic_vector[vector]();
else
{
const char *kind = ", LAPIC";
@@ -245,7 +245,7 @@ static cpumask_t flush_cpumask;
static const void *flush_va;
static unsigned int flush_flags;
-void cf_check invalidate_interrupt(struct cpu_user_regs *regs)
+void cf_check invalidate_interrupt(void)
{
unsigned int flags = flush_flags;
ack_APIC_irq();
@@ -387,14 +387,14 @@ void smp_send_nmi_allbutself(void)
send_IPI_mask(&cpu_online_map, APIC_DM_NMI);
}
-void cf_check event_check_interrupt(struct cpu_user_regs *regs)
+void cf_check event_check_interrupt(void)
{
ack_APIC_irq();
perfc_incr(ipis);
this_cpu(irq_count)++;
}
-void cf_check call_function_interrupt(struct cpu_user_regs *regs)
+void cf_check call_function_interrupt(void)
{
ack_APIC_irq();
perfc_incr(ipis);