| #include <linux/errno.h> |
| #include <linux/signal.h> |
| #include <linux/sched.h> |
| #include <linux/ioport.h> |
| #include <linux/interrupt.h> |
| #include <linux/slab.h> |
| #include <linux/random.h> |
| #include <linux/init.h> |
| #include <linux/kernel_stat.h> |
| #include <linux/sysdev.h> |
| #include <linux/bitops.h> |
| #include <linux/io.h> |
| #include <linux/delay.h> |
| |
| #include <asm/atomic.h> |
| #include <asm/system.h> |
| #include <asm/timer.h> |
| #include <asm/pgtable.h> |
| #include <asm/desc.h> |
| #include <asm/apic.h> |
| #include <asm/arch_hooks.h> |
| #include <asm/i8259.h> |
| #include <asm/traps.h> |
| |
| |
| /* |
| * Note that on a 486, we don't want to do a SIGFPE on an irq13 |
| * as the irq is unreliable, and exception 16 works correctly |
| * (ie as explained in the intel literature). On a 386, you |
| * can't use exception 16 due to bad IBM design, so we have to |
| * rely on the less exact irq13. |
| * |
| * Careful.. Not only is IRQ13 unreliable, but it is also |
| * leads to races. IBM designers who came up with it should |
| * be shot. |
| */ |
| |
| static irqreturn_t math_error_irq(int cpl, void *dev_id) |
| { |
| outb(0, 0xF0); |
| if (ignore_fpu_irq || !boot_cpu_data.hard_math) |
| return IRQ_NONE; |
| math_error((void __user *)get_irq_regs()->ip); |
| return IRQ_HANDLED; |
| } |
| |
| /* |
| * New motherboards sometimes make IRQ 13 be a PCI interrupt, |
| * so allow interrupt sharing. |
| */ |
| static struct irqaction fpu_irq = { |
| .handler = math_error_irq, |
| .mask = CPU_MASK_NONE, |
| .name = "fpu", |
| }; |
| |
| void __init init_ISA_irqs(void) |
| { |
| int i; |
| |
| #ifdef CONFIG_X86_LOCAL_APIC |
| init_bsp_APIC(); |
| #endif |
| init_8259A(0); |
| |
| /* |
| * 16 old-style INTA-cycle interrupts: |
| */ |
| for (i = 0; i < NR_IRQS_LEGACY; i++) { |
| struct irq_desc *desc = irq_to_desc(i); |
| |
| desc->status = IRQ_DISABLED; |
| desc->action = NULL; |
| desc->depth = 1; |
| |
| set_irq_chip_and_handler_name(i, &i8259A_chip, |
| handle_level_irq, "XT"); |
| } |
| } |
| |
| /* |
| * IRQ2 is cascade interrupt to second interrupt controller |
| */ |
| static struct irqaction irq2 = { |
| .handler = no_action, |
| .mask = CPU_MASK_NONE, |
| .name = "cascade", |
| }; |
| |
| DEFINE_PER_CPU(vector_irq_t, vector_irq) = { |
| [0 ... IRQ0_VECTOR - 1] = -1, |
| [IRQ0_VECTOR] = 0, |
| [IRQ1_VECTOR] = 1, |
| [IRQ2_VECTOR] = 2, |
| [IRQ3_VECTOR] = 3, |
| [IRQ4_VECTOR] = 4, |
| [IRQ5_VECTOR] = 5, |
| [IRQ6_VECTOR] = 6, |
| [IRQ7_VECTOR] = 7, |
| [IRQ8_VECTOR] = 8, |
| [IRQ9_VECTOR] = 9, |
| [IRQ10_VECTOR] = 10, |
| [IRQ11_VECTOR] = 11, |
| [IRQ12_VECTOR] = 12, |
| [IRQ13_VECTOR] = 13, |
| [IRQ14_VECTOR] = 14, |
| [IRQ15_VECTOR] = 15, |
| [IRQ15_VECTOR + 1 ... NR_VECTORS - 1] = -1 |
| }; |
| |
| int vector_used_by_percpu_irq(unsigned int vector) |
| { |
| int cpu; |
| |
| for_each_online_cpu(cpu) { |
| if (per_cpu(vector_irq, cpu)[vector] != -1) |
| return 1; |
| } |
| |
| return 0; |
| } |
| |
| /* Overridden in paravirt.c */ |
| void init_IRQ(void) __attribute__((weak, alias("native_init_IRQ"))); |
| |
| void __init native_init_IRQ(void) |
| { |
| int i; |
| |
| /* all the set up before the call gates are initialised */ |
| pre_intr_init_hook(); |
| |
| /* |
| * Cover the whole vector space, no vector can escape |
| * us. (some of these will be overridden and become |
| * 'special' SMP interrupts) |
| */ |
| for (i = FIRST_EXTERNAL_VECTOR; i < NR_VECTORS; i++) { |
| /* SYSCALL_VECTOR was reserved in trap_init. */ |
| if (i != SYSCALL_VECTOR) |
| set_intr_gate(i, interrupt[i-FIRST_EXTERNAL_VECTOR]); |
| } |
| |
| |
| #if defined(CONFIG_X86_LOCAL_APIC) && defined(CONFIG_SMP) |
| /* |
| * The reschedule interrupt is a CPU-to-CPU reschedule-helper |
| * IPI, driven by wakeup. |
| */ |
| alloc_intr_gate(RESCHEDULE_VECTOR, reschedule_interrupt); |
| |
| /* IPIs for invalidation */ |
| alloc_intr_gate(INVALIDATE_TLB_VECTOR_START+0, invalidate_interrupt0); |
| alloc_intr_gate(INVALIDATE_TLB_VECTOR_START+1, invalidate_interrupt1); |
| alloc_intr_gate(INVALIDATE_TLB_VECTOR_START+2, invalidate_interrupt2); |
| alloc_intr_gate(INVALIDATE_TLB_VECTOR_START+3, invalidate_interrupt3); |
| alloc_intr_gate(INVALIDATE_TLB_VECTOR_START+4, invalidate_interrupt4); |
| alloc_intr_gate(INVALIDATE_TLB_VECTOR_START+5, invalidate_interrupt5); |
| alloc_intr_gate(INVALIDATE_TLB_VECTOR_START+6, invalidate_interrupt6); |
| alloc_intr_gate(INVALIDATE_TLB_VECTOR_START+7, invalidate_interrupt7); |
| |
| /* IPI for generic function call */ |
| alloc_intr_gate(CALL_FUNCTION_VECTOR, call_function_interrupt); |
| |
| /* IPI for single call function */ |
| alloc_intr_gate(CALL_FUNCTION_SINGLE_VECTOR, |
| call_function_single_interrupt); |
| |
| /* Low priority IPI to cleanup after moving an irq */ |
| set_intr_gate(IRQ_MOVE_CLEANUP_VECTOR, irq_move_cleanup_interrupt); |
| set_bit(IRQ_MOVE_CLEANUP_VECTOR, used_vectors); |
| #endif |
| |
| #ifdef CONFIG_X86_LOCAL_APIC |
| /* self generated IPI for local APIC timer */ |
| alloc_intr_gate(LOCAL_TIMER_VECTOR, apic_timer_interrupt); |
| |
| /* IPI vectors for APIC spurious and error interrupts */ |
| alloc_intr_gate(SPURIOUS_APIC_VECTOR, spurious_interrupt); |
| alloc_intr_gate(ERROR_APIC_VECTOR, error_interrupt); |
| #endif |
| |
| #if defined(CONFIG_X86_LOCAL_APIC) && defined(CONFIG_X86_MCE_P4THERMAL) |
| /* thermal monitor LVT interrupt */ |
| alloc_intr_gate(THERMAL_APIC_VECTOR, thermal_interrupt); |
| #endif |
| |
| if (!acpi_ioapic) |
| setup_irq(2, &irq2); |
| |
| /* setup after call gates are initialised (usually add in |
| * the architecture specific gates) |
| */ |
| intr_init_hook(); |
| |
| /* |
| * External FPU? Set up irq13 if so, for |
| * original braindamaged IBM FERR coupling. |
| */ |
| if (boot_cpu_data.hard_math && !cpu_has_fpu) |
| setup_irq(FPU_IRQ, &fpu_irq); |
| |
| irq_ctx_init(smp_processor_id()); |
| } |