1 #include <linux/linkage.h>
2 #include <linux/errno.h>
3 #include <linux/signal.h>
4 #include <linux/sched.h>
5 #include <linux/ioport.h>
6 #include <linux/interrupt.h>
8 #include <linux/timex.h>
9 #include <linux/random.h>
10 #include <linux/kprobes.h>
11 #include <linux/init.h>
12 #include <linux/kernel_stat.h>
13 #include <linux/device.h>
14 #include <linux/bitops.h>
15 #include <linux/acpi.h>
17 #include <linux/delay.h>
19 #include <linux/atomic.h>
20 #include <asm/timer.h>
21 #include <asm/hw_irq.h>
22 #include <asm/pgtable.h>
25 #include <asm/setup.h>
26 #include <asm/i8259.h>
27 #include <asm/traps.h>
31 * ISA PIC or low IO-APIC triggered (INTA-cycle or APIC) interrupts:
32 * (these are usually mapped to vectors 0x30-0x3f)
36 * The IO-APIC gives us many more interrupt sources. Most of these
37 * are unused but an SMP system is supposed to have enough memory ...
38 * sometimes (mostly wrt. hw bugs) we get corrupted vectors all
39 * across the spectrum, so we really want to be prepared to get all
40 * of these. Plus, more powerful systems might have more than 64
43 * (these are usually mapped into the 0x30-0xff vector range)
47 * IRQ2 is cascade interrupt to second interrupt controller
49 static struct irqaction irq2 = {
52 .flags = IRQF_NO_THREAD,
55 DEFINE_PER_CPU_USER_MAPPED(vector_irq_t, vector_irq) = {
56 [0 ... NR_VECTORS - 1] = VECTOR_UNUSED,
59 int vector_used_by_percpu_irq(unsigned int vector)
63 for_each_online_cpu(cpu) {
64 if (!IS_ERR_OR_NULL(per_cpu(vector_irq, cpu)[vector]))
71 void __init init_ISA_irqs(void)
73 struct irq_chip *chip = legacy_pic->chip;
76 #if defined(CONFIG_X86_64) || defined(CONFIG_X86_LOCAL_APIC)
81 for (i = 0; i < nr_legacy_irqs(); i++)
82 irq_set_chip_and_handler(i, chip, handle_level_irq);
85 void __init init_IRQ(void)
90 * On cpu 0, Assign ISA_IRQ_VECTOR(irq) to IRQ 0..15.
91 * If these IRQ's are handled by legacy interrupt-controllers like PIC,
92 * then this configuration will likely be static after the boot. If
93 * these IRQ's are handled by more mordern controllers like IO-APIC,
94 * then this vector space can be freed and re-used dynamically as the
97 for (i = 0; i < nr_legacy_irqs(); i++)
98 per_cpu(vector_irq, 0)[ISA_IRQ_VECTOR(i)] = irq_to_desc(i);
100 x86_init.irqs.intr_init();
103 static void __init smp_intr_init(void)
107 * The reschedule interrupt is a CPU-to-CPU reschedule-helper
108 * IPI, driven by wakeup.
110 alloc_intr_gate(RESCHEDULE_VECTOR, reschedule_interrupt);
112 /* IPI for generic function call */
113 alloc_intr_gate(CALL_FUNCTION_VECTOR, call_function_interrupt);
115 /* IPI for generic single function call */
116 alloc_intr_gate(CALL_FUNCTION_SINGLE_VECTOR,
117 call_function_single_interrupt);
119 /* Low priority IPI to cleanup after moving an irq */
120 set_intr_gate(IRQ_MOVE_CLEANUP_VECTOR, irq_move_cleanup_interrupt);
121 set_bit(IRQ_MOVE_CLEANUP_VECTOR, used_vectors);
123 /* IPI used for rebooting/stopping */
124 alloc_intr_gate(REBOOT_VECTOR, reboot_interrupt);
125 #endif /* CONFIG_SMP */
128 static void __init apic_intr_init(void)
132 #ifdef CONFIG_X86_THERMAL_VECTOR
133 alloc_intr_gate(THERMAL_APIC_VECTOR, thermal_interrupt);
135 #ifdef CONFIG_X86_MCE_THRESHOLD
136 alloc_intr_gate(THRESHOLD_APIC_VECTOR, threshold_interrupt);
139 #ifdef CONFIG_X86_MCE_AMD
140 alloc_intr_gate(DEFERRED_ERROR_VECTOR, deferred_error_interrupt);
143 #ifdef CONFIG_X86_LOCAL_APIC
144 /* self generated IPI for local APIC timer */
145 alloc_intr_gate(LOCAL_TIMER_VECTOR, apic_timer_interrupt);
147 /* IPI for X86 platform specific use */
148 alloc_intr_gate(X86_PLATFORM_IPI_VECTOR, x86_platform_ipi);
149 #ifdef CONFIG_HAVE_KVM
150 /* IPI for KVM to deliver posted interrupt */
151 alloc_intr_gate(POSTED_INTR_VECTOR, kvm_posted_intr_ipi);
152 /* IPI for KVM to deliver interrupt to wake up tasks */
153 alloc_intr_gate(POSTED_INTR_WAKEUP_VECTOR, kvm_posted_intr_wakeup_ipi);
156 /* IPI vectors for APIC spurious and error interrupts */
157 alloc_intr_gate(SPURIOUS_APIC_VECTOR, spurious_interrupt);
158 alloc_intr_gate(ERROR_APIC_VECTOR, error_interrupt);
160 /* IRQ work interrupts: */
161 # ifdef CONFIG_IRQ_WORK
162 alloc_intr_gate(IRQ_WORK_VECTOR, irq_work_interrupt);
168 void __init native_init_IRQ(void)
172 /* Execute any quirks before the call gates are initialised: */
173 x86_init.irqs.pre_vector_init();
178 * Cover the whole vector space, no vector can escape
179 * us. (some of these will be overridden and become
180 * 'special' SMP interrupts)
182 i = FIRST_EXTERNAL_VECTOR;
183 #ifndef CONFIG_X86_LOCAL_APIC
184 #define first_system_vector NR_VECTORS
186 for_each_clear_bit_from(i, used_vectors, first_system_vector) {
187 /* IA32_SYSCALL_VECTOR could be used in trap_init already. */
188 set_intr_gate(i, irq_entries_start +
189 8 * (i - FIRST_EXTERNAL_VECTOR));
191 #ifdef CONFIG_X86_LOCAL_APIC
192 for_each_clear_bit_from(i, used_vectors, NR_VECTORS)
193 set_intr_gate(i, spurious_interrupt);
196 if (!acpi_ioapic && !of_ioapic && nr_legacy_irqs())
200 irq_ctx_init(smp_processor_id());