• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * I/O SAPIC support.
3  *
4  * Copyright (C) 1999 Intel Corp.
5  * Copyright (C) 1999 Asit Mallick <asit.k.mallick@intel.com>
6  * Copyright (C) 2000-2002 J.I. Lee <jung-ik.lee@intel.com>
7  * Copyright (C) 1999-2000, 2002-2003 Hewlett-Packard Co.
8  *	David Mosberger-Tang <davidm@hpl.hp.com>
9  * Copyright (C) 1999 VA Linux Systems
10  * Copyright (C) 1999,2000 Walt Drummond <drummond@valinux.com>
11  *
12  * 00/04/19	D. Mosberger	Rewritten to mirror more closely the x86 I/O
13  *				APIC code.  In particular, we now have separate
14  *				handlers for edge and level triggered
15  *				interrupts.
16  * 00/10/27	Asit Mallick, Goutham Rao <goutham.rao@intel.com> IRQ vector
17  *				allocation PCI to vector mapping, shared PCI
18  *				interrupts.
19  * 00/10/27	D. Mosberger	Document things a bit more to make them more
20  *				understandable.  Clean up much of the old
21  *				IOSAPIC cruft.
22  * 01/07/27	J.I. Lee	PCI irq routing, Platform/Legacy interrupts
23  *				and fixes for ACPI S5(SoftOff) support.
24  * 02/01/23	J.I. Lee	iosapic pgm fixes for PCI irq routing from _PRT
25  * 02/01/07     E. Focht        <efocht@ess.nec.de> Redirectable interrupt
26  *				vectors in iosapic_set_affinity(),
27  *				initializations for /proc/irq/#/smp_affinity
28  * 02/04/02	P. Diefenbaugh	Cleaned up ACPI PCI IRQ routing.
29  * 02/04/18	J.I. Lee	bug fix in iosapic_init_pci_irq
30  * 02/04/30	J.I. Lee	bug fix in find_iosapic to fix ACPI PCI IRQ to
31  *				IOSAPIC mapping error
32  * 02/07/29	T. Kochi	Allocate interrupt vectors dynamically
33  * 02/08/04	T. Kochi	Cleaned up terminology (irq, global system
34  *				interrupt, vector, etc.)
35  * 02/09/20	D. Mosberger	Simplified by taking advantage of ACPI's
36  *				pci_irq code.
37  * 03/02/19	B. Helgaas	Make pcat_compat system-wide, not per-IOSAPIC.
38  *				Remove iosapic_address & gsi_base from
39  *				external interfaces.  Rationalize
40  *				__init/__devinit attributes.
41  * 04/12/04 Ashok Raj	<ashok.raj@intel.com> Intel Corporation 2004
42  *				Updated to work with irq migration necessary
43  *				for CPU Hotplug
44  */
45 /*
46  * Here is what the interrupt logic between a PCI device and the kernel looks
47  * like:
48  *
49  * (1) A PCI device raises one of the four interrupt pins (INTA, INTB, INTC,
50  *     INTD).  The device is uniquely identified by its bus-, and slot-number
51  *     (the function number does not matter here because all functions share
52  *     the same interrupt lines).
53  *
54  * (2) The motherboard routes the interrupt line to a pin on a IOSAPIC
55  *     controller.  Multiple interrupt lines may have to share the same
56  *     IOSAPIC pin (if they're level triggered and use the same polarity).
57  *     Each interrupt line has a unique Global System Interrupt (GSI) number
58  *     which can be calculated as the sum of the controller's base GSI number
59  *     and the IOSAPIC pin number to which the line connects.
60  *
61  * (3) The IOSAPIC uses an internal routing table entries (RTEs) to map the
62  * IOSAPIC pin into the IA-64 interrupt vector.  This interrupt vector is then
63  * sent to the CPU.
64  *
65  * (4) The kernel recognizes an interrupt as an IRQ.  The IRQ interface is
66  *     used as architecture-independent interrupt handling mechanism in Linux.
67  *     As an IRQ is a number, we have to have
68  *     IA-64 interrupt vector number <-> IRQ number mapping.  On smaller
69  *     systems, we use one-to-one mapping between IA-64 vector and IRQ.  A
70  *     platform can implement platform_irq_to_vector(irq) and
71  *     platform_local_vector_to_irq(vector) APIs to differentiate the mapping.
72  *     Please see also arch/ia64/include/asm/hw_irq.h for those APIs.
73  *
74  * To sum up, there are three levels of mappings involved:
75  *
76  *	PCI pin -> global system interrupt (GSI) -> IA-64 vector <-> IRQ
77  *
78  * Note: The term "IRQ" is loosely used everywhere in Linux kernel to
79  * describeinterrupts.  Now we use "IRQ" only for Linux IRQ's.  ISA IRQ
80  * (isa_irq) is the only exception in this source code.
81  */
82 
83 #include <linux/acpi.h>
84 #include <linux/init.h>
85 #include <linux/irq.h>
86 #include <linux/kernel.h>
87 #include <linux/list.h>
88 #include <linux/pci.h>
89 #include <linux/smp.h>
90 #include <linux/string.h>
91 #include <linux/bootmem.h>
92 
93 #include <asm/delay.h>
94 #include <asm/hw_irq.h>
95 #include <asm/io.h>
96 #include <asm/iosapic.h>
97 #include <asm/machvec.h>
98 #include <asm/processor.h>
99 #include <asm/ptrace.h>
100 #include <asm/system.h>
101 
102 #undef DEBUG_INTERRUPT_ROUTING
103 
104 #ifdef DEBUG_INTERRUPT_ROUTING
105 #define DBG(fmt...)	printk(fmt)
106 #else
107 #define DBG(fmt...)
108 #endif
109 
110 #define NR_PREALLOCATE_RTE_ENTRIES \
111 	(PAGE_SIZE / sizeof(struct iosapic_rte_info))
112 #define RTE_PREALLOCATED	(1)
113 
114 static DEFINE_SPINLOCK(iosapic_lock);
115 
116 /*
117  * These tables map IA-64 vectors to the IOSAPIC pin that generates this
118  * vector.
119  */
120 
121 #define NO_REF_RTE	0
122 
123 static struct iosapic {
124 	char __iomem	*addr;		/* base address of IOSAPIC */
125 	unsigned int	gsi_base;	/* GSI base */
126 	unsigned short	num_rte;	/* # of RTEs on this IOSAPIC */
127 	int		rtes_inuse;	/* # of RTEs in use on this IOSAPIC */
128 #ifdef CONFIG_NUMA
129 	unsigned short	node;		/* numa node association via pxm */
130 #endif
131 	spinlock_t	lock;		/* lock for indirect reg access */
132 } iosapic_lists[NR_IOSAPICS];
133 
134 struct iosapic_rte_info {
135 	struct list_head rte_list;	/* RTEs sharing the same vector */
136 	char		rte_index;	/* IOSAPIC RTE index */
137 	int		refcnt;		/* reference counter */
138 	unsigned int	flags;		/* flags */
139 	struct iosapic	*iosapic;
140 } ____cacheline_aligned;
141 
142 static struct iosapic_intr_info {
143 	struct list_head rtes;		/* RTEs using this vector (empty =>
144 					 * not an IOSAPIC interrupt) */
145 	int		count;		/* # of registered RTEs */
146 	u32		low32;		/* current value of low word of
147 					 * Redirection table entry */
148 	unsigned int	dest;		/* destination CPU physical ID */
149 	unsigned char	dmode	: 3;	/* delivery mode (see iosapic.h) */
150 	unsigned char 	polarity: 1;	/* interrupt polarity
151 					 * (see iosapic.h) */
152 	unsigned char	trigger	: 1;	/* trigger mode (see iosapic.h) */
153 } iosapic_intr_info[NR_IRQS];
154 
155 static unsigned char pcat_compat __devinitdata;	/* 8259 compatibility flag */
156 
157 static int iosapic_kmalloc_ok;
158 static LIST_HEAD(free_rte_list);
159 
160 static inline void
iosapic_write(struct iosapic * iosapic,unsigned int reg,u32 val)161 iosapic_write(struct iosapic *iosapic, unsigned int reg, u32 val)
162 {
163 	unsigned long flags;
164 
165 	spin_lock_irqsave(&iosapic->lock, flags);
166 	__iosapic_write(iosapic->addr, reg, val);
167 	spin_unlock_irqrestore(&iosapic->lock, flags);
168 }
169 
170 /*
171  * Find an IOSAPIC associated with a GSI
172  */
173 static inline int
find_iosapic(unsigned int gsi)174 find_iosapic (unsigned int gsi)
175 {
176 	int i;
177 
178 	for (i = 0; i < NR_IOSAPICS; i++) {
179 		if ((unsigned) (gsi - iosapic_lists[i].gsi_base) <
180 		    iosapic_lists[i].num_rte)
181 			return i;
182 	}
183 
184 	return -1;
185 }
186 
__gsi_to_irq(unsigned int gsi)187 static inline int __gsi_to_irq(unsigned int gsi)
188 {
189 	int irq;
190 	struct iosapic_intr_info *info;
191 	struct iosapic_rte_info *rte;
192 
193 	for (irq = 0; irq < NR_IRQS; irq++) {
194 		info = &iosapic_intr_info[irq];
195 		list_for_each_entry(rte, &info->rtes, rte_list)
196 			if (rte->iosapic->gsi_base + rte->rte_index == gsi)
197 				return irq;
198 	}
199 	return -1;
200 }
201 
202 int
gsi_to_irq(unsigned int gsi)203 gsi_to_irq (unsigned int gsi)
204 {
205 	unsigned long flags;
206 	int irq;
207 
208 	spin_lock_irqsave(&iosapic_lock, flags);
209 	irq = __gsi_to_irq(gsi);
210 	spin_unlock_irqrestore(&iosapic_lock, flags);
211 	return irq;
212 }
213 
find_rte(unsigned int irq,unsigned int gsi)214 static struct iosapic_rte_info *find_rte(unsigned int irq, unsigned int gsi)
215 {
216 	struct iosapic_rte_info *rte;
217 
218 	list_for_each_entry(rte, &iosapic_intr_info[irq].rtes, rte_list)
219 		if (rte->iosapic->gsi_base + rte->rte_index == gsi)
220 			return rte;
221 	return NULL;
222 }
223 
224 static void
set_rte(unsigned int gsi,unsigned int irq,unsigned int dest,int mask)225 set_rte (unsigned int gsi, unsigned int irq, unsigned int dest, int mask)
226 {
227 	unsigned long pol, trigger, dmode;
228 	u32 low32, high32;
229 	int rte_index;
230 	char redir;
231 	struct iosapic_rte_info *rte;
232 	ia64_vector vector = irq_to_vector(irq);
233 
234 	DBG(KERN_DEBUG"IOSAPIC: routing vector %d to 0x%x\n", vector, dest);
235 
236 	rte = find_rte(irq, gsi);
237 	if (!rte)
238 		return;		/* not an IOSAPIC interrupt */
239 
240 	rte_index = rte->rte_index;
241 	pol     = iosapic_intr_info[irq].polarity;
242 	trigger = iosapic_intr_info[irq].trigger;
243 	dmode   = iosapic_intr_info[irq].dmode;
244 
245 	redir = (dmode == IOSAPIC_LOWEST_PRIORITY) ? 1 : 0;
246 
247 #ifdef CONFIG_SMP
248 	set_irq_affinity_info(irq, (int)(dest & 0xffff), redir);
249 #endif
250 
251 	low32 = ((pol << IOSAPIC_POLARITY_SHIFT) |
252 		 (trigger << IOSAPIC_TRIGGER_SHIFT) |
253 		 (dmode << IOSAPIC_DELIVERY_SHIFT) |
254 		 ((mask ? 1 : 0) << IOSAPIC_MASK_SHIFT) |
255 		 vector);
256 
257 	/* dest contains both id and eid */
258 	high32 = (dest << IOSAPIC_DEST_SHIFT);
259 
260 	iosapic_write(rte->iosapic, IOSAPIC_RTE_HIGH(rte_index), high32);
261 	iosapic_write(rte->iosapic, IOSAPIC_RTE_LOW(rte_index), low32);
262 	iosapic_intr_info[irq].low32 = low32;
263 	iosapic_intr_info[irq].dest = dest;
264 }
265 
266 static void
nop(unsigned int irq)267 nop (unsigned int irq)
268 {
269 	/* do nothing... */
270 }
271 
272 
273 #ifdef CONFIG_KEXEC
274 void
kexec_disable_iosapic(void)275 kexec_disable_iosapic(void)
276 {
277 	struct iosapic_intr_info *info;
278 	struct iosapic_rte_info *rte;
279 	ia64_vector vec;
280 	int irq;
281 
282 	for (irq = 0; irq < NR_IRQS; irq++) {
283 		info = &iosapic_intr_info[irq];
284 		vec = irq_to_vector(irq);
285 		list_for_each_entry(rte, &info->rtes,
286 				rte_list) {
287 			iosapic_write(rte->iosapic,
288 					IOSAPIC_RTE_LOW(rte->rte_index),
289 					IOSAPIC_MASK|vec);
290 			iosapic_eoi(rte->iosapic->addr, vec);
291 		}
292 	}
293 }
294 #endif
295 
296 static void
mask_irq(unsigned int irq)297 mask_irq (unsigned int irq)
298 {
299 	u32 low32;
300 	int rte_index;
301 	struct iosapic_rte_info *rte;
302 
303 	if (!iosapic_intr_info[irq].count)
304 		return;			/* not an IOSAPIC interrupt! */
305 
306 	/* set only the mask bit */
307 	low32 = iosapic_intr_info[irq].low32 |= IOSAPIC_MASK;
308 	list_for_each_entry(rte, &iosapic_intr_info[irq].rtes, rte_list) {
309 		rte_index = rte->rte_index;
310 		iosapic_write(rte->iosapic, IOSAPIC_RTE_LOW(rte_index), low32);
311 	}
312 }
313 
314 static void
unmask_irq(unsigned int irq)315 unmask_irq (unsigned int irq)
316 {
317 	u32 low32;
318 	int rte_index;
319 	struct iosapic_rte_info *rte;
320 
321 	if (!iosapic_intr_info[irq].count)
322 		return;			/* not an IOSAPIC interrupt! */
323 
324 	low32 = iosapic_intr_info[irq].low32 &= ~IOSAPIC_MASK;
325 	list_for_each_entry(rte, &iosapic_intr_info[irq].rtes, rte_list) {
326 		rte_index = rte->rte_index;
327 		iosapic_write(rte->iosapic, IOSAPIC_RTE_LOW(rte_index), low32);
328 	}
329 }
330 
331 
332 static void
iosapic_set_affinity(unsigned int irq,const struct cpumask * mask)333 iosapic_set_affinity(unsigned int irq, const struct cpumask *mask)
334 {
335 #ifdef CONFIG_SMP
336 	u32 high32, low32;
337 	int cpu, dest, rte_index;
338 	int redir = (irq & IA64_IRQ_REDIRECTED) ? 1 : 0;
339 	struct iosapic_rte_info *rte;
340 	struct iosapic *iosapic;
341 
342 	irq &= (~IA64_IRQ_REDIRECTED);
343 
344 	cpu = cpumask_first_and(cpu_online_mask, mask);
345 	if (cpu >= nr_cpu_ids)
346 		return;
347 
348 	if (irq_prepare_move(irq, cpu))
349 		return;
350 
351 	dest = cpu_physical_id(cpu);
352 
353 	if (!iosapic_intr_info[irq].count)
354 		return;			/* not an IOSAPIC interrupt */
355 
356 	set_irq_affinity_info(irq, dest, redir);
357 
358 	/* dest contains both id and eid */
359 	high32 = dest << IOSAPIC_DEST_SHIFT;
360 
361 	low32 = iosapic_intr_info[irq].low32 & ~(7 << IOSAPIC_DELIVERY_SHIFT);
362 	if (redir)
363 		/* change delivery mode to lowest priority */
364 		low32 |= (IOSAPIC_LOWEST_PRIORITY << IOSAPIC_DELIVERY_SHIFT);
365 	else
366 		/* change delivery mode to fixed */
367 		low32 |= (IOSAPIC_FIXED << IOSAPIC_DELIVERY_SHIFT);
368 	low32 &= IOSAPIC_VECTOR_MASK;
369 	low32 |= irq_to_vector(irq);
370 
371 	iosapic_intr_info[irq].low32 = low32;
372 	iosapic_intr_info[irq].dest = dest;
373 	list_for_each_entry(rte, &iosapic_intr_info[irq].rtes, rte_list) {
374 		iosapic = rte->iosapic;
375 		rte_index = rte->rte_index;
376 		iosapic_write(iosapic, IOSAPIC_RTE_HIGH(rte_index), high32);
377 		iosapic_write(iosapic, IOSAPIC_RTE_LOW(rte_index), low32);
378 	}
379 #endif
380 }
381 
382 /*
383  * Handlers for level-triggered interrupts.
384  */
385 
386 static unsigned int
iosapic_startup_level_irq(unsigned int irq)387 iosapic_startup_level_irq (unsigned int irq)
388 {
389 	unmask_irq(irq);
390 	return 0;
391 }
392 
393 static void
iosapic_end_level_irq(unsigned int irq)394 iosapic_end_level_irq (unsigned int irq)
395 {
396 	ia64_vector vec = irq_to_vector(irq);
397 	struct iosapic_rte_info *rte;
398 	int do_unmask_irq = 0;
399 
400 	irq_complete_move(irq);
401 	if (unlikely(irq_desc[irq].status & IRQ_MOVE_PENDING)) {
402 		do_unmask_irq = 1;
403 		mask_irq(irq);
404 	}
405 
406 	list_for_each_entry(rte, &iosapic_intr_info[irq].rtes, rte_list)
407 		iosapic_eoi(rte->iosapic->addr, vec);
408 
409 	if (unlikely(do_unmask_irq)) {
410 		move_masked_irq(irq);
411 		unmask_irq(irq);
412 	}
413 }
414 
415 #define iosapic_shutdown_level_irq	mask_irq
416 #define iosapic_enable_level_irq	unmask_irq
417 #define iosapic_disable_level_irq	mask_irq
418 #define iosapic_ack_level_irq		nop
419 
420 static struct irq_chip irq_type_iosapic_level = {
421 	.name =		"IO-SAPIC-level",
422 	.startup =	iosapic_startup_level_irq,
423 	.shutdown =	iosapic_shutdown_level_irq,
424 	.enable =	iosapic_enable_level_irq,
425 	.disable =	iosapic_disable_level_irq,
426 	.ack =		iosapic_ack_level_irq,
427 	.end =		iosapic_end_level_irq,
428 	.mask =		mask_irq,
429 	.unmask =	unmask_irq,
430 	.set_affinity =	iosapic_set_affinity
431 };
432 
433 /*
434  * Handlers for edge-triggered interrupts.
435  */
436 
437 static unsigned int
iosapic_startup_edge_irq(unsigned int irq)438 iosapic_startup_edge_irq (unsigned int irq)
439 {
440 	unmask_irq(irq);
441 	/*
442 	 * IOSAPIC simply drops interrupts pended while the
443 	 * corresponding pin was masked, so we can't know if an
444 	 * interrupt is pending already.  Let's hope not...
445 	 */
446 	return 0;
447 }
448 
449 static void
iosapic_ack_edge_irq(unsigned int irq)450 iosapic_ack_edge_irq (unsigned int irq)
451 {
452 	irq_desc_t *idesc = irq_desc + irq;
453 
454 	irq_complete_move(irq);
455 	move_native_irq(irq);
456 	/*
457 	 * Once we have recorded IRQ_PENDING already, we can mask the
458 	 * interrupt for real. This prevents IRQ storms from unhandled
459 	 * devices.
460 	 */
461 	if ((idesc->status & (IRQ_PENDING|IRQ_DISABLED)) ==
462 	    (IRQ_PENDING|IRQ_DISABLED))
463 		mask_irq(irq);
464 }
465 
466 #define iosapic_enable_edge_irq		unmask_irq
467 #define iosapic_disable_edge_irq	nop
468 #define iosapic_end_edge_irq		nop
469 
470 static struct irq_chip irq_type_iosapic_edge = {
471 	.name =		"IO-SAPIC-edge",
472 	.startup =	iosapic_startup_edge_irq,
473 	.shutdown =	iosapic_disable_edge_irq,
474 	.enable =	iosapic_enable_edge_irq,
475 	.disable =	iosapic_disable_edge_irq,
476 	.ack =		iosapic_ack_edge_irq,
477 	.end =		iosapic_end_edge_irq,
478 	.mask =		mask_irq,
479 	.unmask =	unmask_irq,
480 	.set_affinity =	iosapic_set_affinity
481 };
482 
483 static unsigned int
iosapic_version(char __iomem * addr)484 iosapic_version (char __iomem *addr)
485 {
486 	/*
487 	 * IOSAPIC Version Register return 32 bit structure like:
488 	 * {
489 	 *	unsigned int version   : 8;
490 	 *	unsigned int reserved1 : 8;
491 	 *	unsigned int max_redir : 8;
492 	 *	unsigned int reserved2 : 8;
493 	 * }
494 	 */
495 	return __iosapic_read(addr, IOSAPIC_VERSION);
496 }
497 
iosapic_find_sharable_irq(unsigned long trigger,unsigned long pol)498 static int iosapic_find_sharable_irq(unsigned long trigger, unsigned long pol)
499 {
500 	int i, irq = -ENOSPC, min_count = -1;
501 	struct iosapic_intr_info *info;
502 
503 	/*
504 	 * shared vectors for edge-triggered interrupts are not
505 	 * supported yet
506 	 */
507 	if (trigger == IOSAPIC_EDGE)
508 		return -EINVAL;
509 
510 	for (i = 0; i < NR_IRQS; i++) {
511 		info = &iosapic_intr_info[i];
512 		if (info->trigger == trigger && info->polarity == pol &&
513 		    (info->dmode == IOSAPIC_FIXED ||
514 		     info->dmode == IOSAPIC_LOWEST_PRIORITY) &&
515 		    can_request_irq(i, IRQF_SHARED)) {
516 			if (min_count == -1 || info->count < min_count) {
517 				irq = i;
518 				min_count = info->count;
519 			}
520 		}
521 	}
522 	return irq;
523 }
524 
525 /*
526  * if the given vector is already owned by other,
527  *  assign a new vector for the other and make the vector available
528  */
529 static void __init
iosapic_reassign_vector(int irq)530 iosapic_reassign_vector (int irq)
531 {
532 	int new_irq;
533 
534 	if (iosapic_intr_info[irq].count) {
535 		new_irq = create_irq();
536 		if (new_irq < 0)
537 			panic("%s: out of interrupt vectors!\n", __func__);
538 		printk(KERN_INFO "Reassigning vector %d to %d\n",
539 		       irq_to_vector(irq), irq_to_vector(new_irq));
540 		memcpy(&iosapic_intr_info[new_irq], &iosapic_intr_info[irq],
541 		       sizeof(struct iosapic_intr_info));
542 		INIT_LIST_HEAD(&iosapic_intr_info[new_irq].rtes);
543 		list_move(iosapic_intr_info[irq].rtes.next,
544 			  &iosapic_intr_info[new_irq].rtes);
545 		memset(&iosapic_intr_info[irq], 0,
546 		       sizeof(struct iosapic_intr_info));
547 		iosapic_intr_info[irq].low32 = IOSAPIC_MASK;
548 		INIT_LIST_HEAD(&iosapic_intr_info[irq].rtes);
549 	}
550 }
551 
iosapic_alloc_rte(void)552 static struct iosapic_rte_info * __init_refok iosapic_alloc_rte (void)
553 {
554 	int i;
555 	struct iosapic_rte_info *rte;
556 	int preallocated = 0;
557 
558 	if (!iosapic_kmalloc_ok && list_empty(&free_rte_list)) {
559 		rte = alloc_bootmem(sizeof(struct iosapic_rte_info) *
560 				    NR_PREALLOCATE_RTE_ENTRIES);
561 		for (i = 0; i < NR_PREALLOCATE_RTE_ENTRIES; i++, rte++)
562 			list_add(&rte->rte_list, &free_rte_list);
563 	}
564 
565 	if (!list_empty(&free_rte_list)) {
566 		rte = list_entry(free_rte_list.next, struct iosapic_rte_info,
567 				 rte_list);
568 		list_del(&rte->rte_list);
569 		preallocated++;
570 	} else {
571 		rte = kmalloc(sizeof(struct iosapic_rte_info), GFP_ATOMIC);
572 		if (!rte)
573 			return NULL;
574 	}
575 
576 	memset(rte, 0, sizeof(struct iosapic_rte_info));
577 	if (preallocated)
578 		rte->flags |= RTE_PREALLOCATED;
579 
580 	return rte;
581 }
582 
irq_is_shared(int irq)583 static inline int irq_is_shared (int irq)
584 {
585 	return (iosapic_intr_info[irq].count > 1);
586 }
587 
588 struct irq_chip*
ia64_native_iosapic_get_irq_chip(unsigned long trigger)589 ia64_native_iosapic_get_irq_chip(unsigned long trigger)
590 {
591 	if (trigger == IOSAPIC_EDGE)
592 		return &irq_type_iosapic_edge;
593 	else
594 		return &irq_type_iosapic_level;
595 }
596 
597 static int
register_intr(unsigned int gsi,int irq,unsigned char delivery,unsigned long polarity,unsigned long trigger)598 register_intr (unsigned int gsi, int irq, unsigned char delivery,
599 	       unsigned long polarity, unsigned long trigger)
600 {
601 	irq_desc_t *idesc;
602 	struct hw_interrupt_type *irq_type;
603 	int index;
604 	struct iosapic_rte_info *rte;
605 
606 	index = find_iosapic(gsi);
607 	if (index < 0) {
608 		printk(KERN_WARNING "%s: No IOSAPIC for GSI %u\n",
609 		       __func__, gsi);
610 		return -ENODEV;
611 	}
612 
613 	rte = find_rte(irq, gsi);
614 	if (!rte) {
615 		rte = iosapic_alloc_rte();
616 		if (!rte) {
617 			printk(KERN_WARNING "%s: cannot allocate memory\n",
618 			       __func__);
619 			return -ENOMEM;
620 		}
621 
622 		rte->iosapic	= &iosapic_lists[index];
623 		rte->rte_index	= gsi - rte->iosapic->gsi_base;
624 		rte->refcnt++;
625 		list_add_tail(&rte->rte_list, &iosapic_intr_info[irq].rtes);
626 		iosapic_intr_info[irq].count++;
627 		iosapic_lists[index].rtes_inuse++;
628 	}
629 	else if (rte->refcnt == NO_REF_RTE) {
630 		struct iosapic_intr_info *info = &iosapic_intr_info[irq];
631 		if (info->count > 0 &&
632 		    (info->trigger != trigger || info->polarity != polarity)){
633 			printk (KERN_WARNING
634 				"%s: cannot override the interrupt\n",
635 				__func__);
636 			return -EINVAL;
637 		}
638 		rte->refcnt++;
639 		iosapic_intr_info[irq].count++;
640 		iosapic_lists[index].rtes_inuse++;
641 	}
642 
643 	iosapic_intr_info[irq].polarity = polarity;
644 	iosapic_intr_info[irq].dmode    = delivery;
645 	iosapic_intr_info[irq].trigger  = trigger;
646 
647 	irq_type = iosapic_get_irq_chip(trigger);
648 
649 	idesc = irq_desc + irq;
650 	if (irq_type != NULL && idesc->chip != irq_type) {
651 		if (idesc->chip != &no_irq_type)
652 			printk(KERN_WARNING
653 			       "%s: changing vector %d from %s to %s\n",
654 			       __func__, irq_to_vector(irq),
655 			       idesc->chip->name, irq_type->name);
656 		idesc->chip = irq_type;
657 	}
658 	return 0;
659 }
660 
661 static unsigned int
get_target_cpu(unsigned int gsi,int irq)662 get_target_cpu (unsigned int gsi, int irq)
663 {
664 #ifdef CONFIG_SMP
665 	static int cpu = -1;
666 	extern int cpe_vector;
667 	cpumask_t domain = irq_to_domain(irq);
668 
669 	/*
670 	 * In case of vector shared by multiple RTEs, all RTEs that
671 	 * share the vector need to use the same destination CPU.
672 	 */
673 	if (iosapic_intr_info[irq].count)
674 		return iosapic_intr_info[irq].dest;
675 
676 	/*
677 	 * If the platform supports redirection via XTP, let it
678 	 * distribute interrupts.
679 	 */
680 	if (smp_int_redirect & SMP_IRQ_REDIRECTION)
681 		return cpu_physical_id(smp_processor_id());
682 
683 	/*
684 	 * Some interrupts (ACPI SCI, for instance) are registered
685 	 * before the BSP is marked as online.
686 	 */
687 	if (!cpu_online(smp_processor_id()))
688 		return cpu_physical_id(smp_processor_id());
689 
690 #ifdef CONFIG_ACPI
691 	if (cpe_vector > 0 && irq_to_vector(irq) == IA64_CPEP_VECTOR)
692 		return get_cpei_target_cpu();
693 #endif
694 
695 #ifdef CONFIG_NUMA
696 	{
697 		int num_cpus, cpu_index, iosapic_index, numa_cpu, i = 0;
698 		const struct cpumask *cpu_mask;
699 
700 		iosapic_index = find_iosapic(gsi);
701 		if (iosapic_index < 0 ||
702 		    iosapic_lists[iosapic_index].node == MAX_NUMNODES)
703 			goto skip_numa_setup;
704 
705 		cpu_mask = cpumask_of_node(iosapic_lists[iosapic_index].node);
706 		num_cpus = 0;
707 		for_each_cpu_and(numa_cpu, cpu_mask, &domain) {
708 			if (cpu_online(numa_cpu))
709 				num_cpus++;
710 		}
711 
712 		if (!num_cpus)
713 			goto skip_numa_setup;
714 
715 		/* Use irq assignment to distribute across cpus in node */
716 		cpu_index = irq % num_cpus;
717 
718 		for_each_cpu_and(numa_cpu, cpu_mask, &domain)
719 			if (cpu_online(numa_cpu) && i++ >= cpu_index)
720 				break;
721 
722 		if (numa_cpu < nr_cpu_ids)
723 			return cpu_physical_id(numa_cpu);
724 	}
725 skip_numa_setup:
726 #endif
727 	/*
728 	 * Otherwise, round-robin interrupt vectors across all the
729 	 * processors.  (It'd be nice if we could be smarter in the
730 	 * case of NUMA.)
731 	 */
732 	do {
733 		if (++cpu >= nr_cpu_ids)
734 			cpu = 0;
735 	} while (!cpu_online(cpu) || !cpu_isset(cpu, domain));
736 
737 	return cpu_physical_id(cpu);
738 #else  /* CONFIG_SMP */
739 	return cpu_physical_id(smp_processor_id());
740 #endif
741 }
742 
choose_dmode(void)743 static inline unsigned char choose_dmode(void)
744 {
745 #ifdef CONFIG_SMP
746 	if (smp_int_redirect & SMP_IRQ_REDIRECTION)
747 		return IOSAPIC_LOWEST_PRIORITY;
748 #endif
749 	return IOSAPIC_FIXED;
750 }
751 
752 /*
753  * ACPI can describe IOSAPIC interrupts via static tables and namespace
754  * methods.  This provides an interface to register those interrupts and
755  * program the IOSAPIC RTE.
756  */
757 int
iosapic_register_intr(unsigned int gsi,unsigned long polarity,unsigned long trigger)758 iosapic_register_intr (unsigned int gsi,
759 		       unsigned long polarity, unsigned long trigger)
760 {
761 	int irq, mask = 1, err;
762 	unsigned int dest;
763 	unsigned long flags;
764 	struct iosapic_rte_info *rte;
765 	u32 low32;
766 	unsigned char dmode;
767 
768 	/*
769 	 * If this GSI has already been registered (i.e., it's a
770 	 * shared interrupt, or we lost a race to register it),
771 	 * don't touch the RTE.
772 	 */
773 	spin_lock_irqsave(&iosapic_lock, flags);
774 	irq = __gsi_to_irq(gsi);
775 	if (irq > 0) {
776 		rte = find_rte(irq, gsi);
777 		if(iosapic_intr_info[irq].count == 0) {
778 			assign_irq_vector(irq);
779 			dynamic_irq_init(irq);
780 		} else if (rte->refcnt != NO_REF_RTE) {
781 			rte->refcnt++;
782 			goto unlock_iosapic_lock;
783 		}
784 	} else
785 		irq = create_irq();
786 
787 	/* If vector is running out, we try to find a sharable vector */
788 	if (irq < 0) {
789 		irq = iosapic_find_sharable_irq(trigger, polarity);
790 		if (irq < 0)
791 			goto unlock_iosapic_lock;
792 	}
793 
794 	spin_lock(&irq_desc[irq].lock);
795 	dest = get_target_cpu(gsi, irq);
796 	dmode = choose_dmode();
797 	err = register_intr(gsi, irq, dmode, polarity, trigger);
798 	if (err < 0) {
799 		spin_unlock(&irq_desc[irq].lock);
800 		irq = err;
801 		goto unlock_iosapic_lock;
802 	}
803 
804 	/*
805 	 * If the vector is shared and already unmasked for other
806 	 * interrupt sources, don't mask it.
807 	 */
808 	low32 = iosapic_intr_info[irq].low32;
809 	if (irq_is_shared(irq) && !(low32 & IOSAPIC_MASK))
810 		mask = 0;
811 	set_rte(gsi, irq, dest, mask);
812 
813 	printk(KERN_INFO "GSI %u (%s, %s) -> CPU %d (0x%04x) vector %d\n",
814 	       gsi, (trigger == IOSAPIC_EDGE ? "edge" : "level"),
815 	       (polarity == IOSAPIC_POL_HIGH ? "high" : "low"),
816 	       cpu_logical_id(dest), dest, irq_to_vector(irq));
817 
818 	spin_unlock(&irq_desc[irq].lock);
819  unlock_iosapic_lock:
820 	spin_unlock_irqrestore(&iosapic_lock, flags);
821 	return irq;
822 }
823 
824 void
iosapic_unregister_intr(unsigned int gsi)825 iosapic_unregister_intr (unsigned int gsi)
826 {
827 	unsigned long flags;
828 	int irq, index;
829 	irq_desc_t *idesc;
830 	u32 low32;
831 	unsigned long trigger, polarity;
832 	unsigned int dest;
833 	struct iosapic_rte_info *rte;
834 
835 	/*
836 	 * If the irq associated with the gsi is not found,
837 	 * iosapic_unregister_intr() is unbalanced. We need to check
838 	 * this again after getting locks.
839 	 */
840 	irq = gsi_to_irq(gsi);
841 	if (irq < 0) {
842 		printk(KERN_ERR "iosapic_unregister_intr(%u) unbalanced\n",
843 		       gsi);
844 		WARN_ON(1);
845 		return;
846 	}
847 
848 	spin_lock_irqsave(&iosapic_lock, flags);
849 	if ((rte = find_rte(irq, gsi)) == NULL) {
850 		printk(KERN_ERR "iosapic_unregister_intr(%u) unbalanced\n",
851 		       gsi);
852 		WARN_ON(1);
853 		goto out;
854 	}
855 
856 	if (--rte->refcnt > 0)
857 		goto out;
858 
859 	idesc = irq_desc + irq;
860 	rte->refcnt = NO_REF_RTE;
861 
862 	/* Mask the interrupt */
863 	low32 = iosapic_intr_info[irq].low32 | IOSAPIC_MASK;
864 	iosapic_write(rte->iosapic, IOSAPIC_RTE_LOW(rte->rte_index), low32);
865 
866 	iosapic_intr_info[irq].count--;
867 	index = find_iosapic(gsi);
868 	iosapic_lists[index].rtes_inuse--;
869 	WARN_ON(iosapic_lists[index].rtes_inuse < 0);
870 
871 	trigger  = iosapic_intr_info[irq].trigger;
872 	polarity = iosapic_intr_info[irq].polarity;
873 	dest     = iosapic_intr_info[irq].dest;
874 	printk(KERN_INFO
875 	       "GSI %u (%s, %s) -> CPU %d (0x%04x) vector %d unregistered\n",
876 	       gsi, (trigger == IOSAPIC_EDGE ? "edge" : "level"),
877 	       (polarity == IOSAPIC_POL_HIGH ? "high" : "low"),
878 	       cpu_logical_id(dest), dest, irq_to_vector(irq));
879 
880 	if (iosapic_intr_info[irq].count == 0) {
881 #ifdef CONFIG_SMP
882 		/* Clear affinity */
883 		cpus_setall(idesc->affinity);
884 #endif
885 		/* Clear the interrupt information */
886 		iosapic_intr_info[irq].dest = 0;
887 		iosapic_intr_info[irq].dmode = 0;
888 		iosapic_intr_info[irq].polarity = 0;
889 		iosapic_intr_info[irq].trigger = 0;
890 		iosapic_intr_info[irq].low32 |= IOSAPIC_MASK;
891 
892 		/* Destroy and reserve IRQ */
893 		destroy_and_reserve_irq(irq);
894 	}
895  out:
896 	spin_unlock_irqrestore(&iosapic_lock, flags);
897 }
898 
899 /*
900  * ACPI calls this when it finds an entry for a platform interrupt.
901  */
902 int __init
iosapic_register_platform_intr(u32 int_type,unsigned int gsi,int iosapic_vector,u16 eid,u16 id,unsigned long polarity,unsigned long trigger)903 iosapic_register_platform_intr (u32 int_type, unsigned int gsi,
904 				int iosapic_vector, u16 eid, u16 id,
905 				unsigned long polarity, unsigned long trigger)
906 {
907 	static const char * const name[] = {"unknown", "PMI", "INIT", "CPEI"};
908 	unsigned char delivery;
909 	int irq, vector, mask = 0;
910 	unsigned int dest = ((id << 8) | eid) & 0xffff;
911 
912 	switch (int_type) {
913 	      case ACPI_INTERRUPT_PMI:
914 		irq = vector = iosapic_vector;
915 		bind_irq_vector(irq, vector, CPU_MASK_ALL);
916 		/*
917 		 * since PMI vector is alloc'd by FW(ACPI) not by kernel,
918 		 * we need to make sure the vector is available
919 		 */
920 		iosapic_reassign_vector(irq);
921 		delivery = IOSAPIC_PMI;
922 		break;
923 	      case ACPI_INTERRUPT_INIT:
924 		irq = create_irq();
925 		if (irq < 0)
926 			panic("%s: out of interrupt vectors!\n", __func__);
927 		vector = irq_to_vector(irq);
928 		delivery = IOSAPIC_INIT;
929 		break;
930 	      case ACPI_INTERRUPT_CPEI:
931 		irq = vector = IA64_CPE_VECTOR;
932 		BUG_ON(bind_irq_vector(irq, vector, CPU_MASK_ALL));
933 		delivery = IOSAPIC_FIXED;
934 		mask = 1;
935 		break;
936 	      default:
937 		printk(KERN_ERR "%s: invalid int type 0x%x\n", __func__,
938 		       int_type);
939 		return -1;
940 	}
941 
942 	register_intr(gsi, irq, delivery, polarity, trigger);
943 
944 	printk(KERN_INFO
945 	       "PLATFORM int %s (0x%x): GSI %u (%s, %s) -> CPU %d (0x%04x)"
946 	       " vector %d\n",
947 	       int_type < ARRAY_SIZE(name) ? name[int_type] : "unknown",
948 	       int_type, gsi, (trigger == IOSAPIC_EDGE ? "edge" : "level"),
949 	       (polarity == IOSAPIC_POL_HIGH ? "high" : "low"),
950 	       cpu_logical_id(dest), dest, vector);
951 
952 	set_rte(gsi, irq, dest, mask);
953 	return vector;
954 }
955 
956 /*
957  * ACPI calls this when it finds an entry for a legacy ISA IRQ override.
958  */
959 void __devinit
iosapic_override_isa_irq(unsigned int isa_irq,unsigned int gsi,unsigned long polarity,unsigned long trigger)960 iosapic_override_isa_irq (unsigned int isa_irq, unsigned int gsi,
961 			  unsigned long polarity,
962 			  unsigned long trigger)
963 {
964 	int vector, irq;
965 	unsigned int dest = cpu_physical_id(smp_processor_id());
966 	unsigned char dmode;
967 
968 	irq = vector = isa_irq_to_vector(isa_irq);
969 	BUG_ON(bind_irq_vector(irq, vector, CPU_MASK_ALL));
970 	dmode = choose_dmode();
971 	register_intr(gsi, irq, dmode, polarity, trigger);
972 
973 	DBG("ISA: IRQ %u -> GSI %u (%s,%s) -> CPU %d (0x%04x) vector %d\n",
974 	    isa_irq, gsi, trigger == IOSAPIC_EDGE ? "edge" : "level",
975 	    polarity == IOSAPIC_POL_HIGH ? "high" : "low",
976 	    cpu_logical_id(dest), dest, vector);
977 
978 	set_rte(gsi, irq, dest, 1);
979 }
980 
981 void __init
ia64_native_iosapic_pcat_compat_init(void)982 ia64_native_iosapic_pcat_compat_init(void)
983 {
984 	if (pcat_compat) {
985 		/*
986 		 * Disable the compatibility mode interrupts (8259 style),
987 		 * needs IN/OUT support enabled.
988 		 */
989 		printk(KERN_INFO
990 		       "%s: Disabling PC-AT compatible 8259 interrupts\n",
991 		       __func__);
992 		outb(0xff, 0xA1);
993 		outb(0xff, 0x21);
994 	}
995 }
996 
997 void __init
iosapic_system_init(int system_pcat_compat)998 iosapic_system_init (int system_pcat_compat)
999 {
1000 	int irq;
1001 
1002 	for (irq = 0; irq < NR_IRQS; ++irq) {
1003 		iosapic_intr_info[irq].low32 = IOSAPIC_MASK;
1004 		/* mark as unused */
1005 		INIT_LIST_HEAD(&iosapic_intr_info[irq].rtes);
1006 
1007 		iosapic_intr_info[irq].count = 0;
1008 	}
1009 
1010 	pcat_compat = system_pcat_compat;
1011 	if (pcat_compat)
1012 		iosapic_pcat_compat_init();
1013 }
1014 
1015 static inline int
iosapic_alloc(void)1016 iosapic_alloc (void)
1017 {
1018 	int index;
1019 
1020 	for (index = 0; index < NR_IOSAPICS; index++)
1021 		if (!iosapic_lists[index].addr)
1022 			return index;
1023 
1024 	printk(KERN_WARNING "%s: failed to allocate iosapic\n", __func__);
1025 	return -1;
1026 }
1027 
1028 static inline void
iosapic_free(int index)1029 iosapic_free (int index)
1030 {
1031 	memset(&iosapic_lists[index], 0, sizeof(iosapic_lists[0]));
1032 }
1033 
1034 static inline int
iosapic_check_gsi_range(unsigned int gsi_base,unsigned int ver)1035 iosapic_check_gsi_range (unsigned int gsi_base, unsigned int ver)
1036 {
1037 	int index;
1038 	unsigned int gsi_end, base, end;
1039 
1040 	/* check gsi range */
1041 	gsi_end = gsi_base + ((ver >> 16) & 0xff);
1042 	for (index = 0; index < NR_IOSAPICS; index++) {
1043 		if (!iosapic_lists[index].addr)
1044 			continue;
1045 
1046 		base = iosapic_lists[index].gsi_base;
1047 		end  = base + iosapic_lists[index].num_rte - 1;
1048 
1049 		if (gsi_end < base || end < gsi_base)
1050 			continue; /* OK */
1051 
1052 		return -EBUSY;
1053 	}
1054 	return 0;
1055 }
1056 
1057 int __devinit
iosapic_init(unsigned long phys_addr,unsigned int gsi_base)1058 iosapic_init (unsigned long phys_addr, unsigned int gsi_base)
1059 {
1060 	int num_rte, err, index;
1061 	unsigned int isa_irq, ver;
1062 	char __iomem *addr;
1063 	unsigned long flags;
1064 
1065 	spin_lock_irqsave(&iosapic_lock, flags);
1066 	index = find_iosapic(gsi_base);
1067 	if (index >= 0) {
1068 		spin_unlock_irqrestore(&iosapic_lock, flags);
1069 		return -EBUSY;
1070 	}
1071 
1072 	addr = ioremap(phys_addr, 0);
1073 	ver = iosapic_version(addr);
1074 	if ((err = iosapic_check_gsi_range(gsi_base, ver))) {
1075 		iounmap(addr);
1076 		spin_unlock_irqrestore(&iosapic_lock, flags);
1077 		return err;
1078 	}
1079 
1080 	/*
1081 	 * The MAX_REDIR register holds the highest input pin number
1082 	 * (starting from 0).  We add 1 so that we can use it for
1083 	 * number of pins (= RTEs)
1084 	 */
1085 	num_rte = ((ver >> 16) & 0xff) + 1;
1086 
1087 	index = iosapic_alloc();
1088 	iosapic_lists[index].addr = addr;
1089 	iosapic_lists[index].gsi_base = gsi_base;
1090 	iosapic_lists[index].num_rte = num_rte;
1091 #ifdef CONFIG_NUMA
1092 	iosapic_lists[index].node = MAX_NUMNODES;
1093 #endif
1094 	spin_lock_init(&iosapic_lists[index].lock);
1095 	spin_unlock_irqrestore(&iosapic_lock, flags);
1096 
1097 	if ((gsi_base == 0) && pcat_compat) {
1098 		/*
1099 		 * Map the legacy ISA devices into the IOSAPIC data.  Some of
1100 		 * these may get reprogrammed later on with data from the ACPI
1101 		 * Interrupt Source Override table.
1102 		 */
1103 		for (isa_irq = 0; isa_irq < 16; ++isa_irq)
1104 			iosapic_override_isa_irq(isa_irq, isa_irq,
1105 						 IOSAPIC_POL_HIGH,
1106 						 IOSAPIC_EDGE);
1107 	}
1108 	return 0;
1109 }
1110 
1111 #ifdef CONFIG_HOTPLUG
1112 int
iosapic_remove(unsigned int gsi_base)1113 iosapic_remove (unsigned int gsi_base)
1114 {
1115 	int index, err = 0;
1116 	unsigned long flags;
1117 
1118 	spin_lock_irqsave(&iosapic_lock, flags);
1119 	index = find_iosapic(gsi_base);
1120 	if (index < 0) {
1121 		printk(KERN_WARNING "%s: No IOSAPIC for GSI base %u\n",
1122 		       __func__, gsi_base);
1123 		goto out;
1124 	}
1125 
1126 	if (iosapic_lists[index].rtes_inuse) {
1127 		err = -EBUSY;
1128 		printk(KERN_WARNING "%s: IOSAPIC for GSI base %u is busy\n",
1129 		       __func__, gsi_base);
1130 		goto out;
1131 	}
1132 
1133 	iounmap(iosapic_lists[index].addr);
1134 	iosapic_free(index);
1135  out:
1136 	spin_unlock_irqrestore(&iosapic_lock, flags);
1137 	return err;
1138 }
1139 #endif /* CONFIG_HOTPLUG */
1140 
1141 #ifdef CONFIG_NUMA
1142 void __devinit
map_iosapic_to_node(unsigned int gsi_base,int node)1143 map_iosapic_to_node(unsigned int gsi_base, int node)
1144 {
1145 	int index;
1146 
1147 	index = find_iosapic(gsi_base);
1148 	if (index < 0) {
1149 		printk(KERN_WARNING "%s: No IOSAPIC for GSI %u\n",
1150 		       __func__, gsi_base);
1151 		return;
1152 	}
1153 	iosapic_lists[index].node = node;
1154 	return;
1155 }
1156 #endif
1157 
iosapic_enable_kmalloc(void)1158 static int __init iosapic_enable_kmalloc (void)
1159 {
1160 	iosapic_kmalloc_ok = 1;
1161 	return 0;
1162 }
1163 core_initcall (iosapic_enable_kmalloc);
1164