• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 #ifndef _LINUX_IRQDESC_H
2 #define _LINUX_IRQDESC_H
3 
4 /*
5  * Core internal functions to deal with irq descriptors
6  *
7  * This include will move to kernel/irq once we cleaned up the tree.
8  * For now it's included from <linux/irq.h>
9  */
10 
11 struct irq_affinity_notify;
12 struct proc_dir_entry;
13 struct module;
14 struct irq_desc;
15 struct irq_domain;
16 struct pt_regs;
17 
18 /**
19  * struct irq_desc - interrupt descriptor
20  * @irq_data:		per irq and chip data passed down to chip functions
21  * @kstat_irqs:		irq stats per cpu
22  * @handle_irq:		highlevel irq-events handler
23  * @preflow_handler:	handler called before the flow handler (currently used by sparc)
24  * @action:		the irq action chain
25  * @status:		status information
26  * @core_internal_state__do_not_mess_with_it: core internal status information
27  * @depth:		disable-depth, for nested irq_disable() calls
28  * @wake_depth:		enable depth, for multiple irq_set_irq_wake() callers
29  * @irq_count:		stats field to detect stalled irqs
30  * @last_unhandled:	aging timer for unhandled count
31  * @irqs_unhandled:	stats field for spurious unhandled interrupts
32  * @threads_handled:	stats field for deferred spurious detection of threaded handlers
33  * @threads_handled_last: comparator field for deferred spurious detection of theraded handlers
34  * @lock:		locking for SMP
35  * @affinity_hint:	hint to user space for preferred irq affinity
36  * @affinity_notify:	context for notification of affinity changes
37  * @pending_mask:	pending rebalanced interrupts
38  * @threads_oneshot:	bitfield to handle shared oneshot threads
39  * @threads_active:	number of irqaction threads currently running
40  * @wait_for_threads:	wait queue for sync_irq to wait for threaded handlers
41  * @nr_actions:		number of installed actions on this descriptor
42  * @no_suspend_depth:	number of irqactions on a irq descriptor with
43  *			IRQF_NO_SUSPEND set
44  * @force_resume_depth:	number of irqactions on a irq descriptor with
45  *			IRQF_FORCE_RESUME set
46  * @dir:		/proc/irq/ procfs entry
47  * @name:		flow handler name for /proc/interrupts output
48  */
49 struct irq_desc {
50 	struct irq_data		irq_data;
51 	unsigned int __percpu	*kstat_irqs;
52 	irq_flow_handler_t	handle_irq;
53 #ifdef CONFIG_IRQ_PREFLOW_FASTEOI
54 	irq_preflow_handler_t	preflow_handler;
55 #endif
56 	struct irqaction	*action;	/* IRQ action list */
57 	unsigned int		status_use_accessors;
58 	unsigned int		core_internal_state__do_not_mess_with_it;
59 	unsigned int		depth;		/* nested irq disables */
60 	unsigned int		wake_depth;	/* nested wake enables */
61 	unsigned int		irq_count;	/* For detecting broken IRQs */
62 	unsigned long		last_unhandled;	/* Aging timer for unhandled count */
63 	unsigned int		irqs_unhandled;
64 	atomic_t		threads_handled;
65 	int			threads_handled_last;
66 	raw_spinlock_t		lock;
67 	struct cpumask		*percpu_enabled;
68 #ifdef CONFIG_SMP
69 	const struct cpumask	*affinity_hint;
70 	struct irq_affinity_notify *affinity_notify;
71 #ifdef CONFIG_GENERIC_PENDING_IRQ
72 	cpumask_var_t		pending_mask;
73 #endif
74 #endif
75 	unsigned long		threads_oneshot;
76 	atomic_t		threads_active;
77 	wait_queue_head_t       wait_for_threads;
78 #ifdef CONFIG_PM_SLEEP
79 	unsigned int		nr_actions;
80 	unsigned int		no_suspend_depth;
81 	unsigned int		force_resume_depth;
82 #endif
83 #ifdef CONFIG_PROC_FS
84 	struct proc_dir_entry	*dir;
85 #endif
86 	int			parent_irq;
87 	struct module		*owner;
88 	const char		*name;
89 } ____cacheline_internodealigned_in_smp;
90 
91 #ifndef CONFIG_SPARSE_IRQ
92 extern struct irq_desc irq_desc[NR_IRQS];
93 #endif
94 
irq_desc_get_irq_data(struct irq_desc * desc)95 static inline struct irq_data *irq_desc_get_irq_data(struct irq_desc *desc)
96 {
97 	return &desc->irq_data;
98 }
99 
irq_desc_get_chip(struct irq_desc * desc)100 static inline struct irq_chip *irq_desc_get_chip(struct irq_desc *desc)
101 {
102 	return desc->irq_data.chip;
103 }
104 
irq_desc_get_chip_data(struct irq_desc * desc)105 static inline void *irq_desc_get_chip_data(struct irq_desc *desc)
106 {
107 	return desc->irq_data.chip_data;
108 }
109 
irq_desc_get_handler_data(struct irq_desc * desc)110 static inline void *irq_desc_get_handler_data(struct irq_desc *desc)
111 {
112 	return desc->irq_data.handler_data;
113 }
114 
irq_desc_get_msi_desc(struct irq_desc * desc)115 static inline struct msi_desc *irq_desc_get_msi_desc(struct irq_desc *desc)
116 {
117 	return desc->irq_data.msi_desc;
118 }
119 
120 /*
121  * Architectures call this to let the generic IRQ layer
122  * handle an interrupt. If the descriptor is attached to an
123  * irqchip-style controller then we call the ->handle_irq() handler,
124  * and it calls __do_IRQ() if it's attached to an irqtype-style controller.
125  */
generic_handle_irq_desc(unsigned int irq,struct irq_desc * desc)126 static inline void generic_handle_irq_desc(unsigned int irq, struct irq_desc *desc)
127 {
128 	desc->handle_irq(irq, desc);
129 }
130 
131 int generic_handle_irq(unsigned int irq);
132 
133 #ifdef CONFIG_HANDLE_DOMAIN_IRQ
134 /*
135  * Convert a HW interrupt number to a logical one using a IRQ domain,
136  * and handle the result interrupt number. Return -EINVAL if
137  * conversion failed. Providing a NULL domain indicates that the
138  * conversion has already been done.
139  */
140 int __handle_domain_irq(struct irq_domain *domain, unsigned int hwirq,
141 			bool lookup, struct pt_regs *regs);
142 
handle_domain_irq(struct irq_domain * domain,unsigned int hwirq,struct pt_regs * regs)143 static inline int handle_domain_irq(struct irq_domain *domain,
144 				    unsigned int hwirq, struct pt_regs *regs)
145 {
146 	return __handle_domain_irq(domain, hwirq, true, regs);
147 }
148 #endif
149 
150 /* Test to see if a driver has successfully requested an irq */
irq_has_action(unsigned int irq)151 static inline int irq_has_action(unsigned int irq)
152 {
153 	struct irq_desc *desc = irq_to_desc(irq);
154 	return desc->action != NULL;
155 }
156 
157 /* caller has locked the irq_desc and both params are valid */
__irq_set_handler_locked(unsigned int irq,irq_flow_handler_t handler)158 static inline void __irq_set_handler_locked(unsigned int irq,
159 					    irq_flow_handler_t handler)
160 {
161 	struct irq_desc *desc;
162 
163 	desc = irq_to_desc(irq);
164 	desc->handle_irq = handler;
165 }
166 
167 /* caller has locked the irq_desc and both params are valid */
168 static inline void
__irq_set_chip_handler_name_locked(unsigned int irq,struct irq_chip * chip,irq_flow_handler_t handler,const char * name)169 __irq_set_chip_handler_name_locked(unsigned int irq, struct irq_chip *chip,
170 				   irq_flow_handler_t handler, const char *name)
171 {
172 	struct irq_desc *desc;
173 
174 	desc = irq_to_desc(irq);
175 	irq_desc_get_irq_data(desc)->chip = chip;
176 	desc->handle_irq = handler;
177 	desc->name = name;
178 }
179 
irq_balancing_disabled(unsigned int irq)180 static inline int irq_balancing_disabled(unsigned int irq)
181 {
182 	struct irq_desc *desc;
183 
184 	desc = irq_to_desc(irq);
185 	return desc->status_use_accessors & IRQ_NO_BALANCING_MASK;
186 }
187 
irq_is_percpu(unsigned int irq)188 static inline int irq_is_percpu(unsigned int irq)
189 {
190 	struct irq_desc *desc;
191 
192 	desc = irq_to_desc(irq);
193 	return desc->status_use_accessors & IRQ_PER_CPU;
194 }
195 
196 static inline void
irq_set_lockdep_class(unsigned int irq,struct lock_class_key * class)197 irq_set_lockdep_class(unsigned int irq, struct lock_class_key *class)
198 {
199 	struct irq_desc *desc = irq_to_desc(irq);
200 
201 	if (desc)
202 		lockdep_set_class(&desc->lock, class);
203 }
204 
205 #ifdef CONFIG_IRQ_PREFLOW_FASTEOI
206 static inline void
__irq_set_preflow_handler(unsigned int irq,irq_preflow_handler_t handler)207 __irq_set_preflow_handler(unsigned int irq, irq_preflow_handler_t handler)
208 {
209 	struct irq_desc *desc;
210 
211 	desc = irq_to_desc(irq);
212 	desc->preflow_handler = handler;
213 }
214 #endif
215 
216 #endif
217