1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * PCI Message Signaled Interrupt (MSI)
4  *
5  * Copyright (C) 2003-2004 Intel
6  * Copyright (C) Tom Long Nguyen (tom.l.nguyen@intel.com)
7  * Copyright (C) 2016 Christoph Hellwig.
8  */
9 #include <linux/bitfield.h>
10 #include <linux/err.h>
11 #include <linux/export.h>
12 #include <linux/irq.h>
13 #include <linux/irqdomain.h>
14 #include <linux/android_kabi.h>
15 ANDROID_KABI_DECLONLY(device_node);
16 ANDROID_KABI_DECLONLY(irq_domain);
17 
18 #include "../pci.h"
19 #include "msi.h"
20 
21 int pci_msi_enable = 1;
22 
23 /**
24  * pci_msi_supported - check whether MSI may be enabled on a device
25  * @dev: pointer to the pci_dev data structure of MSI device function
26  * @nvec: how many MSIs have been requested?
27  *
28  * Look at global flags, the device itself, and its parent buses
29  * to determine if MSI/-X are supported for the device. If MSI/-X is
30  * supported return 1, else return 0.
31  **/
pci_msi_supported(struct pci_dev * dev,int nvec)32 static int pci_msi_supported(struct pci_dev *dev, int nvec)
33 {
34 	struct pci_bus *bus;
35 
36 	/* MSI must be globally enabled and supported by the device */
37 	if (!pci_msi_enable)
38 		return 0;
39 
40 	if (!dev || dev->no_msi)
41 		return 0;
42 
43 	/*
44 	 * You can't ask to have 0 or less MSIs configured.
45 	 *  a) it's stupid ..
46 	 *  b) the list manipulation code assumes nvec >= 1.
47 	 */
48 	if (nvec < 1)
49 		return 0;
50 
51 	/*
52 	 * Any bridge which does NOT route MSI transactions from its
53 	 * secondary bus to its primary bus must set NO_MSI flag on
54 	 * the secondary pci_bus.
55 	 *
56 	 * The NO_MSI flag can either be set directly by:
57 	 * - arch-specific PCI host bus controller drivers (deprecated)
58 	 * - quirks for specific PCI bridges
59 	 *
60 	 * or indirectly by platform-specific PCI host bridge drivers by
61 	 * advertising the 'msi_domain' property, which results in
62 	 * the NO_MSI flag when no MSI domain is found for this bridge
63 	 * at probe time.
64 	 */
65 	for (bus = dev->bus; bus; bus = bus->parent)
66 		if (bus->bus_flags & PCI_BUS_FLAGS_NO_MSI)
67 			return 0;
68 
69 	return 1;
70 }
71 
pcim_msi_release(void * pcidev)72 static void pcim_msi_release(void *pcidev)
73 {
74 	struct pci_dev *dev = pcidev;
75 
76 	dev->is_msi_managed = false;
77 	pci_free_irq_vectors(dev);
78 }
79 
80 /*
81  * Needs to be separate from pcim_release to prevent an ordering problem
82  * vs. msi_device_data_release() in the MSI core code.
83  */
pcim_setup_msi_release(struct pci_dev * dev)84 static int pcim_setup_msi_release(struct pci_dev *dev)
85 {
86 	int ret;
87 
88 	if (!pci_is_managed(dev) || dev->is_msi_managed)
89 		return 0;
90 
91 	ret = devm_add_action(&dev->dev, pcim_msi_release, dev);
92 	if (ret)
93 		return ret;
94 
95 	dev->is_msi_managed = true;
96 	return 0;
97 }
98 
99 /*
100  * Ordering vs. devres: msi device data has to be installed first so that
101  * pcim_msi_release() is invoked before it on device release.
102  */
pci_setup_msi_context(struct pci_dev * dev)103 static int pci_setup_msi_context(struct pci_dev *dev)
104 {
105 	int ret = msi_setup_device_data(&dev->dev);
106 
107 	if (ret)
108 		return ret;
109 
110 	return pcim_setup_msi_release(dev);
111 }
112 
113 /*
114  * Helper functions for mask/unmask and MSI message handling
115  */
116 
pci_msi_update_mask(struct msi_desc * desc,u32 clear,u32 set)117 void pci_msi_update_mask(struct msi_desc *desc, u32 clear, u32 set)
118 {
119 	raw_spinlock_t *lock = &to_pci_dev(desc->dev)->msi_lock;
120 	unsigned long flags;
121 
122 	if (!desc->pci.msi_attrib.can_mask)
123 		return;
124 
125 	raw_spin_lock_irqsave(lock, flags);
126 	desc->pci.msi_mask &= ~clear;
127 	desc->pci.msi_mask |= set;
128 	pci_write_config_dword(msi_desc_to_pci_dev(desc), desc->pci.mask_pos,
129 			       desc->pci.msi_mask);
130 	raw_spin_unlock_irqrestore(lock, flags);
131 }
132 
133 /**
134  * pci_msi_mask_irq - Generic IRQ chip callback to mask PCI/MSI interrupts
135  * @data:	pointer to irqdata associated to that interrupt
136  */
pci_msi_mask_irq(struct irq_data * data)137 void pci_msi_mask_irq(struct irq_data *data)
138 {
139 	struct msi_desc *desc = irq_data_get_msi_desc(data);
140 
141 	__pci_msi_mask_desc(desc, BIT(data->irq - desc->irq));
142 }
143 EXPORT_SYMBOL_GPL(pci_msi_mask_irq);
144 
145 /**
146  * pci_msi_unmask_irq - Generic IRQ chip callback to unmask PCI/MSI interrupts
147  * @data:	pointer to irqdata associated to that interrupt
148  */
pci_msi_unmask_irq(struct irq_data * data)149 void pci_msi_unmask_irq(struct irq_data *data)
150 {
151 	struct msi_desc *desc = irq_data_get_msi_desc(data);
152 
153 	__pci_msi_unmask_desc(desc, BIT(data->irq - desc->irq));
154 }
155 EXPORT_SYMBOL_GPL(pci_msi_unmask_irq);
156 
__pci_read_msi_msg(struct msi_desc * entry,struct msi_msg * msg)157 void __pci_read_msi_msg(struct msi_desc *entry, struct msi_msg *msg)
158 {
159 	struct pci_dev *dev = msi_desc_to_pci_dev(entry);
160 
161 	BUG_ON(dev->current_state != PCI_D0);
162 
163 	if (entry->pci.msi_attrib.is_msix) {
164 		void __iomem *base = pci_msix_desc_addr(entry);
165 
166 		if (WARN_ON_ONCE(entry->pci.msi_attrib.is_virtual))
167 			return;
168 
169 		msg->address_lo = readl(base + PCI_MSIX_ENTRY_LOWER_ADDR);
170 		msg->address_hi = readl(base + PCI_MSIX_ENTRY_UPPER_ADDR);
171 		msg->data = readl(base + PCI_MSIX_ENTRY_DATA);
172 	} else {
173 		int pos = dev->msi_cap;
174 		u16 data;
175 
176 		pci_read_config_dword(dev, pos + PCI_MSI_ADDRESS_LO,
177 				      &msg->address_lo);
178 		if (entry->pci.msi_attrib.is_64) {
179 			pci_read_config_dword(dev, pos + PCI_MSI_ADDRESS_HI,
180 					      &msg->address_hi);
181 			pci_read_config_word(dev, pos + PCI_MSI_DATA_64, &data);
182 		} else {
183 			msg->address_hi = 0;
184 			pci_read_config_word(dev, pos + PCI_MSI_DATA_32, &data);
185 		}
186 		msg->data = data;
187 	}
188 }
189 
pci_write_msg_msi(struct pci_dev * dev,struct msi_desc * desc,struct msi_msg * msg)190 static inline void pci_write_msg_msi(struct pci_dev *dev, struct msi_desc *desc,
191 				     struct msi_msg *msg)
192 {
193 	int pos = dev->msi_cap;
194 	u16 msgctl;
195 
196 	pci_read_config_word(dev, pos + PCI_MSI_FLAGS, &msgctl);
197 	msgctl &= ~PCI_MSI_FLAGS_QSIZE;
198 	msgctl |= FIELD_PREP(PCI_MSI_FLAGS_QSIZE, desc->pci.msi_attrib.multiple);
199 	pci_write_config_word(dev, pos + PCI_MSI_FLAGS, msgctl);
200 
201 	pci_write_config_dword(dev, pos + PCI_MSI_ADDRESS_LO, msg->address_lo);
202 	if (desc->pci.msi_attrib.is_64) {
203 		pci_write_config_dword(dev, pos + PCI_MSI_ADDRESS_HI,  msg->address_hi);
204 		pci_write_config_word(dev, pos + PCI_MSI_DATA_64, msg->data);
205 	} else {
206 		pci_write_config_word(dev, pos + PCI_MSI_DATA_32, msg->data);
207 	}
208 	/* Ensure that the writes are visible in the device */
209 	pci_read_config_word(dev, pos + PCI_MSI_FLAGS, &msgctl);
210 }
211 
pci_write_msg_msix(struct msi_desc * desc,struct msi_msg * msg)212 static inline void pci_write_msg_msix(struct msi_desc *desc, struct msi_msg *msg)
213 {
214 	void __iomem *base = pci_msix_desc_addr(desc);
215 	u32 ctrl = desc->pci.msix_ctrl;
216 	bool unmasked = !(ctrl & PCI_MSIX_ENTRY_CTRL_MASKBIT);
217 
218 	if (desc->pci.msi_attrib.is_virtual)
219 		return;
220 	/*
221 	 * The specification mandates that the entry is masked
222 	 * when the message is modified:
223 	 *
224 	 * "If software changes the Address or Data value of an
225 	 * entry while the entry is unmasked, the result is
226 	 * undefined."
227 	 */
228 	if (unmasked)
229 		pci_msix_write_vector_ctrl(desc, ctrl | PCI_MSIX_ENTRY_CTRL_MASKBIT);
230 
231 	writel(msg->address_lo, base + PCI_MSIX_ENTRY_LOWER_ADDR);
232 	writel(msg->address_hi, base + PCI_MSIX_ENTRY_UPPER_ADDR);
233 	writel(msg->data, base + PCI_MSIX_ENTRY_DATA);
234 
235 	if (unmasked)
236 		pci_msix_write_vector_ctrl(desc, ctrl);
237 
238 	/* Ensure that the writes are visible in the device */
239 	readl(base + PCI_MSIX_ENTRY_DATA);
240 }
241 
__pci_write_msi_msg(struct msi_desc * entry,struct msi_msg * msg)242 void __pci_write_msi_msg(struct msi_desc *entry, struct msi_msg *msg)
243 {
244 	struct pci_dev *dev = msi_desc_to_pci_dev(entry);
245 
246 	if (dev->current_state != PCI_D0 || pci_dev_is_disconnected(dev)) {
247 		/* Don't touch the hardware now */
248 	} else if (entry->pci.msi_attrib.is_msix) {
249 		pci_write_msg_msix(entry, msg);
250 	} else {
251 		pci_write_msg_msi(dev, entry, msg);
252 	}
253 
254 	entry->msg = *msg;
255 
256 	if (entry->write_msi_msg)
257 		entry->write_msi_msg(entry, entry->write_msi_msg_data);
258 }
259 
pci_write_msi_msg(unsigned int irq,struct msi_msg * msg)260 void pci_write_msi_msg(unsigned int irq, struct msi_msg *msg)
261 {
262 	struct msi_desc *entry = irq_get_msi_desc(irq);
263 
264 	__pci_write_msi_msg(entry, msg);
265 }
266 EXPORT_SYMBOL_GPL(pci_write_msi_msg);
267 
268 
269 /* PCI/MSI specific functionality */
270 
pci_intx_for_msi(struct pci_dev * dev,int enable)271 static void pci_intx_for_msi(struct pci_dev *dev, int enable)
272 {
273 	if (!(dev->dev_flags & PCI_DEV_FLAGS_MSI_INTX_DISABLE_BUG))
274 		pci_intx(dev, enable);
275 }
276 
pci_msi_set_enable(struct pci_dev * dev,int enable)277 static void pci_msi_set_enable(struct pci_dev *dev, int enable)
278 {
279 	u16 control;
280 
281 	pci_read_config_word(dev, dev->msi_cap + PCI_MSI_FLAGS, &control);
282 	control &= ~PCI_MSI_FLAGS_ENABLE;
283 	if (enable)
284 		control |= PCI_MSI_FLAGS_ENABLE;
285 	pci_write_config_word(dev, dev->msi_cap + PCI_MSI_FLAGS, control);
286 }
287 
msi_setup_msi_desc(struct pci_dev * dev,int nvec,struct irq_affinity_desc * masks)288 static int msi_setup_msi_desc(struct pci_dev *dev, int nvec,
289 			      struct irq_affinity_desc *masks)
290 {
291 	struct msi_desc desc;
292 	u16 control;
293 
294 	/* MSI Entry Initialization */
295 	memset(&desc, 0, sizeof(desc));
296 
297 	pci_read_config_word(dev, dev->msi_cap + PCI_MSI_FLAGS, &control);
298 	/* Lies, damned lies, and MSIs */
299 	if (dev->dev_flags & PCI_DEV_FLAGS_HAS_MSI_MASKING)
300 		control |= PCI_MSI_FLAGS_MASKBIT;
301 	if (pci_msi_domain_supports(dev, MSI_FLAG_NO_MASK, DENY_LEGACY))
302 		control &= ~PCI_MSI_FLAGS_MASKBIT;
303 
304 	desc.nvec_used			= nvec;
305 	desc.pci.msi_attrib.is_64	= !!(control & PCI_MSI_FLAGS_64BIT);
306 	desc.pci.msi_attrib.can_mask	= !!(control & PCI_MSI_FLAGS_MASKBIT);
307 	desc.pci.msi_attrib.default_irq	= dev->irq;
308 	desc.pci.msi_attrib.multi_cap	= FIELD_GET(PCI_MSI_FLAGS_QMASK, control);
309 	desc.pci.msi_attrib.multiple	= ilog2(__roundup_pow_of_two(nvec));
310 	desc.affinity			= masks;
311 
312 	if (control & PCI_MSI_FLAGS_64BIT)
313 		desc.pci.mask_pos = dev->msi_cap + PCI_MSI_MASK_64;
314 	else
315 		desc.pci.mask_pos = dev->msi_cap + PCI_MSI_MASK_32;
316 
317 	/* Save the initial mask status */
318 	if (desc.pci.msi_attrib.can_mask)
319 		pci_read_config_dword(dev, desc.pci.mask_pos, &desc.pci.msi_mask);
320 
321 	return msi_insert_msi_desc(&dev->dev, &desc);
322 }
323 
msi_verify_entries(struct pci_dev * dev)324 static int msi_verify_entries(struct pci_dev *dev)
325 {
326 	struct msi_desc *entry;
327 
328 	if (!dev->no_64bit_msi)
329 		return 0;
330 
331 	msi_for_each_desc(entry, &dev->dev, MSI_DESC_ALL) {
332 		if (entry->msg.address_hi) {
333 			pci_err(dev, "arch assigned 64-bit MSI address %#x%08x but device only supports 32 bits\n",
334 				entry->msg.address_hi, entry->msg.address_lo);
335 			break;
336 		}
337 	}
338 	return !entry ? 0 : -EIO;
339 }
340 
341 /**
342  * msi_capability_init - configure device's MSI capability structure
343  * @dev: pointer to the pci_dev data structure of MSI device function
344  * @nvec: number of interrupts to allocate
345  * @affd: description of automatic IRQ affinity assignments (may be %NULL)
346  *
347  * Setup the MSI capability structure of the device with the requested
348  * number of interrupts.  A return value of zero indicates the successful
349  * setup of an entry with the new MSI IRQ.  A negative return value indicates
350  * an error, and a positive return value indicates the number of interrupts
351  * which could have been allocated.
352  */
msi_capability_init(struct pci_dev * dev,int nvec,struct irq_affinity * affd)353 static int msi_capability_init(struct pci_dev *dev, int nvec,
354 			       struct irq_affinity *affd)
355 {
356 	struct irq_affinity_desc *masks = NULL;
357 	struct msi_desc *entry, desc;
358 	int ret;
359 
360 	/* Reject multi-MSI early on irq domain enabled architectures */
361 	if (nvec > 1 && !pci_msi_domain_supports(dev, MSI_FLAG_MULTI_PCI_MSI, ALLOW_LEGACY))
362 		return 1;
363 
364 	/*
365 	 * Disable MSI during setup in the hardware, but mark it enabled
366 	 * so that setup code can evaluate it.
367 	 */
368 	pci_msi_set_enable(dev, 0);
369 	dev->msi_enabled = 1;
370 
371 	if (affd)
372 		masks = irq_create_affinity_masks(nvec, affd);
373 
374 	msi_lock_descs(&dev->dev);
375 	ret = msi_setup_msi_desc(dev, nvec, masks);
376 	if (ret)
377 		goto fail;
378 
379 	/* All MSIs are unmasked by default; mask them all */
380 	entry = msi_first_desc(&dev->dev, MSI_DESC_ALL);
381 	pci_msi_mask(entry, msi_multi_mask(entry));
382 	/*
383 	 * Copy the MSI descriptor for the error path because
384 	 * pci_msi_setup_msi_irqs() will free it for the hierarchical
385 	 * interrupt domain case.
386 	 */
387 	memcpy(&desc, entry, sizeof(desc));
388 
389 	/* Configure MSI capability structure */
390 	ret = pci_msi_setup_msi_irqs(dev, nvec, PCI_CAP_ID_MSI);
391 	if (ret)
392 		goto err;
393 
394 	ret = msi_verify_entries(dev);
395 	if (ret)
396 		goto err;
397 
398 	/* Set MSI enabled bits	*/
399 	pci_intx_for_msi(dev, 0);
400 	pci_msi_set_enable(dev, 1);
401 
402 	pcibios_free_irq(dev);
403 	dev->irq = entry->irq;
404 	goto unlock;
405 
406 err:
407 	pci_msi_unmask(&desc, msi_multi_mask(&desc));
408 	pci_free_msi_irqs(dev);
409 fail:
410 	dev->msi_enabled = 0;
411 unlock:
412 	msi_unlock_descs(&dev->dev);
413 	kfree(masks);
414 	return ret;
415 }
416 
__pci_enable_msi_range(struct pci_dev * dev,int minvec,int maxvec,struct irq_affinity * affd)417 int __pci_enable_msi_range(struct pci_dev *dev, int minvec, int maxvec,
418 			   struct irq_affinity *affd)
419 {
420 	int nvec;
421 	int rc;
422 
423 	if (!pci_msi_supported(dev, minvec) || dev->current_state != PCI_D0)
424 		return -EINVAL;
425 
426 	/* Check whether driver already requested MSI-X IRQs */
427 	if (dev->msix_enabled) {
428 		pci_info(dev, "can't enable MSI (MSI-X already enabled)\n");
429 		return -EINVAL;
430 	}
431 
432 	if (maxvec < minvec)
433 		return -ERANGE;
434 
435 	if (WARN_ON_ONCE(dev->msi_enabled))
436 		return -EINVAL;
437 
438 	/* Test for the availability of MSI support */
439 	if (!pci_msi_domain_supports(dev, 0, ALLOW_LEGACY))
440 		return -ENOTSUPP;
441 
442 	nvec = pci_msi_vec_count(dev);
443 	if (nvec < 0)
444 		return nvec;
445 	if (nvec < minvec)
446 		return -ENOSPC;
447 
448 	if (nvec > maxvec)
449 		nvec = maxvec;
450 
451 	rc = pci_setup_msi_context(dev);
452 	if (rc)
453 		return rc;
454 
455 	if (!pci_setup_msi_device_domain(dev))
456 		return -ENODEV;
457 
458 	for (;;) {
459 		if (affd) {
460 			nvec = irq_calc_affinity_vectors(minvec, nvec, affd);
461 			if (nvec < minvec)
462 				return -ENOSPC;
463 		}
464 
465 		rc = msi_capability_init(dev, nvec, affd);
466 		if (rc == 0)
467 			return nvec;
468 
469 		if (rc < 0)
470 			return rc;
471 		if (rc < minvec)
472 			return -ENOSPC;
473 
474 		nvec = rc;
475 	}
476 }
477 
478 /**
479  * pci_msi_vec_count - Return the number of MSI vectors a device can send
480  * @dev: device to report about
481  *
482  * This function returns the number of MSI vectors a device requested via
483  * Multiple Message Capable register. It returns a negative errno if the
484  * device is not capable sending MSI interrupts. Otherwise, the call succeeds
485  * and returns a power of two, up to a maximum of 2^5 (32), according to the
486  * MSI specification.
487  **/
pci_msi_vec_count(struct pci_dev * dev)488 int pci_msi_vec_count(struct pci_dev *dev)
489 {
490 	int ret;
491 	u16 msgctl;
492 
493 	if (!dev->msi_cap)
494 		return -EINVAL;
495 
496 	pci_read_config_word(dev, dev->msi_cap + PCI_MSI_FLAGS, &msgctl);
497 	ret = 1 << FIELD_GET(PCI_MSI_FLAGS_QMASK, msgctl);
498 
499 	return ret;
500 }
501 EXPORT_SYMBOL(pci_msi_vec_count);
502 
503 /*
504  * Architecture override returns true when the PCI MSI message should be
505  * written by the generic restore function.
506  */
arch_restore_msi_irqs(struct pci_dev * dev)507 bool __weak arch_restore_msi_irqs(struct pci_dev *dev)
508 {
509 	return true;
510 }
511 
__pci_restore_msi_state(struct pci_dev * dev)512 void __pci_restore_msi_state(struct pci_dev *dev)
513 {
514 	struct msi_desc *entry;
515 	u16 control;
516 
517 	if (!dev->msi_enabled)
518 		return;
519 
520 	entry = irq_get_msi_desc(dev->irq);
521 
522 	pci_intx_for_msi(dev, 0);
523 	pci_msi_set_enable(dev, 0);
524 	if (arch_restore_msi_irqs(dev))
525 		__pci_write_msi_msg(entry, &entry->msg);
526 
527 	pci_read_config_word(dev, dev->msi_cap + PCI_MSI_FLAGS, &control);
528 	pci_msi_update_mask(entry, 0, 0);
529 	control &= ~PCI_MSI_FLAGS_QSIZE;
530 	control |= PCI_MSI_FLAGS_ENABLE |
531 		   FIELD_PREP(PCI_MSI_FLAGS_QSIZE, entry->pci.msi_attrib.multiple);
532 	pci_write_config_word(dev, dev->msi_cap + PCI_MSI_FLAGS, control);
533 }
534 
pci_msi_shutdown(struct pci_dev * dev)535 void pci_msi_shutdown(struct pci_dev *dev)
536 {
537 	struct msi_desc *desc;
538 
539 	if (!pci_msi_enable || !dev || !dev->msi_enabled)
540 		return;
541 
542 	pci_msi_set_enable(dev, 0);
543 	pci_intx_for_msi(dev, 1);
544 	dev->msi_enabled = 0;
545 
546 	/* Return the device with MSI unmasked as initial states */
547 	desc = msi_first_desc(&dev->dev, MSI_DESC_ALL);
548 	if (!WARN_ON_ONCE(!desc))
549 		pci_msi_unmask(desc, msi_multi_mask(desc));
550 
551 	/* Restore dev->irq to its default pin-assertion IRQ */
552 	dev->irq = desc->pci.msi_attrib.default_irq;
553 	pcibios_alloc_irq(dev);
554 }
555 
556 /* PCI/MSI-X specific functionality */
557 
pci_msix_clear_and_set_ctrl(struct pci_dev * dev,u16 clear,u16 set)558 static void pci_msix_clear_and_set_ctrl(struct pci_dev *dev, u16 clear, u16 set)
559 {
560 	u16 ctrl;
561 
562 	pci_read_config_word(dev, dev->msix_cap + PCI_MSIX_FLAGS, &ctrl);
563 	ctrl &= ~clear;
564 	ctrl |= set;
565 	pci_write_config_word(dev, dev->msix_cap + PCI_MSIX_FLAGS, ctrl);
566 }
567 
msix_map_region(struct pci_dev * dev,unsigned int nr_entries)568 static void __iomem *msix_map_region(struct pci_dev *dev,
569 				     unsigned int nr_entries)
570 {
571 	resource_size_t phys_addr;
572 	u32 table_offset;
573 	unsigned long flags;
574 	u8 bir;
575 
576 	pci_read_config_dword(dev, dev->msix_cap + PCI_MSIX_TABLE,
577 			      &table_offset);
578 	bir = (u8)(table_offset & PCI_MSIX_TABLE_BIR);
579 	flags = pci_resource_flags(dev, bir);
580 	if (!flags || (flags & IORESOURCE_UNSET))
581 		return NULL;
582 
583 	table_offset &= PCI_MSIX_TABLE_OFFSET;
584 	phys_addr = pci_resource_start(dev, bir) + table_offset;
585 
586 	return ioremap(phys_addr, nr_entries * PCI_MSIX_ENTRY_SIZE);
587 }
588 
589 /**
590  * msix_prepare_msi_desc - Prepare a half initialized MSI descriptor for operation
591  * @dev:	The PCI device for which the descriptor is prepared
592  * @desc:	The MSI descriptor for preparation
593  *
594  * This is separate from msix_setup_msi_descs() below to handle dynamic
595  * allocations for MSI-X after initial enablement.
596  *
597  * Ideally the whole MSI-X setup would work that way, but there is no way to
598  * support this for the legacy arch_setup_msi_irqs() mechanism and for the
599  * fake irq domains like the x86 XEN one. Sigh...
600  *
601  * The descriptor is zeroed and only @desc::msi_index and @desc::affinity
602  * are set. When called from msix_setup_msi_descs() then the is_virtual
603  * attribute is initialized as well.
604  *
605  * Fill in the rest.
606  */
msix_prepare_msi_desc(struct pci_dev * dev,struct msi_desc * desc)607 void msix_prepare_msi_desc(struct pci_dev *dev, struct msi_desc *desc)
608 {
609 	desc->nvec_used				= 1;
610 	desc->pci.msi_attrib.is_msix		= 1;
611 	desc->pci.msi_attrib.is_64		= 1;
612 	desc->pci.msi_attrib.default_irq	= dev->irq;
613 	desc->pci.mask_base			= dev->msix_base;
614 
615 
616 	if (!pci_msi_domain_supports(dev, MSI_FLAG_NO_MASK, DENY_LEGACY) &&
617 	    !desc->pci.msi_attrib.is_virtual) {
618 		void __iomem *addr = pci_msix_desc_addr(desc);
619 
620 		desc->pci.msi_attrib.can_mask = 1;
621 		/* Workaround for SUN NIU insanity, which requires write before read */
622 		if (dev->dev_flags & PCI_DEV_FLAGS_MSIX_TOUCH_ENTRY_DATA_FIRST)
623 			writel(0, addr + PCI_MSIX_ENTRY_DATA);
624 		desc->pci.msix_ctrl = readl(addr + PCI_MSIX_ENTRY_VECTOR_CTRL);
625 	}
626 }
627 
msix_setup_msi_descs(struct pci_dev * dev,struct msix_entry * entries,int nvec,struct irq_affinity_desc * masks)628 static int msix_setup_msi_descs(struct pci_dev *dev, struct msix_entry *entries,
629 				int nvec, struct irq_affinity_desc *masks)
630 {
631 	int ret = 0, i, vec_count = pci_msix_vec_count(dev);
632 	struct irq_affinity_desc *curmsk;
633 	struct msi_desc desc;
634 
635 	memset(&desc, 0, sizeof(desc));
636 
637 	for (i = 0, curmsk = masks; i < nvec; i++, curmsk++) {
638 		desc.msi_index = entries ? entries[i].entry : i;
639 		desc.affinity = masks ? curmsk : NULL;
640 		desc.pci.msi_attrib.is_virtual = desc.msi_index >= vec_count;
641 
642 		msix_prepare_msi_desc(dev, &desc);
643 
644 		ret = msi_insert_msi_desc(&dev->dev, &desc);
645 		if (ret)
646 			break;
647 	}
648 	return ret;
649 }
650 
msix_update_entries(struct pci_dev * dev,struct msix_entry * entries)651 static void msix_update_entries(struct pci_dev *dev, struct msix_entry *entries)
652 {
653 	struct msi_desc *desc;
654 
655 	if (entries) {
656 		msi_for_each_desc(desc, &dev->dev, MSI_DESC_ALL) {
657 			entries->vector = desc->irq;
658 			entries++;
659 		}
660 	}
661 }
662 
msix_mask_all(void __iomem * base,int tsize)663 static void msix_mask_all(void __iomem *base, int tsize)
664 {
665 	u32 ctrl = PCI_MSIX_ENTRY_CTRL_MASKBIT;
666 	int i;
667 
668 	for (i = 0; i < tsize; i++, base += PCI_MSIX_ENTRY_SIZE)
669 		writel(ctrl, base + PCI_MSIX_ENTRY_VECTOR_CTRL);
670 }
671 
msix_setup_interrupts(struct pci_dev * dev,struct msix_entry * entries,int nvec,struct irq_affinity * affd)672 static int msix_setup_interrupts(struct pci_dev *dev, struct msix_entry *entries,
673 				 int nvec, struct irq_affinity *affd)
674 {
675 	struct irq_affinity_desc *masks = NULL;
676 	int ret;
677 
678 	if (affd)
679 		masks = irq_create_affinity_masks(nvec, affd);
680 
681 	msi_lock_descs(&dev->dev);
682 	ret = msix_setup_msi_descs(dev, entries, nvec, masks);
683 	if (ret)
684 		goto out_free;
685 
686 	ret = pci_msi_setup_msi_irqs(dev, nvec, PCI_CAP_ID_MSIX);
687 	if (ret)
688 		goto out_free;
689 
690 	/* Check if all MSI entries honor device restrictions */
691 	ret = msi_verify_entries(dev);
692 	if (ret)
693 		goto out_free;
694 
695 	msix_update_entries(dev, entries);
696 	goto out_unlock;
697 
698 out_free:
699 	pci_free_msi_irqs(dev);
700 out_unlock:
701 	msi_unlock_descs(&dev->dev);
702 	kfree(masks);
703 	return ret;
704 }
705 
706 /**
707  * msix_capability_init - configure device's MSI-X capability
708  * @dev: pointer to the pci_dev data structure of MSI-X device function
709  * @entries: pointer to an array of struct msix_entry entries
710  * @nvec: number of @entries
711  * @affd: Optional pointer to enable automatic affinity assignment
712  *
713  * Setup the MSI-X capability structure of device function with a
714  * single MSI-X IRQ. A return of zero indicates the successful setup of
715  * requested MSI-X entries with allocated IRQs or non-zero for otherwise.
716  **/
msix_capability_init(struct pci_dev * dev,struct msix_entry * entries,int nvec,struct irq_affinity * affd)717 static int msix_capability_init(struct pci_dev *dev, struct msix_entry *entries,
718 				int nvec, struct irq_affinity *affd)
719 {
720 	int ret, tsize;
721 	u16 control;
722 
723 	/*
724 	 * Some devices require MSI-X to be enabled before the MSI-X
725 	 * registers can be accessed.  Mask all the vectors to prevent
726 	 * interrupts coming in before they're fully set up.
727 	 */
728 	pci_msix_clear_and_set_ctrl(dev, 0, PCI_MSIX_FLAGS_MASKALL |
729 				    PCI_MSIX_FLAGS_ENABLE);
730 
731 	/* Mark it enabled so setup functions can query it */
732 	dev->msix_enabled = 1;
733 
734 	pci_read_config_word(dev, dev->msix_cap + PCI_MSIX_FLAGS, &control);
735 	/* Request & Map MSI-X table region */
736 	tsize = msix_table_size(control);
737 	dev->msix_base = msix_map_region(dev, tsize);
738 	if (!dev->msix_base) {
739 		ret = -ENOMEM;
740 		goto out_disable;
741 	}
742 
743 	ret = msix_setup_interrupts(dev, entries, nvec, affd);
744 	if (ret)
745 		goto out_disable;
746 
747 	/* Disable INTX */
748 	pci_intx_for_msi(dev, 0);
749 
750 	if (!pci_msi_domain_supports(dev, MSI_FLAG_NO_MASK, DENY_LEGACY)) {
751 		/*
752 		 * Ensure that all table entries are masked to prevent
753 		 * stale entries from firing in a crash kernel.
754 		 *
755 		 * Done late to deal with a broken Marvell NVME device
756 		 * which takes the MSI-X mask bits into account even
757 		 * when MSI-X is disabled, which prevents MSI delivery.
758 		 */
759 		msix_mask_all(dev->msix_base, tsize);
760 	}
761 	pci_msix_clear_and_set_ctrl(dev, PCI_MSIX_FLAGS_MASKALL, 0);
762 
763 	pcibios_free_irq(dev);
764 	return 0;
765 
766 out_disable:
767 	dev->msix_enabled = 0;
768 	pci_msix_clear_and_set_ctrl(dev, PCI_MSIX_FLAGS_MASKALL | PCI_MSIX_FLAGS_ENABLE, 0);
769 
770 	return ret;
771 }
772 
pci_msix_validate_entries(struct pci_dev * dev,struct msix_entry * entries,int nvec)773 static bool pci_msix_validate_entries(struct pci_dev *dev, struct msix_entry *entries, int nvec)
774 {
775 	bool nogap;
776 	int i, j;
777 
778 	if (!entries)
779 		return true;
780 
781 	nogap = pci_msi_domain_supports(dev, MSI_FLAG_MSIX_CONTIGUOUS, DENY_LEGACY);
782 
783 	for (i = 0; i < nvec; i++) {
784 		/* Check for duplicate entries */
785 		for (j = i + 1; j < nvec; j++) {
786 			if (entries[i].entry == entries[j].entry)
787 				return false;
788 		}
789 		/* Check for unsupported gaps */
790 		if (nogap && entries[i].entry != i)
791 			return false;
792 	}
793 	return true;
794 }
795 
__pci_enable_msix_range(struct pci_dev * dev,struct msix_entry * entries,int minvec,int maxvec,struct irq_affinity * affd,int flags)796 int __pci_enable_msix_range(struct pci_dev *dev, struct msix_entry *entries, int minvec,
797 			    int maxvec, struct irq_affinity *affd, int flags)
798 {
799 	int hwsize, rc, nvec = maxvec;
800 
801 	if (maxvec < minvec)
802 		return -ERANGE;
803 
804 	if (dev->msi_enabled) {
805 		pci_info(dev, "can't enable MSI-X (MSI already enabled)\n");
806 		return -EINVAL;
807 	}
808 
809 	if (WARN_ON_ONCE(dev->msix_enabled))
810 		return -EINVAL;
811 
812 	/* Check MSI-X early on irq domain enabled architectures */
813 	if (!pci_msi_domain_supports(dev, MSI_FLAG_PCI_MSIX, ALLOW_LEGACY))
814 		return -ENOTSUPP;
815 
816 	if (!pci_msi_supported(dev, nvec) || dev->current_state != PCI_D0)
817 		return -EINVAL;
818 
819 	hwsize = pci_msix_vec_count(dev);
820 	if (hwsize < 0)
821 		return hwsize;
822 
823 	if (!pci_msix_validate_entries(dev, entries, nvec))
824 		return -EINVAL;
825 
826 	if (hwsize < nvec) {
827 		/* Keep the IRQ virtual hackery working */
828 		if (flags & PCI_IRQ_VIRTUAL)
829 			hwsize = nvec;
830 		else
831 			nvec = hwsize;
832 	}
833 
834 	if (nvec < minvec)
835 		return -ENOSPC;
836 
837 	rc = pci_setup_msi_context(dev);
838 	if (rc)
839 		return rc;
840 
841 	if (!pci_setup_msix_device_domain(dev, hwsize))
842 		return -ENODEV;
843 
844 	for (;;) {
845 		if (affd) {
846 			nvec = irq_calc_affinity_vectors(minvec, nvec, affd);
847 			if (nvec < minvec)
848 				return -ENOSPC;
849 		}
850 
851 		rc = msix_capability_init(dev, entries, nvec, affd);
852 		if (rc == 0)
853 			return nvec;
854 
855 		if (rc < 0)
856 			return rc;
857 		if (rc < minvec)
858 			return -ENOSPC;
859 
860 		nvec = rc;
861 	}
862 }
863 
__pci_restore_msix_state(struct pci_dev * dev)864 void __pci_restore_msix_state(struct pci_dev *dev)
865 {
866 	struct msi_desc *entry;
867 	bool write_msg;
868 
869 	if (!dev->msix_enabled)
870 		return;
871 
872 	/* route the table */
873 	pci_intx_for_msi(dev, 0);
874 	pci_msix_clear_and_set_ctrl(dev, 0,
875 				PCI_MSIX_FLAGS_ENABLE | PCI_MSIX_FLAGS_MASKALL);
876 
877 	write_msg = arch_restore_msi_irqs(dev);
878 
879 	msi_lock_descs(&dev->dev);
880 	msi_for_each_desc(entry, &dev->dev, MSI_DESC_ALL) {
881 		if (write_msg)
882 			__pci_write_msi_msg(entry, &entry->msg);
883 		pci_msix_write_vector_ctrl(entry, entry->pci.msix_ctrl);
884 	}
885 	msi_unlock_descs(&dev->dev);
886 
887 	pci_msix_clear_and_set_ctrl(dev, PCI_MSIX_FLAGS_MASKALL, 0);
888 }
889 
pci_msix_shutdown(struct pci_dev * dev)890 void pci_msix_shutdown(struct pci_dev *dev)
891 {
892 	struct msi_desc *desc;
893 
894 	if (!pci_msi_enable || !dev || !dev->msix_enabled)
895 		return;
896 
897 	if (pci_dev_is_disconnected(dev)) {
898 		dev->msix_enabled = 0;
899 		return;
900 	}
901 
902 	/* Return the device with MSI-X masked as initial states */
903 	msi_for_each_desc(desc, &dev->dev, MSI_DESC_ALL)
904 		pci_msix_mask(desc);
905 
906 	pci_msix_clear_and_set_ctrl(dev, PCI_MSIX_FLAGS_ENABLE, 0);
907 	pci_intx_for_msi(dev, 1);
908 	dev->msix_enabled = 0;
909 	pcibios_alloc_irq(dev);
910 }
911 
912 /* Common interfaces */
913 
pci_free_msi_irqs(struct pci_dev * dev)914 void pci_free_msi_irqs(struct pci_dev *dev)
915 {
916 	pci_msi_teardown_msi_irqs(dev);
917 
918 	if (dev->msix_base) {
919 		iounmap(dev->msix_base);
920 		dev->msix_base = NULL;
921 	}
922 }
923 
924 /* Misc. infrastructure */
925 
msi_desc_to_pci_dev(struct msi_desc * desc)926 struct pci_dev *msi_desc_to_pci_dev(struct msi_desc *desc)
927 {
928 	return to_pci_dev(desc->dev);
929 }
930 EXPORT_SYMBOL(msi_desc_to_pci_dev);
931 
pci_no_msi(void)932 void pci_no_msi(void)
933 {
934 	pci_msi_enable = 0;
935 }
936