| /kernel/linux/linux-6.6/arch/x86/kernel/apic/ |
| D | init.c | 17 DEFINE_APIC_CALL(eoi); 45 apply_override(eoi); in restore_override_callbacks() 66 update_call(eoi); in update_static_calls() 86 apic->native_eoi = apic->eoi; in apic_setup_apic_calls() 101 /* Copy the original eoi() callback as KVM/HyperV might overwrite it */ in apic_install_driver() 103 apic->native_eoi = apic->eoi; in apic_install_driver()
|
| /kernel/linux/linux-5.10/arch/powerpc/kvm/ |
| D | book3s_xive_template.c | 77 /* If the XIVE supports the new "store EOI facility, use it */ in GLUE() 84 * For LSIs the HW EOI cycle is used rather than PQ bits, in GLUE() 93 * Otherwise for EOI, we use the special MMIO that does in GLUE() 95 * except for LSIs where we use the "EOI cycle" special in GLUE() 148 * Snapshot the queue page. The test further down for EOI in GLUE() 151 * to miss an EOI. in GLUE() 165 * we EOI it now, thus re-enabling reception of a new in GLUE() 240 * If this is an EOI that's it, no CPPR adjustment done here, in GLUE() 427 /* If it's not an LSI, set PQ to 11 the EOI will force a resend */ in GLUE() 431 /* EOI the source */ in GLUE() [all …]
|
| /kernel/linux/linux-6.6/arch/powerpc/sysdev/xics/ |
| D | icp-opal.c | 33 * We take the ipi irq but and never return so we need to EOI the IPI, in icp_opal_flush_ipi() 82 /* We might learn about it later, so EOI it */ in icp_opal_get_irq() 115 * EOI tells us whether there are more interrupts to fetch. in icp_opal_eoi() 169 /* EOI the interrupt */ in icp_opal_flush_interrupt() 177 .eoi = icp_opal_eoi,
|
| D | icp-hv.c | 53 pr_err("%s: bad return code eoi xirr=0x%x returned %ld\n", in icp_hv_set_xirr() 94 * need to EOI the IPI, but want to leave our priority 0 in icp_hv_flush_ipi() 122 /* We might learn about it later, so EOI it */ in icp_hv_get_irq() 155 .eoi = icp_hv_eoi,
|
| /kernel/linux/linux-5.10/arch/powerpc/sysdev/xics/ |
| D | icp-opal.c | 32 * We take the ipi irq but and never return so we need to EOI the IPI, in icp_opal_flush_ipi() 81 /* We might learn about it later, so EOI it */ in icp_opal_get_irq() 114 * EOI tells us whether there are more interrupts to fetch. in icp_opal_eoi() 168 /* EOI the interrupt */ in icp_opal_flush_interrupt() 176 .eoi = icp_opal_eoi,
|
| D | icp-hv.c | 52 pr_err("%s: bad return code eoi xirr=0x%x returned %ld\n", in icp_hv_set_xirr() 93 * need to EOI the IPI, but want to leave our priority 0 in icp_hv_flush_ipi() 121 /* We might learn about it later, so EOI it */ in icp_hv_get_irq() 154 .eoi = icp_hv_eoi,
|
| /kernel/linux/linux-6.6/tools/testing/selftests/kvm/include/aarch64/ |
| D | gic.h | 33 * Sets the EOI mode. When split is false, EOI just drops the priority. When 34 * split is true, EOI drops the priority and deactivates the interrupt.
|
| /kernel/linux/linux-6.6/Documentation/ABI/testing/ |
| D | sysfs-devices-xenbus | 19 Summed up time in jiffies the EOI of an interrupt for a Xen 30 trigger delayed EOI processing. 37 before delayed EOI processing is triggered for a Xen pv
|
| /kernel/linux/linux-5.10/drivers/xen/events/ |
| D | events_base.c | 110 unsigned short eoi_cpu; /* EOI must happen on this cpu-1 */ 112 u64 eoi_time; /* Time in jiffies when to EOI. */ 531 struct lateeoi_work *eoi = &per_cpu(lateeoi, info->eoi_cpu); in lateeoi_list_del() local 534 spin_lock_irqsave(&eoi->eoi_list_lock, flags); in lateeoi_list_del() 536 spin_unlock_irqrestore(&eoi->eoi_list_lock, flags); in lateeoi_list_del() 541 struct lateeoi_work *eoi = &per_cpu(lateeoi, info->eoi_cpu); in lateeoi_list_add() local 552 spin_lock_irqsave(&eoi->eoi_list_lock, flags); in lateeoi_list_add() 554 elem = list_first_entry_or_null(&eoi->eoi_list, struct irq_info, in lateeoi_list_add() 557 list_add(&info->eoi_list, &eoi->eoi_list); in lateeoi_list_add() 559 &eoi->delayed, delay); in lateeoi_list_add() [all …]
|
| /kernel/linux/linux-5.10/drivers/xen/xen-pciback/ |
| D | pciback_ops.c | 295 bool eoi = true; in xen_pcibk_test_and_schedule_op() local 301 eoi = false; in xen_pcibk_test_and_schedule_op() 308 eoi = false; in xen_pcibk_test_and_schedule_op() 311 /* EOI if there was nothing to do. */ in xen_pcibk_test_and_schedule_op() 312 if (eoi) in xen_pcibk_test_and_schedule_op() 412 bool eoi; in xen_pcibk_handle_event() local 418 eoi = test_and_set_bit(_EOI_pending, &pdev->flags); in xen_pcibk_handle_event() 419 WARN(eoi, "IRQ while EOI pending\n"); in xen_pcibk_handle_event()
|
| /kernel/linux/linux-6.6/drivers/xen/xen-pciback/ |
| D | pciback_ops.c | 295 bool eoi = true; in xen_pcibk_test_and_schedule_op() local 301 eoi = false; in xen_pcibk_test_and_schedule_op() 308 eoi = false; in xen_pcibk_test_and_schedule_op() 311 /* EOI if there was nothing to do. */ in xen_pcibk_test_and_schedule_op() 312 if (eoi) in xen_pcibk_test_and_schedule_op() 412 bool eoi; in xen_pcibk_handle_event() local 418 eoi = test_and_set_bit(_EOI_pending, &pdev->flags); in xen_pcibk_handle_event() 419 WARN(eoi, "IRQ while EOI pending\n"); in xen_pcibk_handle_event()
|
| /kernel/linux/linux-6.6/drivers/irqchip/ |
| D | irq-i8259.c | 131 * first, _then_ send the EOI, and the order of EOI 164 outb(0x60+(irq&7), PIC_SLAVE_CMD);/* 'Specific EOI' to slave */ in mask_and_ack_8259A() 165 outb(0x60+PIC_CASCADE_IR, PIC_MASTER_CMD); /* 'Specific EOI' to master-IRQ2 */ in mask_and_ack_8259A() 169 outb(0x60+irq, PIC_MASTER_CMD); /* 'Specific EOI to master */ in mask_and_ack_8259A() 245 if (auto_eoi) /* master does Auto EOI */ in init_8259A() 247 else /* master expects normal EOI */ in init_8259A()
|
| D | irq-clps711x.c | 40 phys_addr_t eoi; member 95 writel_relaxed(0, clps711x_intc->base + clps711x_irqs[hwirq].eoi); in clps711x_intc_eoi() 139 } else if (clps711x_irqs[hw].eoi) { in clps711x_intc_irq_map() 144 if (clps711x_irqs[hw].eoi) in clps711x_intc_irq_map() 145 writel_relaxed(0, clps711x_intc->base + clps711x_irqs[hw].eoi); in clps711x_intc_irq_map()
|
| /kernel/linux/linux-5.10/drivers/irqchip/ |
| D | irq-i8259.c | 131 * first, _then_ send the EOI, and the order of EOI 164 outb(0x60+(irq&7), PIC_SLAVE_CMD);/* 'Specific EOI' to slave */ in mask_and_ack_8259A() 165 outb(0x60+PIC_CASCADE_IR, PIC_MASTER_CMD); /* 'Specific EOI' to master-IRQ2 */ in mask_and_ack_8259A() 169 outb(0x60+irq, PIC_MASTER_CMD); /* 'Specific EOI to master */ in mask_and_ack_8259A() 245 if (auto_eoi) /* master does Auto EOI */ in init_8259A() 247 else /* master expects normal EOI */ in init_8259A()
|
| D | irq-clps711x.c | 40 phys_addr_t eoi; member 95 writel_relaxed(0, clps711x_intc->base + clps711x_irqs[hwirq].eoi); in clps711x_intc_eoi() 139 } else if (clps711x_irqs[hw].eoi) { in clps711x_intc_irq_map() 144 if (clps711x_irqs[hw].eoi) in clps711x_intc_irq_map() 145 writel_relaxed(0, clps711x_intc->base + clps711x_irqs[hw].eoi); in clps711x_intc_irq_map()
|
| /kernel/linux/linux-6.6/arch/x86/kvm/ |
| D | ioapic.c | 198 * If no longer has pending EOI in LAPICs, update in ioapic_lazy_update_eoi() 199 * EOI for this vector. in ioapic_lazy_update_eoi() 224 * AMD SVM AVIC accelerate EOI write iff the interrupt is edge in ioapic_set_irq() 226 * to receive the EOI. In this case, we do a lazy update of the in ioapic_set_irq() 227 * pending EOI when trying to set IOAPIC irq. in ioapic_set_irq() 236 * us if the interrupt is waiting for an EOI. in ioapic_set_irq() 239 * if it has been already ack-ed via EOI because coalesced RTC in ioapic_set_irq() 241 * EOI manually for the RTC interrupt. in ioapic_set_irq() 287 /* Make sure we see any missing RTC EOI */ in kvm_ioapic_scan_entry() 363 * explicit EOI on IOAPICs that don't have the EOI register. in ioapic_write_indirect() [all …]
|
| /kernel/linux/linux-6.6/drivers/xen/events/ |
| D | events_base.c | 113 unsigned short eoi_cpu; /* EOI must happen on this cpu-1 */ 115 u64 eoi_time; /* Time in jiffies when to EOI. */ 561 struct lateeoi_work *eoi = &per_cpu(lateeoi, info->eoi_cpu); in lateeoi_list_del() local 564 spin_lock_irqsave(&eoi->eoi_list_lock, flags); in lateeoi_list_del() 566 spin_unlock_irqrestore(&eoi->eoi_list_lock, flags); in lateeoi_list_del() 571 struct lateeoi_work *eoi = &per_cpu(lateeoi, info->eoi_cpu); in lateeoi_list_add() local 582 spin_lock_irqsave(&eoi->eoi_list_lock, flags); in lateeoi_list_add() 584 elem = list_first_entry_or_null(&eoi->eoi_list, struct irq_info, in lateeoi_list_add() 587 list_add(&info->eoi_list, &eoi->eoi_list); in lateeoi_list_add() 589 &eoi->delayed, delay); in lateeoi_list_add() [all …]
|
| /kernel/linux/linux-5.10/arch/x86/kvm/ |
| D | ioapic.c | 199 * If no longer has pending EOI in LAPICs, update in ioapic_lazy_update_eoi() 200 * EOI for this vector. in ioapic_lazy_update_eoi() 225 * AMD SVM AVIC accelerate EOI write iff the interrupt is edge in ioapic_set_irq() 227 * to receive the EOI. In this case, we do a lazy update of the in ioapic_set_irq() 228 * pending EOI when trying to set IOAPIC irq. in ioapic_set_irq() 237 * us if the interrupt is waiting for an EOI. in ioapic_set_irq() 240 * if it has been already ack-ed via EOI because coalesced RTC in ioapic_set_irq() 242 * EOI manually for the RTC interrupt. in ioapic_set_irq() 288 /* Make sure we see any missing RTC EOI */ in kvm_ioapic_scan_entry() 364 * explicit EOI on IOAPICs that don't have the EOI register. in ioapic_write_indirect() [all …]
|
| /kernel/linux/linux-5.10/arch/powerpc/sysdev/xive/ |
| D | common.c | 315 * as a "replay" because EOI decided there was still something in xive_get_irq() 323 * entry (on HW interrupt) from a replay triggered by EOI, in xive_get_irq() 344 * After EOI'ing an interrupt, we need to re-check the queue 356 DBG_VERBOSE("eoi: pending=0x%02x\n", xc->pending_prio); in xive_do_queue_eoi() 362 * EOI an interrupt at the source. There are several methods 368 /* If the XIVE supports the new "store EOI facility, use it */ in xive_do_source_eoi() 379 if (WARN_ON_ONCE(!xive_ops->eoi)) in xive_do_source_eoi() 381 xive_ops->eoi(hw_irq); in xive_do_source_eoi() 386 * Otherwise for EOI, we use the special MMIO that does in xive_do_source_eoi() 388 * except for LSIs where we use the "EOI cycle" special in xive_do_source_eoi() [all …]
|
| /kernel/linux/linux-5.10/arch/x86/kernel/ |
| D | i8259.c | 149 * first, _then_ send the EOI, and the order of EOI 182 /* 'Specific EOI' to slave */ in mask_and_ack_8259A() 184 /* 'Specific EOI' to master-IRQ2 */ in mask_and_ack_8259A() 189 outb(0x60+irq, PIC_MASTER_CMD); /* 'Specific EOI to master */ in mask_and_ack_8259A() 367 if (auto_eoi) /* master does Auto EOI */ in init_8259A() 369 else /* master expects normal EOI */ in init_8259A()
|
| /kernel/linux/linux-6.6/arch/x86/kernel/ |
| D | i8259.c | 149 * first, _then_ send the EOI, and the order of EOI 182 /* 'Specific EOI' to slave */ in mask_and_ack_8259A() 184 /* 'Specific EOI' to master-IRQ2 */ in mask_and_ack_8259A() 189 outb(0x60+irq, PIC_MASTER_CMD); /* 'Specific EOI to master */ in mask_and_ack_8259A() 365 if (auto_eoi) /* master does Auto EOI */ in init_8259A() 367 else /* master expects normal EOI */ in init_8259A()
|
| /kernel/linux/linux-5.10/Documentation/virt/kvm/ |
| D | msr.rst | 319 injection. Value of 1 means that guest can skip writing EOI to the apic 321 EOI by clearing the bit in guest memory - this location will 323 Value of 0 means that the EOI write is required. 326 the APIC EOI write anyway. 337 whether it can skip EOI apic write and between guest 338 clearing it to signal EOI to the hypervisor,
|
| /kernel/linux/linux-6.6/Documentation/virt/kvm/x86/ |
| D | msr.rst | 319 injection. Value of 1 means that guest can skip writing EOI to the apic 321 EOI by clearing the bit in guest memory - this location will 323 Value of 0 means that the EOI write is required. 326 the APIC EOI write anyway. 337 whether it can skip EOI apic write and between guest 338 clearing it to signal EOI to the hypervisor,
|
| /kernel/linux/linux-5.10/arch/x86/kernel/cpu/ |
| D | acrn.c | 43 * The hypervisor requires that the APIC EOI should be acked. in DEFINE_IDTENTRY_SYSVEC() 44 * If the APIC EOI is not acked, the APIC ISR bit for the in DEFINE_IDTENTRY_SYSVEC()
|
| /kernel/linux/linux-6.6/Documentation/devicetree/bindings/interrupt-controller/ |
| D | mstar,mst-intc.yaml | 42 mstar,intc-no-eoi: 44 Mark this controller has no End Of Interrupt(EOI) implementation.
|