/arch/m68k/mac/ |
D | oss.c | 69 int events = oss->irq_pending & in oss_irq() local 73 if ((console_loglevel == 10) && !(events & OSS_IP_SCSI)) { in oss_irq() 79 if (events & OSS_IP_IOPSCC) { in oss_irq() 84 if (events & OSS_IP_SCSI) { in oss_irq() 89 if (events & OSS_IP_IOPISM) { in oss_irq() 103 int events, irq_bit, i; in oss_nubus_irq() local 105 events = oss->irq_pending & OSS_IP_NUBUS; in oss_nubus_irq() 106 if (!events) in oss_nubus_irq() 111 printk("oss_nubus_irq: events = 0x%04X\n", events); in oss_nubus_irq() 121 if (events & irq_bit) { in oss_nubus_irq() [all …]
|
D | via.c | 452 unsigned char irq_bit, events; in via1_irq() local 454 events = via1[vIFR] & via1[vIER] & 0x7F; in via1_irq() 455 if (!events) in via1_irq() 461 if (events & irq_bit) { in via1_irq() 467 } while (events >= irq_bit); in via1_irq() 473 unsigned char irq_bit, events; in via2_irq() local 475 events = via2[gIFR] & via2[gIER] & 0x7F; in via2_irq() 476 if (!events) in via2_irq() 482 if (events & irq_bit) { in via2_irq() 488 } while (events >= irq_bit); in via2_irq() [all …]
|
D | baboon.c | 51 unsigned char events; in baboon_irq() local 59 events = baboon->mb_ifr & 0x07; in baboon_irq() 60 if (!events) in baboon_irq() 66 if (events & irq_bit) { in baboon_irq() 72 } while(events >= irq_bit); in baboon_irq() 76 baboon->mb_ifr &= ~events; in baboon_irq()
|
D | psc.c | 122 unsigned char irq_bit, events; in psc_irq() local 129 events = psc_read_byte(pIFR) & psc_read_byte(pIER) & 0xF; in psc_irq() 130 if (!events) in psc_irq() 136 if (events & irq_bit) { in psc_irq() 142 } while (events >= irq_bit); in psc_irq()
|
/arch/arm/kernel/ |
D | perf_event_xscale.c | 166 struct perf_event *event = cpuc->events[idx]; in xscale1pmu_handle_irq() 201 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); in xscale1pmu_enable_event() local 224 raw_spin_lock_irqsave(&events->pmu_lock, flags); in xscale1pmu_enable_event() 229 raw_spin_unlock_irqrestore(&events->pmu_lock, flags); in xscale1pmu_enable_event() 237 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); in xscale1pmu_disable_event() local 258 raw_spin_lock_irqsave(&events->pmu_lock, flags); in xscale1pmu_disable_event() 263 raw_spin_unlock_irqrestore(&events->pmu_lock, flags); in xscale1pmu_disable_event() 290 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); in xscale1pmu_start() local 292 raw_spin_lock_irqsave(&events->pmu_lock, flags); in xscale1pmu_start() 296 raw_spin_unlock_irqrestore(&events->pmu_lock, flags); in xscale1pmu_start() [all …]
|
D | perf_event_v6.c | 265 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); in armv6pmu_enable_event() local 288 raw_spin_lock_irqsave(&events->pmu_lock, flags); in armv6pmu_enable_event() 293 raw_spin_unlock_irqrestore(&events->pmu_lock, flags); in armv6pmu_enable_event() 320 struct perf_event *event = cpuc->events[idx]; in armv6pmu_handle_irq() 359 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); in armv6pmu_start() local 361 raw_spin_lock_irqsave(&events->pmu_lock, flags); in armv6pmu_start() 365 raw_spin_unlock_irqrestore(&events->pmu_lock, flags); in armv6pmu_start() 371 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); in armv6pmu_stop() local 373 raw_spin_lock_irqsave(&events->pmu_lock, flags); in armv6pmu_stop() 377 raw_spin_unlock_irqrestore(&events->pmu_lock, flags); in armv6pmu_stop() [all …]
|
D | sys_oabi-compat.c | 249 __u32 events; member 265 kernel.events = user.events; in sys_oabi_epoll_ctl() 275 struct oabi_epoll_event __user *events, in sys_oabi_epoll_wait() argument 293 __put_user_error(kbuf[i].events, &events->events, err); in sys_oabi_epoll_wait() 294 __put_user_error(kbuf[i].data, &events->data, err); in sys_oabi_epoll_wait() 295 events++; in sys_oabi_epoll_wait()
|
D | perf_event_cpu.c | 174 struct pmu_hw_events *events = &per_cpu(cpu_hw_events, cpu); in cpu_pmu_init() local 175 events->events = per_cpu(hw_events, cpu); in cpu_pmu_init() 176 events->used_mask = per_cpu(used_mask, cpu); in cpu_pmu_init() 177 raw_spin_lock_init(&events->pmu_lock); in cpu_pmu_init()
|
D | perf_event_v7.c | 708 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); in armv7pmu_enable_event() local 721 raw_spin_lock_irqsave(&events->pmu_lock, flags); in armv7pmu_enable_event() 746 raw_spin_unlock_irqrestore(&events->pmu_lock, flags); in armv7pmu_enable_event() 754 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); in armv7pmu_disable_event() local 766 raw_spin_lock_irqsave(&events->pmu_lock, flags); in armv7pmu_disable_event() 778 raw_spin_unlock_irqrestore(&events->pmu_lock, flags); in armv7pmu_disable_event() 807 struct perf_event *event = cpuc->events[idx]; in armv7pmu_handle_irq() 846 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); in armv7pmu_start() local 848 raw_spin_lock_irqsave(&events->pmu_lock, flags); in armv7pmu_start() 851 raw_spin_unlock_irqrestore(&events->pmu_lock, flags); in armv7pmu_start() [all …]
|
/arch/um/os-Linux/ |
D | irq.c | 52 int os_create_pollfd(int fd, int events, void *tmp_pfd, int size_tmpfds) in os_create_pollfd() argument 72 .events = events, in os_create_pollfd()
|
/arch/arm64/kernel/ |
D | perf_event.c | 274 hw_events->events[idx] = NULL; in armpmu_del() 304 hw_events->events[idx] = event; in armpmu_add() 1026 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); in armv8pmu_enable_event() local 1032 raw_spin_lock_irqsave(&events->pmu_lock, flags); in armv8pmu_enable_event() 1054 raw_spin_unlock_irqrestore(&events->pmu_lock, flags); in armv8pmu_enable_event() 1060 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); in armv8pmu_disable_event() local 1065 raw_spin_lock_irqsave(&events->pmu_lock, flags); in armv8pmu_disable_event() 1077 raw_spin_unlock_irqrestore(&events->pmu_lock, flags); in armv8pmu_disable_event() 1106 struct perf_event *event = cpuc->events[idx]; in armv8pmu_handle_irq() 1145 struct pmu_hw_events *events = cpu_pmu->get_hw_events(); in armv8pmu_start() local [all …]
|
D | trace-events-emulation.h | 34 #define TRACE_INCLUDE_FILE trace-events-emulation
|
/arch/metag/kernel/perf/ |
D | perf_event.c | 288 cpuc->events[idx] = event; in metag_pmu_start() 357 cpuc->events[idx] = NULL; in metag_pmu_del() 600 struct cpu_hw_events *events = this_cpu_ptr(&cpu_hw_events); in metag_pmu_enable_counter() local 605 raw_spin_lock_irqsave(&events->pmu_lock, flags); in metag_pmu_enable_counter() 668 raw_spin_unlock_irqrestore(&events->pmu_lock, flags); in metag_pmu_enable_counter() 673 struct cpu_hw_events *events = this_cpu_ptr(&cpu_hw_events); in metag_pmu_disable_counter() local 696 raw_spin_lock_irqsave(&events->pmu_lock, flags); in metag_pmu_disable_counter() 702 raw_spin_unlock_irqrestore(&events->pmu_lock, flags); in metag_pmu_disable_counter() 721 struct cpu_hw_events *events = this_cpu_ptr(&cpu_hw_events); in metag_pmu_write_counter() local 736 raw_spin_lock_irqsave(&events->pmu_lock, flags); in metag_pmu_write_counter() [all …]
|
D | perf_event.h | 51 struct perf_event *events[MAX_HWEVENTS]; member
|
/arch/um/kernel/ |
D | irq.c | 68 int events, err, n; in activate_fd() local 80 events = UM_POLLIN | UM_POLLPRI; in activate_fd() 81 else events = UM_POLLOUT; in activate_fd() 87 .events = events, in activate_fd() 109 n = os_create_pollfd(fd, events, tmp_pfd, n); in activate_fd()
|
/arch/powerpc/platforms/powernv/ |
D | opal.c | 251 static void opal_do_notifier(uint64_t events) in opal_do_notifier() argument 260 changed_mask = last_notified_mask ^ events; in opal_do_notifier() 261 last_notified_mask = events; in opal_do_notifier() 269 events, (void *)changed_mask); in opal_do_notifier() 362 unsigned long events, void *change) in opal_message_notify() argument 364 if (events & OPAL_EVENT_MSG_PENDING) in opal_message_notify() 609 __be64 events; in opal_interrupt() local 611 opal_handle_interrupt(virq_to_hw(irq), &events); in opal_interrupt() 613 opal_do_notifier(be64_to_cpu(events)); in opal_interrupt()
|
/arch/powerpc/platforms/pseries/ |
D | Kconfig | 63 to return information about hardware error and non-error events 64 which may need OS attention. RTAS returns events for multiple 66 to receive events. 116 bool "Hypervisor supplied PMU events (24x7 & GPCI)" 131 SPLPAR machines can log hypervisor preempt & dispatch events to a 132 kernel buffer. Saying Y here will enable logging these events,
|
/arch/sparc/kernel/ |
D | sparc_ksyms_32.c | 18 short events; member
|
D | sparc_ksyms_64.c | 21 short events; member
|
D | perf_event.c | 96 unsigned long events[MAX_HWEVENTS]; member 826 enc = perf_event_get_enc(cpuc->events[idx]); in sparc_pmu_enable_event() 952 enc = perf_event_get_enc(cpuc->events[i]); in calculate_single_pcr() 1112 cpuc->events[i - 1] = cpuc->events[i]; in sparc_pmu_del() 1221 unsigned long *events, int n_ev) in sparc_check_constraints() argument 1243 msk0 = perf_event_get_msk(events[0]); in sparc_check_constraints() 1250 msk1 = perf_event_get_msk(events[1]); in sparc_check_constraints() 1324 struct perf_event *evts[], unsigned long *events, in collect_events() argument 1334 events[n] = group->hw.event_base; in collect_events() 1343 events[n] = event->hw.event_base; in collect_events() [all …]
|
/arch/um/include/shared/ |
D | irq_user.h | 17 int events; member
|
/arch/powerpc/perf/ |
D | core-book3s.c | 37 u64 events[MAX_HWEVENTS]; member 1262 if (ppmu->compute_mmcr(cpuhw->events, cpuhw->n_events, hwc_index, in power_pmu_enable() 1373 struct perf_event *ctrs[], u64 *events, in collect_events() argument 1384 events[n++] = group->hw.config; in collect_events() 1393 events[n++] = event->hw.config; in collect_events() 1424 cpuhw->events[n0] = event->hw.config; in power_pmu_add() 1448 if (power_check_constraints(cpuhw, cpuhw->events, cpuhw->flags, n0 + 1)) in power_pmu_add() 1450 event->hw.config = cpuhw->events[n0]; in power_pmu_add() 1490 cpuhw->events[i-1] = cpuhw->events[i]; in power_pmu_del() 1627 i = power_check_constraints(cpuhw, cpuhw->events, cpuhw->flags, n); in power_pmu_commit_txn() [all …]
|
/arch/powerpc/include/asm/ |
D | fsl_gtm.h | 45 extern void gtm_ack_timer16(struct gtm_timer *tmr, u16 events);
|
/arch/arm/mach-omap2/ |
D | prm44xx_54xx.h | 55 extern void omap44xx_prm_read_pending_irqs(unsigned long *events);
|
/arch/blackfin/kernel/ |
D | perf_event.c | 231 struct perf_event *events[MAX_HWEVENTS]; member 309 cpuc->events[idx] = NULL; in bfin_pmu_stop() 331 cpuc->events[idx] = event; in bfin_pmu_start() 430 event = cpuc->events[i]; in bfin_pmu_enable()
|