Searched refs:num_events (Results 1 – 15 of 15) sorted by relevance
/arch/sh/kernel/cpu/sh4a/ |
D | ubc.c | 48 for (i = 0; i < sh4a_ubc.num_events; i++) in sh4a_ubc_enable_all() 58 for (i = 0; i < sh4a_ubc.num_events; i++) in sh4a_ubc_disable_all() 68 for (i = 0; i < sh4a_ubc.num_events; i++) in sh4a_ubc_active_mask() 87 .num_events = 2, 114 for (i = 0; i < sh4a_ubc.num_events; i++) { in sh4a_ubc_init()
|
D | perf_event.c | 261 for (i = 0; i < sh4a_pmu.num_events; i++) in sh4a_pmu_disable_all() 269 for (i = 0; i < sh4a_pmu.num_events; i++) in sh4a_pmu_enable_all() 275 .num_events = 2,
|
/arch/sh/kernel/ |
D | hw_breakpoint.c | 35 static struct sh_ubc ubc_dummy = { .num_events = 0 }; 52 for (i = 0; i < sh_ubc->num_events; i++) { in arch_install_hw_breakpoint() 61 if (WARN_ONCE(i == sh_ubc->num_events, "Can't find any breakpoint slot")) in arch_install_hw_breakpoint() 84 for (i = 0; i < sh_ubc->num_events; i++) { in arch_uninstall_hw_breakpoint() 93 if (WARN_ONCE(i == sh_ubc->num_events, "Can't find any breakpoint slot")) in arch_uninstall_hw_breakpoint() 267 for (i = 0; i < sh_ubc->num_events; i++) { in flush_ptrace_hw_breakpoint() 297 for (i = 0; i < sh_ubc->num_events; i++) { in hw_breakpoint_handler() 405 WARN_ON(ubc->num_events > HBP_NUM); in register_sh_ubc()
|
D | perf_event.c | 39 static atomic_t num_events; variable 65 if (!atomic_add_unless(&num_events, -1, 1)) { in hw_perf_event_destroy() 67 if (atomic_dec_return(&num_events) == 0) in hw_perf_event_destroy() 118 if (!atomic_inc_not_zero(&num_events)) { in __hw_perf_event_init() 120 if (atomic_read(&num_events) == 0 && in __hw_perf_event_init() 124 atomic_inc(&num_events); in __hw_perf_event_init() 254 idx = find_first_zero_bit(cpuc->used_mask, sh_pmu->num_events); in sh_pmu_add() 255 if (idx == sh_pmu->num_events) in sh_pmu_add() 357 WARN_ON(_pmu->num_events > MAX_HWEVENTS); in register_sh_pmu()
|
/arch/powerpc/perf/ |
D | e500-pmu.c | 83 static int num_events = 128; variable 91 if (event_low >= num_events) in e500_xlate_event() 125 num_events = 256; in init_e500_pmu()
|
D | e6500-pmu.c | 85 static int num_events = 512; variable 91 if (event_low >= num_events || in e6500_xlate_event()
|
D | core-fsl-emb.c | 30 static atomic_t num_events; variable 209 if (atomic_read(&num_events)) { in fsl_emb_pmu_disable() 431 if (!atomic_add_unless(&num_events, -1, 1)) { in hw_perf_event_destroy() 433 if (atomic_dec_return(&num_events) == 0) in hw_perf_event_destroy() 556 if (!atomic_inc_not_zero(&num_events)) { in fsl_emb_pmu_event_init() 558 if (atomic_read(&num_events) == 0 && in fsl_emb_pmu_event_init() 562 atomic_inc(&num_events); in fsl_emb_pmu_event_init()
|
D | core-book3s.c | 1944 static atomic_t num_events; variable 1953 if (!atomic_add_unless(&num_events, -1, 1)) { in hw_perf_event_destroy() 1955 if (atomic_dec_return(&num_events) == 0) in hw_perf_event_destroy() 2184 if (!atomic_inc_not_zero(&num_events)) { in power_pmu_event_init() 2186 if (atomic_read(&num_events) == 0 && in power_pmu_event_init() 2190 atomic_inc(&num_events); in power_pmu_event_init()
|
/arch/sh/kernel/cpu/sh4/ |
D | perf_event.c | 227 for (i = 0; i < sh7750_pmu.num_events; i++) in sh7750_pmu_disable_all() 235 for (i = 0; i < sh7750_pmu.num_events; i++) in sh7750_pmu_enable_all() 241 .num_events = 2,
|
/arch/sh/include/asm/ |
D | perf_event.h | 11 unsigned int num_events; member
|
D | hw_breakpoint.h | 31 unsigned int num_events; member
|
/arch/arm/kernel/ |
D | perf_event_xscale.c | 173 for (idx = 0; idx < cpu_pmu->num_events; ++idx) { in xscale1pmu_handle_irq() 383 cpu_pmu->num_events = 3; in xscale1pmu_init() 519 for (idx = 0; idx < cpu_pmu->num_events; ++idx) { in xscale2pmu_handle_irq() 752 cpu_pmu->num_events = 5; in xscale2pmu_init()
|
D | perf_event_v6.c | 326 for (idx = 0; idx < cpu_pmu->num_events; ++idx) { in armv6pmu_handle_irq() 504 cpu_pmu->num_events = 3; in armv6pmu_init() 555 cpu_pmu->num_events = 3; in armv6mpcore_pmu_init()
|
D | perf_event_v7.c | 657 (ARMV7_IDX_CYCLE_COUNTER + cpu_pmu->num_events - 1) 973 for (idx = 0; idx < cpu_pmu->num_events; ++idx) { in armv7pmu_handle_irq() 1052 for (idx = ARMV7_IDX_COUNTER0; idx < cpu_pmu->num_events; ++idx) { in armv7pmu_get_event_idx() 1096 u32 idx, nb_cnt = cpu_pmu->num_events, val; in armv7pmu_reset() 1197 &arm_pmu->num_events, 1); in armv7_probe_num_events() 1559 u32 idx, nb_cnt = cpu_pmu->num_events; in krait_pmu_reset() 1892 u32 idx, nb_cnt = cpu_pmu->num_events; in scorpion_pmu_reset()
|
/arch/s390/kernel/ |
D | perf_cpum_sf.c | 588 static atomic_t num_events; variable 647 if (!atomic_add_unless(&num_events, -1, 1)) { in hw_perf_event_destroy() 649 if (atomic_dec_return(&num_events) == 0) in hw_perf_event_destroy() 817 if (!atomic_inc_not_zero(&num_events)) { in __hw_perf_event_init() 819 if (atomic_read(&num_events) == 0 && reserve_pmc_hardware()) in __hw_perf_event_init() 822 atomic_inc(&num_events); in __hw_perf_event_init() 2146 if (!atomic_read(&num_events)) in cpusf_pmu_setup()
|