/arch/sh/kernel/ |
D | perf_event.c | 30 unsigned long used_mask[BITS_TO_LONGS(MAX_HWEVENTS)]; member 239 __clear_bit(event->hw.idx, cpuc->used_mask); in sh_pmu_del() 253 if (__test_and_set_bit(idx, cpuc->used_mask)) { in sh_pmu_add() 254 idx = find_first_zero_bit(cpuc->used_mask, sh_pmu->num_events); in sh_pmu_add() 258 __set_bit(idx, cpuc->used_mask); in sh_pmu_add()
|
/arch/xtensa/kernel/ |
D | perf_event.c | 63 unsigned long used_mask[BITS_TO_LONGS(XCHAL_NUM_PERF_COUNTERS)]; member 296 if (__test_and_set_bit(idx, ev->used_mask)) { in xtensa_pmu_add() 297 idx = find_first_zero_bit(ev->used_mask, in xtensa_pmu_add() 302 __set_bit(idx, ev->used_mask); in xtensa_pmu_add() 321 __clear_bit(event->hw.idx, ev->used_mask); in xtensa_pmu_del() 374 for_each_set_bit(i, ev->used_mask, XCHAL_NUM_PERF_COUNTERS) { in xtensa_pmu_irq_handler()
|
/arch/arm/kernel/ |
D | perf_event_xscale.c | 280 if (test_and_set_bit(XSCALE_CYCLE_COUNTER, cpuc->used_mask)) in xscale1pmu_get_event_idx() 285 if (!test_and_set_bit(XSCALE_COUNTER1, cpuc->used_mask)) in xscale1pmu_get_event_idx() 288 if (!test_and_set_bit(XSCALE_COUNTER0, cpuc->used_mask)) in xscale1pmu_get_event_idx() 298 clear_bit(event->hw.idx, cpuc->used_mask); in xscalepmu_clear_event_idx() 656 if (!test_and_set_bit(XSCALE_COUNTER3, cpuc->used_mask)) in xscale2pmu_get_event_idx() 658 else if (!test_and_set_bit(XSCALE_COUNTER2, cpuc->used_mask)) in xscale2pmu_get_event_idx()
|
D | perf_event_v6.c | 394 if (test_and_set_bit(ARMV6_CYCLE_COUNTER, cpuc->used_mask)) in armv6pmu_get_event_idx() 403 if (!test_and_set_bit(ARMV6_COUNTER1, cpuc->used_mask)) in armv6pmu_get_event_idx() 406 if (!test_and_set_bit(ARMV6_COUNTER0, cpuc->used_mask)) in armv6pmu_get_event_idx() 417 clear_bit(event->hw.idx, cpuc->used_mask); in armv6pmu_clear_event_idx()
|
D | perf_event_v7.c | 1042 if (test_and_set_bit(ARMV7_IDX_CYCLE_COUNTER, cpuc->used_mask)) in armv7pmu_get_event_idx() 1053 if (!test_and_set_bit(idx, cpuc->used_mask)) in armv7pmu_get_event_idx() 1064 clear_bit(event->hw.idx, cpuc->used_mask); in armv7pmu_clear_event_idx() 1626 if (test_and_set_bit(bit, cpuc->used_mask)) in krait_pmu_get_event_idx() 1632 clear_bit(bit, cpuc->used_mask); in krait_pmu_get_event_idx() 1650 clear_bit(bit, cpuc->used_mask); in krait_pmu_clear_event_idx() 1956 if (test_and_set_bit(bit, cpuc->used_mask)) in scorpion_pmu_get_event_idx() 1962 clear_bit(bit, cpuc->used_mask); in scorpion_pmu_get_event_idx() 1980 clear_bit(bit, cpuc->used_mask); in scorpion_pmu_clear_event_idx()
|
/arch/nds32/kernel/ |
D | perf_event_cpu.c | 574 if (!test_and_set_bit(idx, cpuc->used_mask)) in nds32_pmu_get_event_idx() 576 if (!test_and_set_bit(NDS32_IDX_COUNTER0, cpuc->used_mask)) in nds32_pmu_get_event_idx() 578 if (!test_and_set_bit(NDS32_IDX_COUNTER1, cpuc->used_mask)) in nds32_pmu_get_event_idx() 581 if (!test_and_set_bit(idx, cpuc->used_mask)) in nds32_pmu_get_event_idx() 583 else if (!test_and_set_bit(NDS32_IDX_COUNTER1, cpuc->used_mask)) in nds32_pmu_get_event_idx() 586 (NDS32_IDX_CYCLE_COUNTER, cpuc->used_mask)) in nds32_pmu_get_event_idx() 589 if (!test_and_set_bit(idx, cpuc->used_mask)) in nds32_pmu_get_event_idx() 698 int enabled = bitmap_weight(hw_events->used_mask, in nds32_pmu_enable() 993 clear_bit(idx, hw_events->used_mask); in nds32_pmu_del()
|
/arch/arc/kernel/ |
D | perf_event.c | 49 unsigned long used_mask[BITS_TO_LONGS(ARC_PERF_MAX_COUNTERS)]; member 352 __clear_bit(event->hw.idx, pmu_cpu->used_mask); in arc_pmu_del() 366 idx = ffz(pmu_cpu->used_mask[0]); in arc_pmu_add() 370 __set_bit(idx, pmu_cpu->used_mask); in arc_pmu_add()
|
/arch/arm64/kernel/ |
D | perf_event.c | 850 if (!test_and_set_bit(idx, cpuc->used_mask)) in armv8pmu_get_single_idx() 866 if (!test_and_set_bit(idx, cpuc->used_mask)) { in armv8pmu_get_chain_idx() 868 if (!test_and_set_bit(idx - 1, cpuc->used_mask)) in armv8pmu_get_chain_idx() 871 clear_bit(idx, cpuc->used_mask); in armv8pmu_get_chain_idx() 886 if (!test_and_set_bit(ARMV8_IDX_CYCLE_COUNTER, cpuc->used_mask)) in armv8pmu_get_event_idx() 905 clear_bit(idx, cpuc->used_mask); in armv8pmu_clear_event_idx() 907 clear_bit(idx - 1, cpuc->used_mask); in armv8pmu_clear_event_idx()
|
/arch/x86/events/intel/ |
D | p4.c | 1203 static int p4_next_cntr(int thread, unsigned long *used_mask, in p4_next_cntr() argument 1210 if (j != -1 && !test_bit(j, used_mask)) in p4_next_cntr() 1219 unsigned long used_mask[BITS_TO_LONGS(X86_PMC_IDX_MAX)]; in p4_pmu_schedule_events() local 1229 bitmap_zero(used_mask, X86_PMC_IDX_MAX); in p4_pmu_schedule_events() 1259 cntr_idx = p4_next_cntr(thread, used_mask, bind); in p4_pmu_schedule_events() 1293 set_bit(cntr_idx, used_mask); in p4_pmu_schedule_events()
|
D | uncore.c | 445 unsigned long used_mask[BITS_TO_LONGS(UNCORE_PMC_IDX_MAX)]; in uncore_assign_events() local 450 bitmap_zero(used_mask, UNCORE_PMC_IDX_MAX); in uncore_assign_events() 473 if (test_bit(hwc->idx, used_mask)) in uncore_assign_events() 476 __set_bit(hwc->idx, used_mask); in uncore_assign_events()
|
/arch/nds32/include/asm/ |
D | pmu.h | 54 unsigned long used_mask[BITS_TO_LONGS(MAX_COUNTERS)]; member
|
/arch/mips/kernel/ |
D | perf_event_mipsxx.c | 39 unsigned long used_mask[BITS_TO_LONGS(MIPS_MAX_HWEVENTS)]; member 341 !test_and_set_bit(i, cpuc->used_mask)) in mipsxx_pmu_alloc_counter() 545 clear_bit(idx, cpuc->used_mask); in mipspmu_del() 1667 if (!test_bit(n, cpuc->used_mask)) in mipsxx_pmu_handle_shared_irq()
|
/arch/x86/events/ |
D | core.c | 979 u64 used_mask = 0; in x86_schedule_events() local 1039 if (used_mask & mask) in x86_schedule_events() 1042 used_mask |= mask; in x86_schedule_events()
|
/arch/csky/kernel/ |
D | perf_event.c | 33 unsigned long used_mask[BITS_TO_LONGS(CSKY_PMU_MAX_EVENTS)]; member
|