/arch/sh/kernel/ |
D | perf_event.c | 179 u64 prev_raw_count, new_raw_count; in sh_perf_event_update() local 197 new_raw_count = sh_pmu->read(idx); in sh_perf_event_update() 200 new_raw_count) != prev_raw_count) in sh_perf_event_update() 211 delta = (new_raw_count << shift) - (prev_raw_count << shift); in sh_perf_event_update()
|
/arch/riscv/kernel/ |
D | perf_event.c | 217 u64 prev_raw_count, new_raw_count; in riscv_pmu_read() local 224 new_raw_count = read_counter(idx); in riscv_pmu_read() 227 new_raw_count); in riscv_pmu_read() 233 delta = (new_raw_count - prev_raw_count) & in riscv_pmu_read()
|
/arch/xtensa/kernel/ |
D | perf_event.c | 148 uint64_t prev_raw_count, new_raw_count; in xtensa_perf_event_update() local 153 new_raw_count = xtensa_pmu_read_counter(event->hw.idx); in xtensa_perf_event_update() 155 new_raw_count) != prev_raw_count); in xtensa_perf_event_update() 157 delta = (new_raw_count - prev_raw_count) & XTENSA_PMU_COUNTER_MASK; in xtensa_perf_event_update()
|
/arch/alpha/kernel/ |
D | perf_event.c | 310 long prev_raw_count, new_raw_count; in alpha_perf_event_update() local 315 new_raw_count = alpha_read_pmc(idx); in alpha_perf_event_update() 318 new_raw_count) != prev_raw_count) in alpha_perf_event_update() 321 delta = (new_raw_count - (prev_raw_count & alpha_pmu->pmc_count_mask[idx])) + ovf; in alpha_perf_event_update() 333 return new_raw_count; in alpha_perf_event_update()
|
/arch/arm/mach-imx/ |
D | mmdc.c | 310 u64 delta, prev_raw_count, new_raw_count; in mmdc_pmu_event_update() local 314 new_raw_count = mmdc_pmu_read_counter(pmu_mmdc, in mmdc_pmu_event_update() 317 new_raw_count) != prev_raw_count); in mmdc_pmu_event_update() 319 delta = (new_raw_count - prev_raw_count) & 0xFFFFFFFF; in mmdc_pmu_event_update()
|
/arch/x86/events/ |
D | rapl.c | 178 u64 prev_raw_count, new_raw_count; in rapl_event_update() local 184 rdmsrl(event->hw.event_base, new_raw_count); in rapl_event_update() 187 new_raw_count) != prev_raw_count) { in rapl_event_update() 200 delta = (new_raw_count << shift) - (prev_raw_count << shift); in rapl_event_update() 207 return new_raw_count; in rapl_event_update()
|
D | core.c | 104 u64 prev_raw_count, new_raw_count; in x86_perf_event_update() local 122 rdpmcl(hwc->event_base_rdpmc, new_raw_count); in x86_perf_event_update() 125 new_raw_count) != prev_raw_count) in x86_perf_event_update() 136 delta = (new_raw_count << shift) - (prev_raw_count << shift); in x86_perf_event_update() 142 return new_raw_count; in x86_perf_event_update()
|
/arch/x86/events/intel/ |
D | cstate.c | 361 u64 prev_raw_count, new_raw_count; in cstate_pmu_event_update() local 365 new_raw_count = cstate_pmu_read_counter(event); in cstate_pmu_event_update() 368 new_raw_count) != prev_raw_count) in cstate_pmu_event_update() 371 local64_add(new_raw_count - prev_raw_count, &event->count); in cstate_pmu_event_update()
|
D | ds.c | 1677 u64 prev_raw_count, new_raw_count; in intel_pmu_save_and_restart_reload() local 1688 rdpmcl(hwc->event_base_rdpmc, new_raw_count); in intel_pmu_save_and_restart_reload() 1689 local64_set(&hwc->prev_count, new_raw_count); in intel_pmu_save_and_restart_reload() 1718 new = ((s64)(new_raw_count << shift) >> shift); in intel_pmu_save_and_restart_reload()
|
/arch/arc/kernel/ |
D | perf_event.c | 122 u64 new_raw_count = arc_pmu_read_counter(idx); in arc_perf_event_update() local 123 s64 delta = new_raw_count - prev_raw_count; in arc_perf_event_update() 129 local64_set(&hwc->prev_count, new_raw_count); in arc_perf_event_update()
|
/arch/nds32/kernel/ |
D | perf_event_cpu.c | 947 u64 delta, prev_raw_count, new_raw_count; in nds32_pmu_event_update() local 951 new_raw_count = nds32_pmu->read_counter(event); in nds32_pmu_event_update() 954 new_raw_count) != prev_raw_count) { in nds32_pmu_event_update() 961 delta = (new_raw_count - prev_raw_count) & nds32_pmu->max_period; in nds32_pmu_event_update() 966 return new_raw_count; in nds32_pmu_event_update()
|
/arch/x86/events/amd/ |
D | ibs.c | 156 perf_event_try_update(struct perf_event *event, u64 new_raw_count, int width) in perf_event_try_update() argument 172 new_raw_count) != prev_raw_count) in perf_event_try_update() 183 delta = (new_raw_count << shift) - (prev_raw_count << shift); in perf_event_try_update()
|
/arch/sparc/kernel/ |
D | perf_event.c | 867 u64 prev_raw_count, new_raw_count; in sparc_perf_event_update() local 872 new_raw_count = sparc_pmu->read_pmc(idx); in sparc_perf_event_update() 875 new_raw_count) != prev_raw_count) in sparc_perf_event_update() 878 delta = (new_raw_count << shift) - (prev_raw_count << shift); in sparc_perf_event_update() 884 return new_raw_count; in sparc_perf_event_update()
|
/arch/csky/kernel/ |
D | perf_event.c | 927 uint64_t new_raw_count = sign_extend64( in csky_perf_event_update() local 929 int64_t delta = new_raw_count - prev_raw_count; in csky_perf_event_update() 935 local64_set(&hwc->prev_count, new_raw_count); in csky_perf_event_update()
|
/arch/mips/kernel/ |
D | perf_event_mipsxx.c | 453 u64 prev_raw_count, new_raw_count; in mipspmu_event_update() local 458 new_raw_count = mipspmu.read_counter(idx); in mipspmu_event_update() 461 new_raw_count) != prev_raw_count) in mipspmu_event_update() 464 delta = new_raw_count - prev_raw_count; in mipspmu_event_update()
|