/arch/sh/kernel/ |
D | perf_event.c | 182 u64 prev_raw_count, new_raw_count; in sh_perf_event_update() local 200 new_raw_count = sh_pmu->read(idx); in sh_perf_event_update() 203 new_raw_count) != prev_raw_count) in sh_perf_event_update() 214 delta = (new_raw_count << shift) - (prev_raw_count << shift); in sh_perf_event_update()
|
/arch/blackfin/kernel/ |
D | perf_event.c | 263 u64 prev_raw_count, new_raw_count; in bfin_perf_event_update() local 281 new_raw_count = bfin_pfmon_read(idx); in bfin_perf_event_update() 284 new_raw_count) != prev_raw_count) in bfin_perf_event_update() 295 delta = (new_raw_count << shift) - (prev_raw_count << shift); in bfin_perf_event_update()
|
/arch/xtensa/kernel/ |
D | perf_event.c | 142 uint64_t prev_raw_count, new_raw_count; in xtensa_perf_event_update() local 147 new_raw_count = xtensa_pmu_read_counter(event->hw.idx); in xtensa_perf_event_update() 149 new_raw_count) != prev_raw_count); in xtensa_perf_event_update() 151 delta = (new_raw_count - prev_raw_count) & XTENSA_PMU_COUNTER_MASK; in xtensa_perf_event_update()
|
/arch/alpha/kernel/ |
D | perf_event.c | 310 long prev_raw_count, new_raw_count; in alpha_perf_event_update() local 315 new_raw_count = alpha_read_pmc(idx); in alpha_perf_event_update() 318 new_raw_count) != prev_raw_count) in alpha_perf_event_update() 321 delta = (new_raw_count - (prev_raw_count & alpha_pmu->pmc_count_mask[idx])) + ovf; in alpha_perf_event_update() 333 return new_raw_count; in alpha_perf_event_update()
|
/arch/x86/events/intel/ |
D | rapl.c | 198 u64 prev_raw_count, new_raw_count; in rapl_event_update() local 204 rdmsrl(event->hw.event_base, new_raw_count); in rapl_event_update() 207 new_raw_count) != prev_raw_count) { in rapl_event_update() 220 delta = (new_raw_count << shift) - (prev_raw_count << shift); in rapl_event_update() 227 return new_raw_count; in rapl_event_update()
|
D | cstate.c | 335 u64 prev_raw_count, new_raw_count; in cstate_pmu_event_update() local 339 new_raw_count = cstate_pmu_read_counter(event); in cstate_pmu_event_update() 342 new_raw_count) != prev_raw_count) in cstate_pmu_event_update() 345 local64_add(new_raw_count - prev_raw_count, &event->count); in cstate_pmu_event_update()
|
D | ds.c | 1326 u64 prev_raw_count, new_raw_count; in intel_pmu_save_and_restart_reload() local 1337 rdpmcl(hwc->event_base_rdpmc, new_raw_count); in intel_pmu_save_and_restart_reload() 1338 local64_set(&hwc->prev_count, new_raw_count); in intel_pmu_save_and_restart_reload() 1367 new = ((s64)(new_raw_count << shift) >> shift); in intel_pmu_save_and_restart_reload()
|
/arch/arm/mach-imx/ |
D | mmdc.c | 319 u64 delta, prev_raw_count, new_raw_count; in mmdc_pmu_event_update() local 323 new_raw_count = mmdc_pmu_read_counter(pmu_mmdc, in mmdc_pmu_event_update() 326 new_raw_count) != prev_raw_count); in mmdc_pmu_event_update() 328 delta = (new_raw_count - prev_raw_count) & 0xFFFFFFFF; in mmdc_pmu_event_update()
|
/arch/tile/kernel/ |
D | perf_event.c | 506 u64 prev_raw_count, new_raw_count; in tile_perf_event_update() local 520 new_raw_count = read_counter(idx); in tile_perf_event_update() 523 new_raw_count); in tile_perf_event_update() 535 delta = (new_raw_count << shift) - (prev_raw_count << shift); in tile_perf_event_update() 541 return new_raw_count; in tile_perf_event_update()
|
/arch/metag/kernel/perf/ |
D | perf_event.c | 192 u64 prev_raw_count, new_raw_count; in metag_pmu_event_update() local 206 new_raw_count = metag_pmu->read(idx); in metag_pmu_event_update() 209 new_raw_count) != prev_raw_count) in metag_pmu_event_update() 215 delta = (new_raw_count - prev_raw_count) & MAX_PERIOD; in metag_pmu_event_update()
|
/arch/arc/kernel/ |
D | perf_event.c | 107 uint64_t new_raw_count = arc_pmu_read_counter(idx); in arc_perf_event_update() local 108 int64_t delta = new_raw_count - prev_raw_count; in arc_perf_event_update() 114 local64_set(&hwc->prev_count, new_raw_count); in arc_perf_event_update()
|
/arch/x86/events/amd/ |
D | ibs.c | 154 perf_event_try_update(struct perf_event *event, u64 new_raw_count, int width) in perf_event_try_update() argument 170 new_raw_count) != prev_raw_count) in perf_event_try_update() 181 delta = (new_raw_count << shift) - (prev_raw_count << shift); in perf_event_try_update()
|
/arch/sparc/kernel/ |
D | perf_event.c | 867 u64 prev_raw_count, new_raw_count; in sparc_perf_event_update() local 872 new_raw_count = sparc_pmu->read_pmc(idx); in sparc_perf_event_update() 875 new_raw_count) != prev_raw_count) in sparc_perf_event_update() 878 delta = (new_raw_count << shift) - (prev_raw_count << shift); in sparc_perf_event_update() 884 return new_raw_count; in sparc_perf_event_update()
|
/arch/mips/kernel/ |
D | perf_event_mipsxx.c | 404 u64 prev_raw_count, new_raw_count; in mipspmu_event_update() local 409 new_raw_count = mipspmu.read_counter(idx); in mipspmu_event_update() 412 new_raw_count) != prev_raw_count) in mipspmu_event_update() 415 delta = new_raw_count - prev_raw_count; in mipspmu_event_update()
|
/arch/x86/events/ |
D | core.c | 72 u64 prev_raw_count, new_raw_count; in x86_perf_event_update() local 88 rdpmcl(hwc->event_base_rdpmc, new_raw_count); in x86_perf_event_update() 91 new_raw_count) != prev_raw_count) in x86_perf_event_update() 102 delta = (new_raw_count << shift) - (prev_raw_count << shift); in x86_perf_event_update() 108 return new_raw_count; in x86_perf_event_update()
|