Home
last modified time | relevance | path

Searched refs:pmu (Results 1 – 25 of 107) sorted by relevance

12345

/arch/x86/kvm/
Dpmu.c50 struct kvm_pmu *pmu = &pmc->vcpu->arch.pmu; in pmc_bitmask() local
52 return pmu->counter_bitmask[pmc->type]; in pmc_bitmask()
57 struct kvm_pmu *pmu = &pmc->vcpu->arch.pmu; in pmc_enabled() local
58 return test_bit(pmc->idx, (unsigned long *)&pmu->global_ctrl); in pmc_enabled()
61 static inline struct kvm_pmc *get_gp_pmc(struct kvm_pmu *pmu, u32 msr, in get_gp_pmc() argument
64 if (msr >= base && msr < base + pmu->nr_arch_gp_counters) in get_gp_pmc()
65 return &pmu->gp_counters[msr - base]; in get_gp_pmc()
69 static inline struct kvm_pmc *get_fixed_pmc(struct kvm_pmu *pmu, u32 msr) in get_fixed_pmc() argument
72 if (msr >= base && msr < base + pmu->nr_arch_fixed_counters) in get_fixed_pmc()
73 return &pmu->fixed_counters[msr - base]; in get_fixed_pmc()
[all …]
/arch/x86/kernel/cpu/
Dperf_event_intel_rapl.c111 struct pmu *pmu; /* pointer to rapl_pmu_class */ member
116 static struct pmu rapl_pmu_class;
176 static void rapl_start_hrtimer(struct rapl_pmu *pmu) in rapl_start_hrtimer() argument
178 __hrtimer_start_range_ns(&pmu->hrtimer, in rapl_start_hrtimer()
179 pmu->timer_interval, 0, in rapl_start_hrtimer()
183 static void rapl_stop_hrtimer(struct rapl_pmu *pmu) in rapl_stop_hrtimer() argument
185 hrtimer_cancel(&pmu->hrtimer); in rapl_stop_hrtimer()
190 struct rapl_pmu *pmu = __this_cpu_read(rapl_pmu); in rapl_hrtimer_handle() local
194 if (!pmu->n_active) in rapl_hrtimer_handle()
197 spin_lock_irqsave(&pmu->lock, flags); in rapl_hrtimer_handle()
[all …]
Dperf_event_intel_uncore.h55 struct pmu *pmu; /* for custom pmu ops */ member
76 struct pmu pmu; member
103 struct intel_uncore_pmu *pmu; member
141 return box->pmu->type->box_ctl; in uncore_pci_box_ctl()
146 return box->pmu->type->fixed_ctl; in uncore_pci_fixed_ctl()
151 return box->pmu->type->fixed_ctr; in uncore_pci_fixed_ctr()
157 return idx * 4 + box->pmu->type->event_ctl; in uncore_pci_event_ctl()
163 return idx * 8 + box->pmu->type->perf_ctr; in uncore_pci_perf_ctr()
168 struct intel_uncore_pmu *pmu = box->pmu; in uncore_msr_box_offset() local
169 return pmu->type->msr_offsets ? in uncore_msr_box_offset()
[all …]
Dperf_event_intel_uncore.c33 return container_of(event->pmu, struct intel_uncore_pmu, pmu); in uncore_event_to_pmu()
36 struct intel_uncore_box *uncore_pmu_to_box(struct intel_uncore_pmu *pmu, int cpu) in uncore_pmu_to_box() argument
40 box = *per_cpu_ptr(pmu->box, cpu); in uncore_pmu_to_box()
46 if (*per_cpu_ptr(pmu->box, cpu)) in uncore_pmu_to_box()
48 list_for_each_entry(box, &pmu->box_list, list) { in uncore_pmu_to_box()
51 *per_cpu_ptr(pmu->box, cpu) = box; in uncore_pmu_to_box()
58 return *per_cpu_ptr(pmu->box, cpu); in uncore_pmu_to_box()
287 return event->pmu->event_init == uncore_pmu_event_init; in is_uncore_event()
296 max_count = box->pmu->type->num_counters; in uncore_collect_events()
297 if (box->pmu->type->fixed_ctl) in uncore_collect_events()
[all …]
Dperf_event_amd_uncore.c40 struct pmu *pmu; member
48 static struct pmu amd_nb_pmu;
49 static struct pmu amd_l2_pmu;
56 return event->pmu->type == amd_nb_pmu.type; in is_nb_event()
61 return event->pmu->type == amd_l2_pmu.type; in is_l2_event()
180 if (event->attr.type != event->pmu->type) in amd_uncore_event_init()
224 struct pmu *pmu = dev_get_drvdata(dev); in amd_uncore_attr_show_cpumask() local
226 if (pmu->type == amd_nb_pmu.type) in amd_uncore_attr_show_cpumask()
228 else if (pmu->type == amd_l2_pmu.type) in amd_uncore_attr_show_cpumask()
269 static struct pmu amd_nb_pmu = {
[all …]
Dperf_event_amd_iommu.c39 struct pmu pmu; member
207 if (event->attr.type != event->pmu->type) in perf_iommu_event_init()
228 if (event->pmu != &perf_iommu->pmu) in perf_iommu_event_init()
371 container_of(event->pmu, struct perf_amd_iommu, pmu); in perf_iommu_add()
392 container_of(event->pmu, struct perf_amd_iommu, pmu); in perf_iommu_del()
458 perf_iommu->pmu.attr_groups = perf_iommu->attr_groups; in _init_perf_amd_iommu()
460 ret = perf_pmu_register(&perf_iommu->pmu, name, -1); in _init_perf_amd_iommu()
474 .pmu = {
Dperf_event_amd_ibs.c45 struct pmu pmu; member
157 if (perf_ibs_fetch.pmu.type == type) in get_ibs_pmu()
159 if (perf_ibs_op.pmu.type == type) in get_ibs_pmu()
244 if (event->pmu != &perf_ibs->pmu) in perf_ibs_init()
368 struct perf_ibs *perf_ibs = container_of(event->pmu, struct perf_ibs, pmu); in perf_ibs_start()
388 struct perf_ibs *perf_ibs = container_of(event->pmu, struct perf_ibs, pmu); in perf_ibs_stop()
422 struct perf_ibs *perf_ibs = container_of(event->pmu, struct perf_ibs, pmu); in perf_ibs_add()
440 struct perf_ibs *perf_ibs = container_of(event->pmu, struct perf_ibs, pmu); in perf_ibs_del()
469 .pmu = {
493 .pmu = {
[all …]
/arch/powerpc/perf/
DMakefile6 obj64-$(CONFIG_PPC_PERF_CTRS) += power4-pmu.o ppc970-pmu.o power5-pmu.o \
7 power5+-pmu.o power6-pmu.o power7-pmu.o \
8 power8-pmu.o
9 obj32-$(CONFIG_PPC_PERF_CTRS) += mpc7450-pmu.o
12 obj-$(CONFIG_FSL_EMB_PERF_EVENT_E500) += e500-pmu.o e6500-pmu.o
Dcore-fsl-emb.c207 static void fsl_emb_pmu_disable(struct pmu *pmu) in fsl_emb_pmu_disable() argument
246 static void fsl_emb_pmu_enable(struct pmu *pmu) in fsl_emb_pmu_enable() argument
301 perf_pmu_disable(event->pmu); in fsl_emb_pmu_add()
347 perf_pmu_enable(event->pmu); in fsl_emb_pmu_add()
357 perf_pmu_disable(event->pmu); in fsl_emb_pmu_del()
385 perf_pmu_enable(event->pmu); in fsl_emb_pmu_del()
404 perf_pmu_disable(event->pmu); in fsl_emb_pmu_start()
411 perf_pmu_enable(event->pmu); in fsl_emb_pmu_start()
426 perf_pmu_disable(event->pmu); in fsl_emb_pmu_stop()
433 perf_pmu_enable(event->pmu); in fsl_emb_pmu_stop()
[all …]
/arch/arm/kernel/
Dperf_event_cpu.c191 cpu_pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in cpu_pmu_init()
247 static int probe_current_pmu(struct arm_pmu *pmu) in probe_current_pmu() argument
257 ret = armv6_1136_pmu_init(pmu); in probe_current_pmu()
260 ret = armv6_1156_pmu_init(pmu); in probe_current_pmu()
263 ret = armv6_1176_pmu_init(pmu); in probe_current_pmu()
266 ret = armv6mpcore_pmu_init(pmu); in probe_current_pmu()
269 ret = armv7_a8_pmu_init(pmu); in probe_current_pmu()
272 ret = armv7_a9_pmu_init(pmu); in probe_current_pmu()
279 ret = xscale1pmu_init(pmu); in probe_current_pmu()
282 ret = xscale2pmu_init(pmu); in probe_current_pmu()
[all …]
Dperf_event.c98 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_set_period()
132 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_update()
161 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_stop()
177 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_start()
202 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_del()
219 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_add()
225 perf_pmu_disable(event->pmu); in armpmu_add()
250 perf_pmu_enable(event->pmu); in armpmu_add()
258 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in validate_event()
354 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in hw_perf_event_destroy()
[all …]
/arch/arc/kernel/
Dperf_event.c21 struct pmu pmu; member
50 struct arc_pmu *arc_pmu = container_of(event->pmu, struct arc_pmu, pmu); in arc_perf_event_update()
98 struct arc_pmu *arc_pmu = container_of(event->pmu, struct arc_pmu, pmu); in arc_pmu_event_init()
124 static void arc_pmu_enable(struct pmu *pmu) in arc_pmu_enable() argument
132 static void arc_pmu_disable(struct pmu *pmu) in arc_pmu_disable() argument
186 struct arc_pmu *arc_pmu = container_of(event->pmu, struct arc_pmu, pmu); in arc_pmu_del()
197 struct arc_pmu *arc_pmu = container_of(event->pmu, struct arc_pmu, pmu); in arc_pmu_add()
284 arc_pmu->pmu = (struct pmu) { in arc_pmu_device_probe()
296 arc_pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in arc_pmu_device_probe()
298 ret = perf_pmu_register(&arc_pmu->pmu, pdev->name, PERF_TYPE_RAW); in arc_pmu_device_probe()
/arch/sh/kernel/
Dperf_event.c272 perf_pmu_disable(event->pmu); in sh_pmu_add()
292 perf_pmu_enable(event->pmu); in sh_pmu_add()
328 static void sh_pmu_enable(struct pmu *pmu) in sh_pmu_enable() argument
336 static void sh_pmu_disable(struct pmu *pmu) in sh_pmu_disable() argument
344 static struct pmu pmu = { variable
392 pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in register_sh_pmu()
396 perf_pmu_register(&pmu, "cpu", PERF_TYPE_RAW); in register_sh_pmu()
/arch/arm64/kernel/
Dperf_event.c49 #define to_arm_pmu(p) (container_of(p, struct arm_pmu, pmu))
153 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_set_period()
189 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_update()
223 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_stop()
241 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_start()
266 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_del()
283 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_add()
289 perf_pmu_disable(event->pmu); in armpmu_add()
314 perf_pmu_enable(event->pmu); in armpmu_add()
319 validate_event(struct pmu *pmu, struct pmu_hw_events *hw_events, in validate_event() argument
[all …]
/arch/arm64/include/asm/
Dpmu.h45 struct pmu pmu; member
69 #define to_arm_pmu(p) (container_of(p, struct arm_pmu, pmu))
/arch/s390/kernel/
Dperf_cpum_cf.c174 static void cpumf_pmu_enable(struct pmu *pmu) in cpumf_pmu_enable() argument
197 static void cpumf_pmu_disable(struct pmu *pmu) in cpumf_pmu_disable() argument
576 static void cpumf_pmu_start_txn(struct pmu *pmu) in cpumf_pmu_start_txn() argument
580 perf_pmu_disable(pmu); in cpumf_pmu_start_txn()
590 static void cpumf_pmu_cancel_txn(struct pmu *pmu) in cpumf_pmu_cancel_txn() argument
597 perf_pmu_enable(pmu); in cpumf_pmu_cancel_txn()
605 static int cpumf_pmu_commit_txn(struct pmu *pmu) in cpumf_pmu_commit_txn() argument
617 perf_pmu_enable(pmu); in cpumf_pmu_commit_txn()
622 static struct pmu cpumf_pmu = {
/arch/blackfin/kernel/
Dperf_event.c353 perf_pmu_disable(event->pmu); in bfin_pmu_add()
373 perf_pmu_enable(event->pmu); in bfin_pmu_add()
422 static void bfin_pmu_enable(struct pmu *pmu) in bfin_pmu_enable() argument
440 static void bfin_pmu_disable(struct pmu *pmu) in bfin_pmu_disable() argument
445 static struct pmu pmu = { variable
490 pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in bfin_pmu_init()
492 ret = perf_pmu_register(&pmu, "cpu", PERF_TYPE_RAW); in bfin_pmu_init()
/arch/arm/include/asm/
Dpmu.h85 struct pmu pmu; member
113 #define to_arm_pmu(p) (container_of(p, struct arm_pmu, pmu))
/arch/x86/include/asm/
Dhw_breakpoint.h51 struct pmu;
73 extern struct pmu perf_ops_bp;
/arch/arm/plat-iop/
DMakefile12 obj-$(CONFIG_ARCH_IOP32X) += pmu.o
22 obj-$(CONFIG_ARCH_IOP33X) += pmu.o
/arch/sh/include/asm/
Dhw_breakpoint.h45 struct pmu;
68 extern struct pmu perf_ops_bp;
/arch/metag/kernel/perf/
Dperf_event.h77 struct pmu pmu; member
Dperf_event.c136 static void metag_pmu_enable(struct pmu *pmu) in metag_pmu_enable() argument
140 static void metag_pmu_disable(struct pmu *pmu) in metag_pmu_disable() argument
313 perf_pmu_disable(event->pmu); in metag_pmu_add()
345 perf_pmu_enable(event->pmu); in metag_pmu_add()
374 static struct pmu pmu = { variable
853 metag_pmu->pmu = pmu; in init_hw_perf_events()
865 metag_pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in init_hw_perf_events()
884 ret = perf_pmu_register(&pmu, metag_pmu->name, PERF_TYPE_RAW); in init_hw_perf_events()
/arch/arm/boot/dts/
Ddra72x.dtsi26 pmu {
27 compatible = "arm,cortex-a15-pmu";
/arch/powerpc/include/asm/
Dhw_breakpoint.h56 struct pmu;
72 extern struct pmu perf_ops_bp;

12345