/arch/x86/kvm/ |
D | pmu.c | 50 struct kvm_pmu *pmu = &pmc->vcpu->arch.pmu; in pmc_bitmask() local 52 return pmu->counter_bitmask[pmc->type]; in pmc_bitmask() 57 struct kvm_pmu *pmu = &pmc->vcpu->arch.pmu; in pmc_enabled() local 58 return test_bit(pmc->idx, (unsigned long *)&pmu->global_ctrl); in pmc_enabled() 61 static inline struct kvm_pmc *get_gp_pmc(struct kvm_pmu *pmu, u32 msr, in get_gp_pmc() argument 64 if (msr >= base && msr < base + pmu->nr_arch_gp_counters) in get_gp_pmc() 65 return &pmu->gp_counters[msr - base]; in get_gp_pmc() 69 static inline struct kvm_pmc *get_fixed_pmc(struct kvm_pmu *pmu, u32 msr) in get_fixed_pmc() argument 72 if (msr >= base && msr < base + pmu->nr_arch_fixed_counters) in get_fixed_pmc() 73 return &pmu->fixed_counters[msr - base]; in get_fixed_pmc() [all …]
|
/arch/x86/kernel/cpu/ |
D | perf_event_intel_rapl.c | 111 struct pmu *pmu; /* pointer to rapl_pmu_class */ member 116 static struct pmu rapl_pmu_class; 176 static void rapl_start_hrtimer(struct rapl_pmu *pmu) in rapl_start_hrtimer() argument 178 __hrtimer_start_range_ns(&pmu->hrtimer, in rapl_start_hrtimer() 179 pmu->timer_interval, 0, in rapl_start_hrtimer() 183 static void rapl_stop_hrtimer(struct rapl_pmu *pmu) in rapl_stop_hrtimer() argument 185 hrtimer_cancel(&pmu->hrtimer); in rapl_stop_hrtimer() 190 struct rapl_pmu *pmu = __this_cpu_read(rapl_pmu); in rapl_hrtimer_handle() local 194 if (!pmu->n_active) in rapl_hrtimer_handle() 197 spin_lock_irqsave(&pmu->lock, flags); in rapl_hrtimer_handle() [all …]
|
D | perf_event_intel_uncore.h | 55 struct pmu *pmu; /* for custom pmu ops */ member 76 struct pmu pmu; member 103 struct intel_uncore_pmu *pmu; member 141 return box->pmu->type->box_ctl; in uncore_pci_box_ctl() 146 return box->pmu->type->fixed_ctl; in uncore_pci_fixed_ctl() 151 return box->pmu->type->fixed_ctr; in uncore_pci_fixed_ctr() 157 return idx * 4 + box->pmu->type->event_ctl; in uncore_pci_event_ctl() 163 return idx * 8 + box->pmu->type->perf_ctr; in uncore_pci_perf_ctr() 168 struct intel_uncore_pmu *pmu = box->pmu; in uncore_msr_box_offset() local 169 return pmu->type->msr_offsets ? in uncore_msr_box_offset() [all …]
|
D | perf_event_intel_uncore.c | 33 return container_of(event->pmu, struct intel_uncore_pmu, pmu); in uncore_event_to_pmu() 36 struct intel_uncore_box *uncore_pmu_to_box(struct intel_uncore_pmu *pmu, int cpu) in uncore_pmu_to_box() argument 40 box = *per_cpu_ptr(pmu->box, cpu); in uncore_pmu_to_box() 46 if (*per_cpu_ptr(pmu->box, cpu)) in uncore_pmu_to_box() 48 list_for_each_entry(box, &pmu->box_list, list) { in uncore_pmu_to_box() 51 *per_cpu_ptr(pmu->box, cpu) = box; in uncore_pmu_to_box() 58 return *per_cpu_ptr(pmu->box, cpu); in uncore_pmu_to_box() 287 return event->pmu->event_init == uncore_pmu_event_init; in is_uncore_event() 296 max_count = box->pmu->type->num_counters; in uncore_collect_events() 297 if (box->pmu->type->fixed_ctl) in uncore_collect_events() [all …]
|
D | perf_event_amd_uncore.c | 40 struct pmu *pmu; member 48 static struct pmu amd_nb_pmu; 49 static struct pmu amd_l2_pmu; 56 return event->pmu->type == amd_nb_pmu.type; in is_nb_event() 61 return event->pmu->type == amd_l2_pmu.type; in is_l2_event() 180 if (event->attr.type != event->pmu->type) in amd_uncore_event_init() 224 struct pmu *pmu = dev_get_drvdata(dev); in amd_uncore_attr_show_cpumask() local 226 if (pmu->type == amd_nb_pmu.type) in amd_uncore_attr_show_cpumask() 228 else if (pmu->type == amd_l2_pmu.type) in amd_uncore_attr_show_cpumask() 269 static struct pmu amd_nb_pmu = { [all …]
|
D | perf_event_amd_iommu.c | 39 struct pmu pmu; member 207 if (event->attr.type != event->pmu->type) in perf_iommu_event_init() 228 if (event->pmu != &perf_iommu->pmu) in perf_iommu_event_init() 371 container_of(event->pmu, struct perf_amd_iommu, pmu); in perf_iommu_add() 392 container_of(event->pmu, struct perf_amd_iommu, pmu); in perf_iommu_del() 458 perf_iommu->pmu.attr_groups = perf_iommu->attr_groups; in _init_perf_amd_iommu() 460 ret = perf_pmu_register(&perf_iommu->pmu, name, -1); in _init_perf_amd_iommu() 474 .pmu = {
|
D | perf_event_amd_ibs.c | 45 struct pmu pmu; member 157 if (perf_ibs_fetch.pmu.type == type) in get_ibs_pmu() 159 if (perf_ibs_op.pmu.type == type) in get_ibs_pmu() 244 if (event->pmu != &perf_ibs->pmu) in perf_ibs_init() 368 struct perf_ibs *perf_ibs = container_of(event->pmu, struct perf_ibs, pmu); in perf_ibs_start() 388 struct perf_ibs *perf_ibs = container_of(event->pmu, struct perf_ibs, pmu); in perf_ibs_stop() 422 struct perf_ibs *perf_ibs = container_of(event->pmu, struct perf_ibs, pmu); in perf_ibs_add() 440 struct perf_ibs *perf_ibs = container_of(event->pmu, struct perf_ibs, pmu); in perf_ibs_del() 469 .pmu = { 493 .pmu = { [all …]
|
/arch/powerpc/perf/ |
D | Makefile | 6 obj64-$(CONFIG_PPC_PERF_CTRS) += power4-pmu.o ppc970-pmu.o power5-pmu.o \ 7 power5+-pmu.o power6-pmu.o power7-pmu.o \ 8 power8-pmu.o 9 obj32-$(CONFIG_PPC_PERF_CTRS) += mpc7450-pmu.o 12 obj-$(CONFIG_FSL_EMB_PERF_EVENT_E500) += e500-pmu.o e6500-pmu.o
|
D | core-fsl-emb.c | 207 static void fsl_emb_pmu_disable(struct pmu *pmu) in fsl_emb_pmu_disable() argument 246 static void fsl_emb_pmu_enable(struct pmu *pmu) in fsl_emb_pmu_enable() argument 301 perf_pmu_disable(event->pmu); in fsl_emb_pmu_add() 347 perf_pmu_enable(event->pmu); in fsl_emb_pmu_add() 357 perf_pmu_disable(event->pmu); in fsl_emb_pmu_del() 385 perf_pmu_enable(event->pmu); in fsl_emb_pmu_del() 404 perf_pmu_disable(event->pmu); in fsl_emb_pmu_start() 411 perf_pmu_enable(event->pmu); in fsl_emb_pmu_start() 426 perf_pmu_disable(event->pmu); in fsl_emb_pmu_stop() 433 perf_pmu_enable(event->pmu); in fsl_emb_pmu_stop() [all …]
|
/arch/arm/kernel/ |
D | perf_event_cpu.c | 191 cpu_pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in cpu_pmu_init() 247 static int probe_current_pmu(struct arm_pmu *pmu) in probe_current_pmu() argument 257 ret = armv6_1136_pmu_init(pmu); in probe_current_pmu() 260 ret = armv6_1156_pmu_init(pmu); in probe_current_pmu() 263 ret = armv6_1176_pmu_init(pmu); in probe_current_pmu() 266 ret = armv6mpcore_pmu_init(pmu); in probe_current_pmu() 269 ret = armv7_a8_pmu_init(pmu); in probe_current_pmu() 272 ret = armv7_a9_pmu_init(pmu); in probe_current_pmu() 279 ret = xscale1pmu_init(pmu); in probe_current_pmu() 282 ret = xscale2pmu_init(pmu); in probe_current_pmu() [all …]
|
D | perf_event.c | 98 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_set_period() 132 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_update() 161 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_stop() 177 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_start() 202 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_del() 219 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_add() 225 perf_pmu_disable(event->pmu); in armpmu_add() 250 perf_pmu_enable(event->pmu); in armpmu_add() 258 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in validate_event() 354 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in hw_perf_event_destroy() [all …]
|
/arch/arc/kernel/ |
D | perf_event.c | 21 struct pmu pmu; member 50 struct arc_pmu *arc_pmu = container_of(event->pmu, struct arc_pmu, pmu); in arc_perf_event_update() 98 struct arc_pmu *arc_pmu = container_of(event->pmu, struct arc_pmu, pmu); in arc_pmu_event_init() 124 static void arc_pmu_enable(struct pmu *pmu) in arc_pmu_enable() argument 132 static void arc_pmu_disable(struct pmu *pmu) in arc_pmu_disable() argument 186 struct arc_pmu *arc_pmu = container_of(event->pmu, struct arc_pmu, pmu); in arc_pmu_del() 197 struct arc_pmu *arc_pmu = container_of(event->pmu, struct arc_pmu, pmu); in arc_pmu_add() 284 arc_pmu->pmu = (struct pmu) { in arc_pmu_device_probe() 296 arc_pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in arc_pmu_device_probe() 298 ret = perf_pmu_register(&arc_pmu->pmu, pdev->name, PERF_TYPE_RAW); in arc_pmu_device_probe()
|
/arch/sh/kernel/ |
D | perf_event.c | 272 perf_pmu_disable(event->pmu); in sh_pmu_add() 292 perf_pmu_enable(event->pmu); in sh_pmu_add() 328 static void sh_pmu_enable(struct pmu *pmu) in sh_pmu_enable() argument 336 static void sh_pmu_disable(struct pmu *pmu) in sh_pmu_disable() argument 344 static struct pmu pmu = { variable 392 pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in register_sh_pmu() 396 perf_pmu_register(&pmu, "cpu", PERF_TYPE_RAW); in register_sh_pmu()
|
/arch/arm64/kernel/ |
D | perf_event.c | 49 #define to_arm_pmu(p) (container_of(p, struct arm_pmu, pmu)) 153 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_set_period() 189 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_update() 223 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_stop() 241 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_start() 266 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_del() 283 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_add() 289 perf_pmu_disable(event->pmu); in armpmu_add() 314 perf_pmu_enable(event->pmu); in armpmu_add() 319 validate_event(struct pmu *pmu, struct pmu_hw_events *hw_events, in validate_event() argument [all …]
|
/arch/arm64/include/asm/ |
D | pmu.h | 45 struct pmu pmu; member 69 #define to_arm_pmu(p) (container_of(p, struct arm_pmu, pmu))
|
/arch/s390/kernel/ |
D | perf_cpum_cf.c | 174 static void cpumf_pmu_enable(struct pmu *pmu) in cpumf_pmu_enable() argument 197 static void cpumf_pmu_disable(struct pmu *pmu) in cpumf_pmu_disable() argument 576 static void cpumf_pmu_start_txn(struct pmu *pmu) in cpumf_pmu_start_txn() argument 580 perf_pmu_disable(pmu); in cpumf_pmu_start_txn() 590 static void cpumf_pmu_cancel_txn(struct pmu *pmu) in cpumf_pmu_cancel_txn() argument 597 perf_pmu_enable(pmu); in cpumf_pmu_cancel_txn() 605 static int cpumf_pmu_commit_txn(struct pmu *pmu) in cpumf_pmu_commit_txn() argument 617 perf_pmu_enable(pmu); in cpumf_pmu_commit_txn() 622 static struct pmu cpumf_pmu = {
|
/arch/blackfin/kernel/ |
D | perf_event.c | 353 perf_pmu_disable(event->pmu); in bfin_pmu_add() 373 perf_pmu_enable(event->pmu); in bfin_pmu_add() 422 static void bfin_pmu_enable(struct pmu *pmu) in bfin_pmu_enable() argument 440 static void bfin_pmu_disable(struct pmu *pmu) in bfin_pmu_disable() argument 445 static struct pmu pmu = { variable 490 pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in bfin_pmu_init() 492 ret = perf_pmu_register(&pmu, "cpu", PERF_TYPE_RAW); in bfin_pmu_init()
|
/arch/arm/include/asm/ |
D | pmu.h | 85 struct pmu pmu; member 113 #define to_arm_pmu(p) (container_of(p, struct arm_pmu, pmu))
|
/arch/x86/include/asm/ |
D | hw_breakpoint.h | 51 struct pmu; 73 extern struct pmu perf_ops_bp;
|
/arch/arm/plat-iop/ |
D | Makefile | 12 obj-$(CONFIG_ARCH_IOP32X) += pmu.o 22 obj-$(CONFIG_ARCH_IOP33X) += pmu.o
|
/arch/sh/include/asm/ |
D | hw_breakpoint.h | 45 struct pmu; 68 extern struct pmu perf_ops_bp;
|
/arch/metag/kernel/perf/ |
D | perf_event.h | 77 struct pmu pmu; member
|
D | perf_event.c | 136 static void metag_pmu_enable(struct pmu *pmu) in metag_pmu_enable() argument 140 static void metag_pmu_disable(struct pmu *pmu) in metag_pmu_disable() argument 313 perf_pmu_disable(event->pmu); in metag_pmu_add() 345 perf_pmu_enable(event->pmu); in metag_pmu_add() 374 static struct pmu pmu = { variable 853 metag_pmu->pmu = pmu; in init_hw_perf_events() 865 metag_pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in init_hw_perf_events() 884 ret = perf_pmu_register(&pmu, metag_pmu->name, PERF_TYPE_RAW); in init_hw_perf_events()
|
/arch/arm/boot/dts/ |
D | dra72x.dtsi | 26 pmu { 27 compatible = "arm,cortex-a15-pmu";
|
/arch/powerpc/include/asm/ |
D | hw_breakpoint.h | 56 struct pmu; 72 extern struct pmu perf_ops_bp;
|