Home
last modified time | relevance | path

Searched refs:counters (Results 1 – 23 of 23) sorted by relevance

/arch/mips/oprofile/
Dop_model_mipsxx.c87 static inline unsigned int counters_total_to_per_cpu(unsigned int counters) in counters_total_to_per_cpu() argument
89 return counters >> vpe_shift(); in counters_total_to_per_cpu()
92 static inline unsigned int counters_per_cpu_to_total(unsigned int counters) in counters_per_cpu_to_total() argument
94 return counters << vpe_shift(); in counters_per_cpu_to_total()
152 unsigned int counters = op_model_mipsxx_ops.num_counters; in mipsxx_reg_setup() local
156 for (i = 0; i < counters; i++) { in mipsxx_reg_setup()
181 unsigned int counters = op_model_mipsxx_ops.num_counters; in mipsxx_cpu_setup() local
186 switch (counters) { in mipsxx_cpu_setup()
205 unsigned int counters = op_model_mipsxx_ops.num_counters; in mipsxx_cpu_start() local
210 switch (counters) { in mipsxx_cpu_start()
[all …]
/arch/x86/oprofile/
Dnmi_int.c77 struct op_msr *counters = msrs->counters; in nmi_cpu_save_registers() local
82 if (counters[i].addr) in nmi_cpu_save_registers()
83 rdmsrl(counters[i].addr, counters[i].saved); in nmi_cpu_save_registers()
202 struct op_msr *counters = msrs->counters; in nmi_cpu_save_mpx_registers() local
208 if (counters[i].addr) in nmi_cpu_save_mpx_registers()
209 rdmsrl(counters[i].addr, multiplex[virt].saved); in nmi_cpu_save_mpx_registers()
215 struct op_msr *counters = msrs->counters; in nmi_cpu_restore_mpx_registers() local
221 if (counters[i].addr) in nmi_cpu_restore_mpx_registers()
222 wrmsrl(counters[i].addr, multiplex[virt].saved); in nmi_cpu_restore_mpx_registers()
307 kfree(per_cpu(cpu_msrs, i).counters); in free_msrs()
[all …]
Dop_model_ppro.c38 if (!msrs->counters[i].addr) in ppro_shutdown()
57 msrs->counters[i].addr = MSR_P6_PERFCTR0 + i; in ppro_fill_in_addresses()
108 wrmsrl(msrs->counters[i].addr, -1LL); in ppro_setup_ctrs()
113 if (counter_config[i].enabled && msrs->counters[i].addr) { in ppro_setup_ctrs()
115 wrmsrl(msrs->counters[i].addr, -reset_value[i]); in ppro_setup_ctrs()
136 rdmsrl(msrs->counters[i].addr, val); in ppro_check_ctrs()
140 wrmsrl(msrs->counters[i].addr, -reset_value[i]); in ppro_check_ctrs()
Dop_model_amd.c295 if (!msrs->counters[i].addr) in op_amd_shutdown()
315 msrs->counters[i].addr = MSR_F15H_PERF_CTR + (i << 1); in op_amd_fill_in_addresses()
319 msrs->counters[i].addr = MSR_K7_PERFCTR0 + i; in op_amd_fill_in_addresses()
342 && msrs->counters[op_x86_virt_to_phys(i)].addr) in op_amd_setup_ctrs()
361 wrmsrl(msrs->counters[i].addr, -1LL); in op_amd_setup_ctrs()
371 wrmsrl(msrs->counters[i].addr, -(u64)reset_value[virt]); in op_amd_setup_ctrs()
391 rdmsrl(msrs->counters[i].addr, val); in op_amd_check_ctrs()
396 wrmsrl(msrs->counters[i].addr, -(u64)reset_value[virt]); in op_amd_check_ctrs()
Dop_x86_model.h24 struct op_msr *counters; member
Dop_model_p4.c393 if (msrs->counters[i].addr) in p4_shutdown()
394 release_perfctr_nmi(msrs->counters[i].addr); in p4_shutdown()
420 msrs->counters[i].addr = addr; in p4_fill_in_addresses()
/arch/mips/include/asm/
Dspinlock.h40 u32 counters = ACCESS_ONCE(lock->lock); in arch_spin_is_locked() local
42 return ((counters >> 16) ^ counters) & 0xffff; in arch_spin_is_locked()
56 u32 counters = ACCESS_ONCE(lock->lock); in arch_spin_is_contended() local
58 return (((counters >> 16) - counters) & 0xffff) > 1; in arch_spin_is_contended()
/arch/mips/kernel/
Dperf_event_mipsxx.c168 static unsigned int counters_total_to_per_cpu(unsigned int counters) in counters_total_to_per_cpu() argument
170 return counters >> vpe_shift(); in counters_total_to_per_cpu()
771 int counters; in n_counters() local
775 counters = 2; in n_counters()
781 counters = 4; in n_counters()
785 counters = __n_counters(); in n_counters()
788 return counters; in n_counters()
793 int counters = (int)(long)arg; in reset_counters() local
794 switch (counters) { in reset_counters()
1421 unsigned int counters = mipspmu.num_counters; in mipsxx_pmu_handle_shared_irq() local
[all …]
/arch/x86/kvm/
Dpmu_amd.c84 struct kvm_pmc *counters; in amd_msr_idx_to_pmc() local
89 counters = pmu->gp_counters; in amd_msr_idx_to_pmc()
91 return &counters[idx]; in amd_msr_idx_to_pmc()
Dpmu_intel.c137 struct kvm_pmc *counters; in intel_msr_idx_to_pmc() local
142 counters = pmu->fixed_counters; in intel_msr_idx_to_pmc()
145 counters = pmu->gp_counters; in intel_msr_idx_to_pmc()
150 return &counters[array_index_nospec(idx, num_counters)]; in intel_msr_idx_to_pmc()
/arch/mips/kvm/
DKconfig38 bool "Maintain counters for COP0 accesses"
/arch/powerpc/platforms/pseries/
DKconfig124 Enable access to hypervisor supplied counters in perf. Currently,
126 interfaces to retrieve counters. GPCI exists on Power 6 and later
/arch/s390/kvm/
Dinterrupt.c645 fi->counters[FIRQ_CNTR_PFAULT] -= 1; in __deliver_pfault_done()
695 fi->counters[FIRQ_CNTR_VIRTIO] -= 1; in __deliver_virtio()
746 fi->counters[FIRQ_CNTR_IO] -= 1; in __deliver_io()
1195 fi->counters[FIRQ_CNTR_IO] -= 1; in get_io_int()
1256 if (fi->counters[FIRQ_CNTR_VIRTIO] >= KVM_S390_MAX_VIRTIO_IRQS) { in __inject_virtio()
1260 fi->counters[FIRQ_CNTR_VIRTIO] += 1; in __inject_virtio()
1273 if (fi->counters[FIRQ_CNTR_PFAULT] >= in __inject_pfault_done()
1278 fi->counters[FIRQ_CNTR_PFAULT] += 1; in __inject_pfault_done()
1308 if (fi->counters[FIRQ_CNTR_IO] >= KVM_S390_MAX_FLOAT_IRQS) { in __inject_io()
1312 fi->counters[FIRQ_CNTR_IO] += 1; in __inject_io()
[all …]
/arch/x86/include/asm/xen/
Dinterface.h260 uint32_t counters; member
/arch/x86/xen/
Dpmu.c267 counter_regs = field_offset(ctxt, counters); in xen_amd_pmu_emulate()
345 counter_regs = field_offset(ctxt, counters); in xen_amd_read_pmc()
/arch/powerpc/boot/dts/
Da4m072.dts39 fsl,init-fd-counters = <0x3333>;
/arch/mips/include/asm/mach-cavium-octeon/
Dkernel-entry-init.h41 # counters interrupt to IRQ 6
/arch/s390/include/asm/
Dkvm_host.h459 int counters[FIRQ_MAX_COUNT]; member
/arch/m68k/fpsp040/
Dskeleton.S48 | The following counters are used for standalone testing
/arch/
DKconfig26 The number of hardware counters is limited. The multiplexing
27 feature enables OProfile to gather more events than counters
/arch/parisc/kernel/
Dperf_asm.S41 ; Enable the performance counters
/arch/arc/
DKconfig524 The counters can be cleared via Debugfs as well
/arch/x86/
DKconfig947 performance counters), and the NMI watchdog which detects hard