Home
last modified time | relevance | path

Searched refs:num_counters (Results 1 – 25 of 93) sorted by relevance

1234

/kernel/linux/linux-5.10/arch/x86/oprofile/
Dop_model_ppro.c26 static int num_counters = 2; variable
37 for (i = 0; i < num_counters; ++i) { in ppro_shutdown()
49 for (i = 0; i < num_counters; i++) { in ppro_fill_in_addresses()
96 for (i = 0; i < num_counters; ++i) { in ppro_setup_ctrs()
112 for (i = 0; i < num_counters; ++i) { in ppro_setup_ctrs()
133 for (i = 0; i < num_counters; ++i) { in ppro_check_ctrs()
163 for (i = 0; i < num_counters; ++i) { in ppro_start()
178 for (i = 0; i < num_counters; ++i) { in ppro_stop()
188 .num_counters = 2,
218 eax.split.num_counters = 2; in arch_perfmon_setup_counters()
[all …]
Dop_model_amd.c42 static int num_counters; variable
275 for (i = 0; i < num_counters; ++i) { in op_mux_switch_ctrl()
294 for (i = 0; i < num_counters; ++i) { in op_amd_shutdown()
306 for (i = 0; i < num_counters; i++) { in op_amd_fill_in_addresses()
314 if (num_counters == AMD64_NUM_COUNTERS_CORE) { in op_amd_fill_in_addresses()
349 for (i = 0; i < num_counters; ++i) { in op_amd_setup_ctrs()
365 for (i = 0; i < num_counters; ++i) { in op_amd_setup_ctrs()
387 for (i = 0; i < num_counters; ++i) { in op_amd_check_ctrs()
410 for (i = 0; i < num_counters; ++i) { in op_amd_start()
430 for (i = 0; i < num_counters; ++i) { in op_amd_stop()
[all …]
Dop_model_p4.c37 static unsigned int num_counters = NUM_COUNTERS_NON_HT; variable
47 num_counters = NUM_COUNTERS_HT2; in setup_num_counters()
384 #define VIRT_CTR(stagger, i) ((i) + ((num_counters) * (stagger)))
392 for (i = 0; i < num_counters; ++i) { in p4_shutdown()
401 for (i = num_counters; i < num_controls; ++i) { in p4_shutdown()
416 for (i = 0; i < num_counters; ++i) { in p4_fill_in_addresses()
468 if (num_counters == NUM_COUNTERS_NON_HT) { in p4_fill_in_addresses()
490 for (i = 0; i < num_counters; ++i) { in p4_fill_in_addresses()
585 for (i = 0; i < num_counters; i++) { in p4_setup_ctrs()
595 for (i = num_counters; i < num_controls; i++) { in p4_setup_ctrs()
[all …]
Dnmi_int.c81 for (i = 0; i < model->num_counters; ++i) { in nmi_cpu_save_registers()
145 return virt % model->num_counters; in op_x86_virt_to_phys()
206 for (i = 0; i < model->num_counters; ++i) { in nmi_cpu_save_mpx_registers()
219 for (i = 0; i < model->num_counters; ++i) { in nmi_cpu_restore_mpx_registers()
236 si += model->num_counters; in nmi_cpu_switch()
256 return counter_config[model->num_counters].count ? 0 : -EINVAL; in nmi_multiplex_on()
318 size_t counters_size = sizeof(struct op_msr) * model->num_counters; in allocate_msrs()
367 for (i = 0; i < model->num_counters; ++i) { in nmi_cpu_restore_registers()
468 sizeof(struct op_msr) * model->num_counters); in nmi_setup()
767 model->num_virt_counters = model->num_counters; in op_nmi_init()
/kernel/linux/linux-5.10/drivers/oprofile/
Doprofile_perf.c31 static int num_counters; variable
42 for (id = 0; id < num_counters; ++id) in op_overflow_handler()
46 if (id != num_counters) in op_overflow_handler()
64 for (i = 0; i < num_counters; ++i) { in op_perf_setup()
120 for (event = 0; event < num_counters; ++event) { in op_perf_start()
138 for (event = 0; event < num_counters; ++event) in op_perf_stop()
146 for (i = 0; i < num_counters; i++) { in oprofile_perf_create_files()
261 for (id = 0; id < num_counters; ++id) { in oprofile_perf_exit()
282 num_counters = perf_num_counters(); in oprofile_perf_init()
283 if (num_counters <= 0) { in oprofile_perf_init()
[all …]
/kernel/linux/linux-5.10/arch/mips/oprofile/
Dop_model_mipsxx.c137 unsigned int counters = op_model_mipsxx_ops.num_counters; in mipsxx_reg_setup()
166 unsigned int counters = op_model_mipsxx_ops.num_counters; in mipsxx_cpu_setup()
193 unsigned int counters = op_model_mipsxx_ops.num_counters; in mipsxx_cpu_start()
216 unsigned int counters = op_model_mipsxx_ops.num_counters; in mipsxx_cpu_stop()
238 unsigned int counters = op_model_mipsxx_ops.num_counters; in mipsxx_perfcount_handler()
348 op_model_mipsxx_ops.num_counters = counters; in mipsxx_init()
461 int counters = op_model_mipsxx_ops.num_counters; in mipsxx_exit()
/kernel/linux/linux-5.10/arch/x86/events/intel/
Duncore_snbep.c865 .num_counters = 2,
1022 .num_counters = 4,
1132 .num_counters = 4,
1225 .num_counters = 4,
1233 .num_counters = 4,
1245 .num_counters = 4,
1261 .num_counters = 4,
1270 .num_counters = 3,
1593 .num_counters = 2,
1724 .num_counters = 4,
[all …]
Duncore_snb.c212 .num_counters = 2,
230 .num_counters = 2,
290 .num_counters = 4,
328 .num_counters = 2,
355 .num_counters = 1,
369 .num_counters = 1,
700 .num_counters = 5,
1155 .num_counters = 8,
1315 .num_counters = 3,
/kernel/linux/linux-5.10/arch/powerpc/oprofile/
Dop_model_fsl_emb.c24 static int num_counters; variable
262 for (i = 0;i < num_counters;i++) { in fsl_emb_cpu_setup()
279 num_counters = num_ctrs; in fsl_emb_reg_setup()
286 for (i = 0; i < num_counters; ++i) in fsl_emb_reg_setup()
298 for (i = 0; i < num_counters; ++i) { in fsl_emb_start()
351 for (i = 0; i < num_counters; ++i) { in fsl_emb_handle_interrupt()
Dcommon.c56 op_per_cpu_rc = model->reg_setup(ctr, &sys, model->num_counters); in op_powerpc_setup()
154 for (i = 0; i < model->num_counters; ++i) { in op_powerpc_create_files()
225 model->num_counters = cur_cpu_spec->num_pmcs; in oprofile_arch_init()
Dop_model_cell.c179 static int num_counters; variable
479 for (i = 0; i < num_counters; i++) in cell_virtual_cntr()
498 for (i = 0; i < num_counters; i++) { in cell_virtual_cntr()
527 for (i = 0; i < num_counters; i++) { in cell_virtual_cntr()
735 num_counters = 1; /* Only support one SPU event at a time */ in cell_reg_setup_spu_events()
756 num_counters = num_ctrs; in cell_reg_setup_ppu()
809 for (i = 0; i < num_counters; ++i) { in cell_reg_setup_ppu()
825 for (i = 0; i < num_counters; ++i) { in cell_reg_setup_ppu()
921 for (i = 0; i < num_counters; ++i) { in cell_cpu_setup()
1409 for (i = 0; i < num_counters; ++i) { in cell_global_start_ppu()
[all …]
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/
Dcounters.c251 int ret, num_counters; in mlx5_ib_get_hw_stats() local
257 num_counters = cnts->num_q_counters + in mlx5_ib_get_hw_stats()
295 return num_counters; in mlx5_ib_get_hw_stats()
434 u32 num_counters; in __mlx5_ib_alloc_counters() local
436 num_counters = ARRAY_SIZE(basic_q_cnts); in __mlx5_ib_alloc_counters()
439 num_counters += ARRAY_SIZE(out_of_seq_q_cnts); in __mlx5_ib_alloc_counters()
442 num_counters += ARRAY_SIZE(retrans_q_cnts); in __mlx5_ib_alloc_counters()
445 num_counters += ARRAY_SIZE(extended_err_cnts); in __mlx5_ib_alloc_counters()
448 num_counters += ARRAY_SIZE(roce_accl_cnts); in __mlx5_ib_alloc_counters()
450 cnts->num_q_counters = num_counters; in __mlx5_ib_alloc_counters()
[all …]
/kernel/linux/linux-5.10/kernel/gcov/
Dclang.c78 u32 num_counters; member
167 void llvm_gcda_emit_arcs(u32 num_counters, u64 *counters) in llvm_gcda_emit_arcs() argument
172 info->num_counters = num_counters; in llvm_gcda_emit_arcs()
270 sizeof(fn->counters[0]) * fn->num_counters); in gcov_info_reset()
328 for (i = 0; i < sfn_ptr->num_counters; i++) in gcov_info_add()
349 cv_size = fn->num_counters * sizeof(fn->counters[0]); in gcov_fn_info_dup()
373 cv_size = fn->num_counters * sizeof(fn->counters[0]); in gcov_fn_info_dup()
556 pos += store_gcov_u32(buffer, pos, fi_ptr->num_counters * 2); in convert_to_gcda()
557 for (i = 0; i < fi_ptr->num_counters; i++) in convert_to_gcda()
/kernel/linux/linux-5.10/drivers/perf/
Darm_dsu_pmu.c121 s8 num_counters; member
245 return (idx < dsu_pmu->num_counters) || in dsu_pmu_counter_valid()
314 idx = find_first_zero_bit(used_mask, dsu_pmu->num_counters); in dsu_pmu_get_event_idx()
315 if (idx >= dsu_pmu->num_counters) in dsu_pmu_get_event_idx()
602 dsu_pmu->num_counters = -1; in dsu_pmu_alloc()
671 u64 num_counters; in dsu_pmu_probe_pmu() local
674 num_counters = (__dsu_pmu_read_pmcr() >> CLUSTERPMCR_N_SHIFT) & in dsu_pmu_probe_pmu()
677 if (WARN_ON(num_counters > 31)) in dsu_pmu_probe_pmu()
678 num_counters = 31; in dsu_pmu_probe_pmu()
679 dsu_pmu->num_counters = num_counters; in dsu_pmu_probe_pmu()
[all …]
Darm_smmuv3_pmu.c109 unsigned int num_counters; member
278 unsigned int cur_idx, num_ctrs = smmu_pmu->num_counters; in smmu_pmu_apply_event_filter()
309 unsigned int num_ctrs = smmu_pmu->num_counters; in smmu_pmu_get_event_idx()
378 if (++group_num_events > smmu_pmu->num_counters) in smmu_pmu_event_init()
389 if (++group_num_events > smmu_pmu->num_counters) in smmu_pmu_event_init()
619 for_each_set_bit(idx, (unsigned long *)&ovsr, smmu_pmu->num_counters) { in smmu_pmu_handle_irq()
700 u64 counter_present_mask = GENMASK_ULL(smmu_pmu->num_counters - 1, 0); in smmu_pmu_reset()
785 smmu_pmu->num_counters = FIELD_GET(SMMU_PMCG_CFGR_NCTR, cfgr) + 1; in smmu_pmu_probe()
830 &res_0->start, smmu_pmu->num_counters, in smmu_pmu_probe()
Dqcom_l2_pmu.c113 int num_counters; member
344 int num_ctrs = cluster->l2cache_pmu->num_counters - 1; in l2_cache_get_event_idx()
388 int num_counters = cluster->l2cache_pmu->num_counters; in l2_cache_handle_irq() local
396 for_each_set_bit(idx, cluster->used_counters, num_counters) { in l2_cache_handle_irq()
926 l2cache_pmu->num_counters = get_num_counters(); in l2_cache_pmu_probe()
933 l2_cycle_ctr_idx = l2cache_pmu->num_counters - 1; in l2_cache_pmu_probe()
934 l2_counter_present_mask = GENMASK(l2cache_pmu->num_counters - 2, 0) | in l2_cache_pmu_probe()
/kernel/linux/linux-5.10/drivers/perf/hisilicon/
Dhisi_uncore_pmu.c96 return counters <= hisi_pmu->num_counters; in hisi_validate_event_group()
101 return idx >= 0 && idx < hisi_pmu->num_counters; in hisi_uncore_pmu_counter_valid()
109 u32 num_counters = hisi_pmu->num_counters; in hisi_uncore_pmu_get_event_idx() local
112 idx = find_first_zero_bit(used_mask, num_counters); in hisi_uncore_pmu_get_event_idx()
113 if (idx == num_counters) in hisi_uncore_pmu_get_event_idx()
338 hisi_pmu->num_counters); in hisi_uncore_pmu_enable()
/kernel/linux/linux-5.10/net/bridge/netfilter/
Debtables.c979 if (repl->num_counters) { in do_replace_finish()
980 unsigned long size = repl->num_counters * sizeof(*counterstmp); in do_replace_finish()
1007 if (repl->num_counters && repl->num_counters != t->private->nentries) { in do_replace_finish()
1022 if (repl->num_counters) in do_replace_finish()
1034 if (repl->num_counters && in do_replace_finish()
1036 repl->num_counters * sizeof(struct ebt_counter))) { in do_replace_finish()
1086 if (tmp.num_counters >= INT_MAX / sizeof(struct ebt_counter)) in do_replace()
1265 unsigned int num_counters, unsigned int len) in do_update_counters() argument
1271 if (num_counters == 0) in do_update_counters()
1274 tmp = vmalloc(array_size(num_counters, sizeof(*tmp))); in do_update_counters()
[all …]
/kernel/linux/linux-5.10/arch/powerpc/include/asm/
Doprofile_impl.h46 int num_counters);
56 int num_counters; member
/kernel/linux/linux-5.10/arch/x86/events/amd/
Dcore.c396 for (i = 0; i < x86_pmu.num_counters; i++) { in __amd_put_nb_event_constraints()
463 for_each_set_bit(idx, c->idxmsk, x86_pmu.num_counters) { in __amd_get_nb_event_constraints()
506 for (i = 0; i < x86_pmu.num_counters; i++) { in amd_alloc_nb()
629 for (idx = 0; idx < x86_pmu.num_counters; idx++) { in amd_pmu_disable_all()
922 .num_counters = AMD64_NUM_COUNTERS,
959 x86_pmu.num_counters = AMD64_NUM_COUNTERS_CORE; in amd_core_pmu_init()
978 for (i = 0; i < x86_pmu.num_counters - 1; i += 2) in amd_core_pmu_init()
983 x86_pmu.num_counters / 2, 0, in amd_core_pmu_init()
Duncore.c44 int num_counters; member
136 for (i = 0; i < uncore->num_counters; i++) { in amd_uncore_add()
145 for (i = 0; i < uncore->num_counters; i++) { in amd_uncore_add()
175 for (i = 0; i < uncore->num_counters; i++) { in amd_uncore_del()
380 uncore_nb->num_counters = num_counters_nb; in amd_uncore_cpu_up_prepare()
394 uncore_llc->num_counters = num_counters_llc; in amd_uncore_cpu_up_prepare()
/kernel/linux/linux-5.10/arch/x86/kvm/vmx/
Dpmu_intel.c143 unsigned int num_counters; in intel_rdpmc_ecx_to_pmc() local
148 num_counters = pmu->nr_arch_fixed_counters; in intel_rdpmc_ecx_to_pmc()
151 num_counters = pmu->nr_arch_gp_counters; in intel_rdpmc_ecx_to_pmc()
153 if (idx >= num_counters) in intel_rdpmc_ecx_to_pmc()
156 return &counters[array_index_nospec(idx, num_counters)]; in intel_rdpmc_ecx_to_pmc()
357 pmu->nr_arch_gp_counters = min_t(int, eax.split.num_counters, in intel_pmu_refresh()
/kernel/linux/linux-5.10/arch/x86/events/zhaoxin/
Dcore.c533 x86_pmu.num_counters = eax.split.num_counters; in zhaoxin_pmu_init()
601 x86_pmu.intel_ctrl = (1 << (x86_pmu.num_counters)) - 1; in zhaoxin_pmu_init()
606 c->idxmsk64 |= (1ULL << x86_pmu.num_counters) - 1; in zhaoxin_pmu_init()
607 c->weight += x86_pmu.num_counters; in zhaoxin_pmu_init()
/kernel/linux/linux-5.10/drivers/staging/comedi/drivers/
Dni_tio.h109 unsigned int num_counters; member
124 unsigned int num_counters,
/kernel/linux/linux-5.10/arch/x86/events/
Dcore.c185 for (i = 0; i < x86_pmu.num_counters; i++) { in reserve_pmc_hardware()
190 for (i = 0; i < x86_pmu.num_counters; i++) { in reserve_pmc_hardware()
201 i = x86_pmu.num_counters; in reserve_pmc_hardware()
214 for (i = 0; i < x86_pmu.num_counters; i++) { in release_pmc_hardware()
238 for (i = 0; i < x86_pmu.num_counters; i++) { in check_hw_exists()
656 for (idx = 0; idx < x86_pmu.num_counters; idx++) { in x86_pmu_disable_all()
707 for (idx = 0; idx < x86_pmu.num_counters; idx++) { in x86_pmu_enable_all()
1011 int gpmax = x86_pmu.num_counters; in x86_schedule_events()
1032 gpmax = x86_pmu.num_counters - cpuc->n_pair; in x86_schedule_events()
1121 max_count = x86_pmu.num_counters + x86_pmu.num_counters_fixed; in collect_events()
[all …]

1234