/kernel/linux/linux-5.10/arch/x86/oprofile/ |
D | op_model_ppro.c | 26 static int num_counters = 2; variable 37 for (i = 0; i < num_counters; ++i) { in ppro_shutdown() 49 for (i = 0; i < num_counters; i++) { in ppro_fill_in_addresses() 96 for (i = 0; i < num_counters; ++i) { in ppro_setup_ctrs() 112 for (i = 0; i < num_counters; ++i) { in ppro_setup_ctrs() 133 for (i = 0; i < num_counters; ++i) { in ppro_check_ctrs() 163 for (i = 0; i < num_counters; ++i) { in ppro_start() 178 for (i = 0; i < num_counters; ++i) { in ppro_stop() 188 .num_counters = 2, 218 eax.split.num_counters = 2; in arch_perfmon_setup_counters() [all …]
|
D | op_model_amd.c | 42 static int num_counters; variable 275 for (i = 0; i < num_counters; ++i) { in op_mux_switch_ctrl() 294 for (i = 0; i < num_counters; ++i) { in op_amd_shutdown() 306 for (i = 0; i < num_counters; i++) { in op_amd_fill_in_addresses() 314 if (num_counters == AMD64_NUM_COUNTERS_CORE) { in op_amd_fill_in_addresses() 349 for (i = 0; i < num_counters; ++i) { in op_amd_setup_ctrs() 365 for (i = 0; i < num_counters; ++i) { in op_amd_setup_ctrs() 387 for (i = 0; i < num_counters; ++i) { in op_amd_check_ctrs() 410 for (i = 0; i < num_counters; ++i) { in op_amd_start() 430 for (i = 0; i < num_counters; ++i) { in op_amd_stop() [all …]
|
D | op_model_p4.c | 37 static unsigned int num_counters = NUM_COUNTERS_NON_HT; variable 47 num_counters = NUM_COUNTERS_HT2; in setup_num_counters() 384 #define VIRT_CTR(stagger, i) ((i) + ((num_counters) * (stagger))) 392 for (i = 0; i < num_counters; ++i) { in p4_shutdown() 401 for (i = num_counters; i < num_controls; ++i) { in p4_shutdown() 416 for (i = 0; i < num_counters; ++i) { in p4_fill_in_addresses() 468 if (num_counters == NUM_COUNTERS_NON_HT) { in p4_fill_in_addresses() 490 for (i = 0; i < num_counters; ++i) { in p4_fill_in_addresses() 585 for (i = 0; i < num_counters; i++) { in p4_setup_ctrs() 595 for (i = num_counters; i < num_controls; i++) { in p4_setup_ctrs() [all …]
|
D | nmi_int.c | 81 for (i = 0; i < model->num_counters; ++i) { in nmi_cpu_save_registers() 145 return virt % model->num_counters; in op_x86_virt_to_phys() 206 for (i = 0; i < model->num_counters; ++i) { in nmi_cpu_save_mpx_registers() 219 for (i = 0; i < model->num_counters; ++i) { in nmi_cpu_restore_mpx_registers() 236 si += model->num_counters; in nmi_cpu_switch() 256 return counter_config[model->num_counters].count ? 0 : -EINVAL; in nmi_multiplex_on() 318 size_t counters_size = sizeof(struct op_msr) * model->num_counters; in allocate_msrs() 367 for (i = 0; i < model->num_counters; ++i) { in nmi_cpu_restore_registers() 468 sizeof(struct op_msr) * model->num_counters); in nmi_setup() 767 model->num_virt_counters = model->num_counters; in op_nmi_init()
|
/kernel/linux/linux-5.10/drivers/oprofile/ |
D | oprofile_perf.c | 31 static int num_counters; variable 42 for (id = 0; id < num_counters; ++id) in op_overflow_handler() 46 if (id != num_counters) in op_overflow_handler() 64 for (i = 0; i < num_counters; ++i) { in op_perf_setup() 120 for (event = 0; event < num_counters; ++event) { in op_perf_start() 138 for (event = 0; event < num_counters; ++event) in op_perf_stop() 146 for (i = 0; i < num_counters; i++) { in oprofile_perf_create_files() 261 for (id = 0; id < num_counters; ++id) { in oprofile_perf_exit() 282 num_counters = perf_num_counters(); in oprofile_perf_init() 283 if (num_counters <= 0) { in oprofile_perf_init() [all …]
|
/kernel/linux/linux-5.10/arch/mips/oprofile/ |
D | op_model_mipsxx.c | 137 unsigned int counters = op_model_mipsxx_ops.num_counters; in mipsxx_reg_setup() 166 unsigned int counters = op_model_mipsxx_ops.num_counters; in mipsxx_cpu_setup() 193 unsigned int counters = op_model_mipsxx_ops.num_counters; in mipsxx_cpu_start() 216 unsigned int counters = op_model_mipsxx_ops.num_counters; in mipsxx_cpu_stop() 238 unsigned int counters = op_model_mipsxx_ops.num_counters; in mipsxx_perfcount_handler() 348 op_model_mipsxx_ops.num_counters = counters; in mipsxx_init() 461 int counters = op_model_mipsxx_ops.num_counters; in mipsxx_exit()
|
/kernel/linux/linux-5.10/arch/x86/events/intel/ |
D | uncore_snbep.c | 865 .num_counters = 2, 1022 .num_counters = 4, 1132 .num_counters = 4, 1225 .num_counters = 4, 1233 .num_counters = 4, 1245 .num_counters = 4, 1261 .num_counters = 4, 1270 .num_counters = 3, 1593 .num_counters = 2, 1724 .num_counters = 4, [all …]
|
D | uncore_snb.c | 212 .num_counters = 2, 230 .num_counters = 2, 290 .num_counters = 4, 328 .num_counters = 2, 355 .num_counters = 1, 369 .num_counters = 1, 700 .num_counters = 5, 1155 .num_counters = 8, 1315 .num_counters = 3,
|
/kernel/linux/linux-5.10/arch/powerpc/oprofile/ |
D | op_model_fsl_emb.c | 24 static int num_counters; variable 262 for (i = 0;i < num_counters;i++) { in fsl_emb_cpu_setup() 279 num_counters = num_ctrs; in fsl_emb_reg_setup() 286 for (i = 0; i < num_counters; ++i) in fsl_emb_reg_setup() 298 for (i = 0; i < num_counters; ++i) { in fsl_emb_start() 351 for (i = 0; i < num_counters; ++i) { in fsl_emb_handle_interrupt()
|
D | common.c | 56 op_per_cpu_rc = model->reg_setup(ctr, &sys, model->num_counters); in op_powerpc_setup() 154 for (i = 0; i < model->num_counters; ++i) { in op_powerpc_create_files() 225 model->num_counters = cur_cpu_spec->num_pmcs; in oprofile_arch_init()
|
D | op_model_cell.c | 179 static int num_counters; variable 479 for (i = 0; i < num_counters; i++) in cell_virtual_cntr() 498 for (i = 0; i < num_counters; i++) { in cell_virtual_cntr() 527 for (i = 0; i < num_counters; i++) { in cell_virtual_cntr() 735 num_counters = 1; /* Only support one SPU event at a time */ in cell_reg_setup_spu_events() 756 num_counters = num_ctrs; in cell_reg_setup_ppu() 809 for (i = 0; i < num_counters; ++i) { in cell_reg_setup_ppu() 825 for (i = 0; i < num_counters; ++i) { in cell_reg_setup_ppu() 921 for (i = 0; i < num_counters; ++i) { in cell_cpu_setup() 1409 for (i = 0; i < num_counters; ++i) { in cell_global_start_ppu() [all …]
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/ |
D | counters.c | 251 int ret, num_counters; in mlx5_ib_get_hw_stats() local 257 num_counters = cnts->num_q_counters + in mlx5_ib_get_hw_stats() 295 return num_counters; in mlx5_ib_get_hw_stats() 434 u32 num_counters; in __mlx5_ib_alloc_counters() local 436 num_counters = ARRAY_SIZE(basic_q_cnts); in __mlx5_ib_alloc_counters() 439 num_counters += ARRAY_SIZE(out_of_seq_q_cnts); in __mlx5_ib_alloc_counters() 442 num_counters += ARRAY_SIZE(retrans_q_cnts); in __mlx5_ib_alloc_counters() 445 num_counters += ARRAY_SIZE(extended_err_cnts); in __mlx5_ib_alloc_counters() 448 num_counters += ARRAY_SIZE(roce_accl_cnts); in __mlx5_ib_alloc_counters() 450 cnts->num_q_counters = num_counters; in __mlx5_ib_alloc_counters() [all …]
|
/kernel/linux/linux-5.10/kernel/gcov/ |
D | clang.c | 78 u32 num_counters; member 167 void llvm_gcda_emit_arcs(u32 num_counters, u64 *counters) in llvm_gcda_emit_arcs() argument 172 info->num_counters = num_counters; in llvm_gcda_emit_arcs() 270 sizeof(fn->counters[0]) * fn->num_counters); in gcov_info_reset() 328 for (i = 0; i < sfn_ptr->num_counters; i++) in gcov_info_add() 349 cv_size = fn->num_counters * sizeof(fn->counters[0]); in gcov_fn_info_dup() 373 cv_size = fn->num_counters * sizeof(fn->counters[0]); in gcov_fn_info_dup() 556 pos += store_gcov_u32(buffer, pos, fi_ptr->num_counters * 2); in convert_to_gcda() 557 for (i = 0; i < fi_ptr->num_counters; i++) in convert_to_gcda()
|
/kernel/linux/linux-5.10/drivers/perf/ |
D | arm_dsu_pmu.c | 121 s8 num_counters; member 245 return (idx < dsu_pmu->num_counters) || in dsu_pmu_counter_valid() 314 idx = find_first_zero_bit(used_mask, dsu_pmu->num_counters); in dsu_pmu_get_event_idx() 315 if (idx >= dsu_pmu->num_counters) in dsu_pmu_get_event_idx() 602 dsu_pmu->num_counters = -1; in dsu_pmu_alloc() 671 u64 num_counters; in dsu_pmu_probe_pmu() local 674 num_counters = (__dsu_pmu_read_pmcr() >> CLUSTERPMCR_N_SHIFT) & in dsu_pmu_probe_pmu() 677 if (WARN_ON(num_counters > 31)) in dsu_pmu_probe_pmu() 678 num_counters = 31; in dsu_pmu_probe_pmu() 679 dsu_pmu->num_counters = num_counters; in dsu_pmu_probe_pmu() [all …]
|
D | arm_smmuv3_pmu.c | 109 unsigned int num_counters; member 278 unsigned int cur_idx, num_ctrs = smmu_pmu->num_counters; in smmu_pmu_apply_event_filter() 309 unsigned int num_ctrs = smmu_pmu->num_counters; in smmu_pmu_get_event_idx() 378 if (++group_num_events > smmu_pmu->num_counters) in smmu_pmu_event_init() 389 if (++group_num_events > smmu_pmu->num_counters) in smmu_pmu_event_init() 619 for_each_set_bit(idx, (unsigned long *)&ovsr, smmu_pmu->num_counters) { in smmu_pmu_handle_irq() 700 u64 counter_present_mask = GENMASK_ULL(smmu_pmu->num_counters - 1, 0); in smmu_pmu_reset() 785 smmu_pmu->num_counters = FIELD_GET(SMMU_PMCG_CFGR_NCTR, cfgr) + 1; in smmu_pmu_probe() 830 &res_0->start, smmu_pmu->num_counters, in smmu_pmu_probe()
|
D | qcom_l2_pmu.c | 113 int num_counters; member 344 int num_ctrs = cluster->l2cache_pmu->num_counters - 1; in l2_cache_get_event_idx() 388 int num_counters = cluster->l2cache_pmu->num_counters; in l2_cache_handle_irq() local 396 for_each_set_bit(idx, cluster->used_counters, num_counters) { in l2_cache_handle_irq() 926 l2cache_pmu->num_counters = get_num_counters(); in l2_cache_pmu_probe() 933 l2_cycle_ctr_idx = l2cache_pmu->num_counters - 1; in l2_cache_pmu_probe() 934 l2_counter_present_mask = GENMASK(l2cache_pmu->num_counters - 2, 0) | in l2_cache_pmu_probe()
|
/kernel/linux/linux-5.10/drivers/perf/hisilicon/ |
D | hisi_uncore_pmu.c | 96 return counters <= hisi_pmu->num_counters; in hisi_validate_event_group() 101 return idx >= 0 && idx < hisi_pmu->num_counters; in hisi_uncore_pmu_counter_valid() 109 u32 num_counters = hisi_pmu->num_counters; in hisi_uncore_pmu_get_event_idx() local 112 idx = find_first_zero_bit(used_mask, num_counters); in hisi_uncore_pmu_get_event_idx() 113 if (idx == num_counters) in hisi_uncore_pmu_get_event_idx() 338 hisi_pmu->num_counters); in hisi_uncore_pmu_enable()
|
/kernel/linux/linux-5.10/net/bridge/netfilter/ |
D | ebtables.c | 979 if (repl->num_counters) { in do_replace_finish() 980 unsigned long size = repl->num_counters * sizeof(*counterstmp); in do_replace_finish() 1007 if (repl->num_counters && repl->num_counters != t->private->nentries) { in do_replace_finish() 1022 if (repl->num_counters) in do_replace_finish() 1034 if (repl->num_counters && in do_replace_finish() 1036 repl->num_counters * sizeof(struct ebt_counter))) { in do_replace_finish() 1086 if (tmp.num_counters >= INT_MAX / sizeof(struct ebt_counter)) in do_replace() 1265 unsigned int num_counters, unsigned int len) in do_update_counters() argument 1271 if (num_counters == 0) in do_update_counters() 1274 tmp = vmalloc(array_size(num_counters, sizeof(*tmp))); in do_update_counters() [all …]
|
/kernel/linux/linux-5.10/arch/powerpc/include/asm/ |
D | oprofile_impl.h | 46 int num_counters); 56 int num_counters; member
|
/kernel/linux/linux-5.10/arch/x86/events/amd/ |
D | core.c | 396 for (i = 0; i < x86_pmu.num_counters; i++) { in __amd_put_nb_event_constraints() 463 for_each_set_bit(idx, c->idxmsk, x86_pmu.num_counters) { in __amd_get_nb_event_constraints() 506 for (i = 0; i < x86_pmu.num_counters; i++) { in amd_alloc_nb() 629 for (idx = 0; idx < x86_pmu.num_counters; idx++) { in amd_pmu_disable_all() 922 .num_counters = AMD64_NUM_COUNTERS, 959 x86_pmu.num_counters = AMD64_NUM_COUNTERS_CORE; in amd_core_pmu_init() 978 for (i = 0; i < x86_pmu.num_counters - 1; i += 2) in amd_core_pmu_init() 983 x86_pmu.num_counters / 2, 0, in amd_core_pmu_init()
|
D | uncore.c | 44 int num_counters; member 136 for (i = 0; i < uncore->num_counters; i++) { in amd_uncore_add() 145 for (i = 0; i < uncore->num_counters; i++) { in amd_uncore_add() 175 for (i = 0; i < uncore->num_counters; i++) { in amd_uncore_del() 380 uncore_nb->num_counters = num_counters_nb; in amd_uncore_cpu_up_prepare() 394 uncore_llc->num_counters = num_counters_llc; in amd_uncore_cpu_up_prepare()
|
/kernel/linux/linux-5.10/arch/x86/kvm/vmx/ |
D | pmu_intel.c | 143 unsigned int num_counters; in intel_rdpmc_ecx_to_pmc() local 148 num_counters = pmu->nr_arch_fixed_counters; in intel_rdpmc_ecx_to_pmc() 151 num_counters = pmu->nr_arch_gp_counters; in intel_rdpmc_ecx_to_pmc() 153 if (idx >= num_counters) in intel_rdpmc_ecx_to_pmc() 156 return &counters[array_index_nospec(idx, num_counters)]; in intel_rdpmc_ecx_to_pmc() 357 pmu->nr_arch_gp_counters = min_t(int, eax.split.num_counters, in intel_pmu_refresh()
|
/kernel/linux/linux-5.10/arch/x86/events/zhaoxin/ |
D | core.c | 533 x86_pmu.num_counters = eax.split.num_counters; in zhaoxin_pmu_init() 601 x86_pmu.intel_ctrl = (1 << (x86_pmu.num_counters)) - 1; in zhaoxin_pmu_init() 606 c->idxmsk64 |= (1ULL << x86_pmu.num_counters) - 1; in zhaoxin_pmu_init() 607 c->weight += x86_pmu.num_counters; in zhaoxin_pmu_init()
|
/kernel/linux/linux-5.10/drivers/staging/comedi/drivers/ |
D | ni_tio.h | 109 unsigned int num_counters; member 124 unsigned int num_counters,
|
/kernel/linux/linux-5.10/arch/x86/events/ |
D | core.c | 185 for (i = 0; i < x86_pmu.num_counters; i++) { in reserve_pmc_hardware() 190 for (i = 0; i < x86_pmu.num_counters; i++) { in reserve_pmc_hardware() 201 i = x86_pmu.num_counters; in reserve_pmc_hardware() 214 for (i = 0; i < x86_pmu.num_counters; i++) { in release_pmc_hardware() 238 for (i = 0; i < x86_pmu.num_counters; i++) { in check_hw_exists() 656 for (idx = 0; idx < x86_pmu.num_counters; idx++) { in x86_pmu_disable_all() 707 for (idx = 0; idx < x86_pmu.num_counters; idx++) { in x86_pmu_enable_all() 1011 int gpmax = x86_pmu.num_counters; in x86_schedule_events() 1032 gpmax = x86_pmu.num_counters - cpuc->n_pair; in x86_schedule_events() 1121 max_count = x86_pmu.num_counters + x86_pmu.num_counters_fixed; in collect_events() [all …]
|