Home
last modified time | relevance | path

Searched refs:cpu_data (Results 1 – 25 of 148) sorted by relevance

123456

/kernel/linux/linux-5.10/drivers/cpuidle/governors/
Dteo.c119 struct teo_cpu *cpu_data = per_cpu_ptr(&teo_cpus, dev->cpu); in teo_update() local
123 if (cpu_data->time_span_ns >= cpu_data->sleep_length_ns) { in teo_update()
158 unsigned int early_hits = cpu_data->states[i].early_hits; in teo_update()
160 cpu_data->states[i].early_hits -= early_hits >> DECAY_SHIFT; in teo_update()
162 if (drv->states[i].target_residency_ns <= cpu_data->sleep_length_ns) { in teo_update()
178 unsigned int hits = cpu_data->states[idx_timer].hits; in teo_update()
179 unsigned int misses = cpu_data->states[idx_timer].misses; in teo_update()
187 cpu_data->states[idx_hit].early_hits += PULSE; in teo_update()
192 cpu_data->states[idx_timer].misses = misses; in teo_update()
193 cpu_data->states[idx_timer].hits = hits; in teo_update()
[all …]
/kernel/linux/linux-5.10/arch/mips/include/asm/
Dcpu-features.h17 #define __ase(ase) (cpu_data[0].ases & (ase))
18 #define __isa(isa) (cpu_data[0].isa_level & (isa))
19 #define __opt(opt) (cpu_data[0].options & (opt))
234 #define cpu_has_vtag_icache (cpu_data[0].icache.flags & MIPS_CACHE_VTAG)
237 #define cpu_has_dc_aliases (cpu_data[0].dcache.flags & MIPS_CACHE_ALIASES)
240 #define cpu_has_ic_fills_f_dc (cpu_data[0].icache.flags & MIPS_CACHE_IC_F_DC)
243 #define cpu_has_pindexed_dcache (cpu_data[0].dcache.flags & MIPS_CACHE_PINDEX)
259 #define cpu_icache_snoops_remote_store (cpu_data[0].icache.flags & MIPS_IC_SNOOPS_REMOTE)
444 # define cpu_has_64bits (cpu_data[0].isa_level & MIPS_CPU_ISA_64BIT)
447 # define cpu_has_64bit_zero_reg (cpu_data[0].isa_level & MIPS_CPU_ISA_64BIT)
[all …]
Dcpu-info.h119 extern struct cpuinfo_mips cpu_data[];
120 #define current_cpu_data cpu_data[smp_processor_id()]
121 #define raw_current_cpu_data cpu_data[raw_smp_processor_id()]
122 #define boot_cpu_data cpu_data[0]
183 struct cpuinfo_mips *infoa = &cpu_data[cpua]; in cpus_are_siblings()
184 struct cpuinfo_mips *infob = &cpu_data[cpub]; in cpus_are_siblings()
Dmmu_context.h96 unsigned long asid_mask = cpu_asid_mask(&cpu_data[cpu]); in asid_version_mask()
123 #define asid_cache(cpu) (cpu_data[cpu].asid_cache)
125 (cpu_context((cpu), (mm)) & cpu_asid_mask(&cpu_data[cpu]))
216 write_c0_memorymapid(ctx & cpu_asid_mask(&cpu_data[cpu])); in drop_mmu_context()
Dtopology.h15 #define topology_physical_package_id(cpu) (cpu_data[cpu].package)
16 #define topology_core_id(cpu) (cpu_core(&cpu_data[cpu]))
/kernel/linux/linux-5.10/arch/mips/kernel/
Dproc.c40 unsigned int version = cpu_data[n].processor_id; in show_cpuinfo()
41 unsigned int fp_vers = cpu_data[n].fpu_id; in show_cpuinfo()
62 cpu_data[n].options & MIPS_CPU_FPU ? " FPU V%d.%d" : ""); in show_cpuinfo()
67 cpu_data[n].udelay_val / (500000/HZ), in show_cpuinfo()
68 (cpu_data[n].udelay_val / (5000/HZ)) % 100); in show_cpuinfo()
72 seq_printf(m, "tlb_entries\t\t: %d\n", cpu_data[n].tlbsize); in show_cpuinfo()
79 cpu_data[n].watch_reg_count); in show_cpuinfo()
80 for (i = 0; i < cpu_data[n].watch_reg_count; i++) in show_cpuinfo()
82 cpu_data[n].watch_reg_masks[i]); in show_cpuinfo()
142 cpu_data[n].srsets); in show_cpuinfo()
[all …]
Dsmp-cps.c74 cpu_set_cluster(&cpu_data[nvpes + v], cl); in cps_smp_setup()
75 cpu_set_core(&cpu_data[nvpes + v], c); in cps_smp_setup()
76 cpu_set_vpe_id(&cpu_data[nvpes + v], v); in cps_smp_setup()
88 set_cpu_possible(v, cpu_cluster(&cpu_data[v]) == 0); in cps_smp_setup()
89 set_cpu_present(v, cpu_cluster(&cpu_data[v]) == 0); in cps_smp_setup()
296 unsigned core = cpu_core(&cpu_data[cpu]); in cps_boot_secondary()
297 unsigned vpe_id = cpu_vpe_id(&cpu_data[cpu]); in cps_boot_secondary()
305 if (cpu_cluster(&cpu_data[cpu]) != cpu_cluster(&raw_current_cpu_data)) in cps_boot_secondary()
312 atomic_or(1 << cpu_vpe_id(&cpu_data[cpu]), &core_cfg->vpe_mask); in cps_boot_secondary()
409 core = cpu_core(&cpu_data[cpu]); in cps_shutdown_this_cpu()
[all …]
/kernel/linux/linux-5.10/arch/sparc/kernel/
Dprom_64.c416 int proc_id = cpu_data(cpu).proc_id; in arch_find_n_match_cpu_physical_id()
530 cpu_data(cpuid).clock_tick = in fill_in_one_cpu()
534 cpu_data(cpuid).dcache_size = in fill_in_one_cpu()
537 cpu_data(cpuid).dcache_line_size = in fill_in_one_cpu()
540 cpu_data(cpuid).icache_size = in fill_in_one_cpu()
543 cpu_data(cpuid).icache_line_size = in fill_in_one_cpu()
546 cpu_data(cpuid).ecache_size = in fill_in_one_cpu()
548 cpu_data(cpuid).ecache_line_size = in fill_in_one_cpu()
550 if (!cpu_data(cpuid).ecache_size || in fill_in_one_cpu()
551 !cpu_data(cpuid).ecache_line_size) { in fill_in_one_cpu()
[all …]
Dsmp_32.c61 cpu_data(id).udelay_val = loops_per_jiffy; in smp_store_cpu_info()
64 cpu_data(id).clock_tick = prom_getintdefault(cpu_node, in smp_store_cpu_info()
66 cpu_data(id).prom_node = cpu_node; in smp_store_cpu_info()
73 cpu_data(id).mid = mid; in smp_store_cpu_info()
83 bogosum += cpu_data(cpu).udelay_val; in smp_cpus_done()
382 cpu_data(i).udelay_val/(500000/HZ), in smp_bogo()
383 (cpu_data(i).udelay_val/(5000/HZ))%100); in smp_bogo()
/kernel/linux/linux-5.10/arch/ia64/mm/
Dcontig.c35 static void *cpu_data; variable
59 memcpy(cpu_data, src, __per_cpu_end - __per_cpu_start); in per_cpu_init()
60 __per_cpu_offset[cpu] = (char *)cpu_data - __per_cpu_start; in per_cpu_init()
73 ia64_set_kr(IA64_KR_PER_CPU_DATA, __pa(cpu_data) - in per_cpu_init()
76 cpu_data += PERCPU_PAGE_SIZE; in per_cpu_init()
87 cpu_data = memblock_alloc_from(size, PERCPU_PAGE_SIZE, in alloc_per_cpu_data()
89 if (!cpu_data) in alloc_per_cpu_data()
Ddiscontig.c135 static void *per_cpu_node_setup(void *cpu_data, int node) in per_cpu_node_setup() argument
146 memcpy(__va(cpu_data), src, __per_cpu_end - __per_cpu_start); in per_cpu_node_setup()
147 __per_cpu_offset[cpu] = (char *)__va(cpu_data) - in per_cpu_node_setup()
161 (unsigned long)cpu_data - in per_cpu_node_setup()
164 cpu_data += PERCPU_PAGE_SIZE; in per_cpu_node_setup()
167 return cpu_data; in per_cpu_node_setup()
261 void *cpu_data; in fill_pernode() local
268 cpu_data = (void *)pernode; in fill_pernode()
279 cpu_data = per_cpu_node_setup(cpu_data, node); in fill_pernode()
/kernel/linux/linux-5.10/kernel/trace/
Dtrace_functions_graph.c31 struct fgraph_cpu_data __percpu *cpu_data; member
393 last_pid = &(per_cpu_ptr(data->cpu_data, cpu)->last_pid); in verif_pid()
644 struct fgraph_cpu_data *cpu_data; in print_graph_entry_leaf() local
646 cpu_data = per_cpu_ptr(data->cpu_data, cpu); in print_graph_entry_leaf()
653 cpu_data->depth = call->depth - 1; in print_graph_entry_leaf()
658 cpu_data->enter_funcs[call->depth] = 0; in print_graph_entry_leaf()
687 struct fgraph_cpu_data *cpu_data; in print_graph_entry_nested() local
690 cpu_data = per_cpu_ptr(data->cpu_data, cpu); in print_graph_entry_nested()
691 cpu_data->depth = call->depth; in print_graph_entry_nested()
696 cpu_data->enter_funcs[call->depth] = call->func; in print_graph_entry_nested()
[all …]
/kernel/linux/linux-5.10/arch/x86/include/asm/
Dtopology.h107 #define topology_logical_package_id(cpu) (cpu_data(cpu).logical_proc_id)
108 #define topology_physical_package_id(cpu) (cpu_data(cpu).phys_proc_id)
109 #define topology_logical_die_id(cpu) (cpu_data(cpu).logical_die_id)
110 #define topology_die_id(cpu) (cpu_data(cpu).cpu_die_id)
111 #define topology_core_id(cpu) (cpu_data(cpu).cpu_core_id)
/kernel/linux/linux-5.10/arch/ia64/kernel/
Dsmpboot.c410 last_cpuinfo = cpu_data(cpuid - 1); in smp_callin()
580 if (cpu_data(cpu)->threads_per_core == 1 && in remove_siblinginfo()
581 cpu_data(cpu)->cores_per_socket == 1) { in remove_siblinginfo()
690 bogosum += cpu_data(cpu)->loops_per_jiffy; in smp_cpus_done()
702 if ((cpu_data(cpu)->socket_id == cpu_data(i)->socket_id)) { in set_cpu_sibling_map()
705 if (cpu_data(cpu)->core_id == cpu_data(i)->core_id) { in set_cpu_sibling_map()
738 if (cpu_data(cpu)->threads_per_core == 1 && in __cpu_up()
739 cpu_data(cpu)->cores_per_socket == 1) { in __cpu_up()
834 if ((cpu_data(j)->socket_id == cpu_data(i)->socket_id)) { in is_multithreading_enabled()
835 if (cpu_data(j)->core_id == cpu_data(i)->core_id) in is_multithreading_enabled()
Dtopology.c35 if (cpu_data(num)->socket_id == -1) in arch_fix_phys_package_id()
36 cpu_data(num)->socket_id = slot; in arch_fix_phys_package_id()
145 if (cpu_data(cpu)->threads_per_core <= 1 && in cache_shared_cpu_map_setup()
146 cpu_data(cpu)->cores_per_socket <= 1) { in cache_shared_cpu_map_setup()
160 if (cpu_data(cpu)->socket_id == cpu_data(j)->socket_id in cache_shared_cpu_map_setup()
161 && cpu_data(j)->core_id == csi.log1_cid in cache_shared_cpu_map_setup()
162 && cpu_data(j)->thread_id == csi.log1_tid) in cache_shared_cpu_map_setup()
/kernel/linux/linux-5.10/arch/alpha/include/asm/
Dmmu_context.h93 #define cpu_last_asn(cpuid) (cpu_data[cpuid].last_asn)
145 cpu_data[cpu].asn_lock = 1; in ev5_switch_mm()
156 cpu_data[cpu].need_new_asn = 1; in ev5_switch_mm()
191 cpu_data[cpu].asn_lock = 0; \
193 if (cpu_data[cpu].need_new_asn) { \
195 cpu_data[cpu].need_new_asn = 0; \
/kernel/linux/linux-5.10/arch/mips/loongson64/
Dsmp.c355 cpu_set_core(&cpu_data[cpu], in loongson3_init_secondary()
357 cpu_data[cpu].package = in loongson3_init_secondary()
370 if (cpu_data[cpu].package) in loongson3_init_secondary()
423 cpu_set_core(&cpu_data[0], in loongson3_smp_setup()
425 cpu_data[0].package = cpu_logical_map(0) / loongson_sysconf.cores_per_package; in loongson3_smp_setup()
528 [sets] "r" (cpu_data[smp_processor_id()].dcache.sets)); in loongson3_type1_play_dead()
590 [sets] "r" (cpu_data[smp_processor_id()].dcache.sets)); in loongson3_type2_play_dead()
674 [sets] "r" (cpu_data[smp_processor_id()].dcache.sets), in loongson3_type3_play_dead()
675 [vsets] "r" (cpu_data[smp_processor_id()].vcache.sets)); in loongson3_type3_play_dead()
754 uint64_t core_id = cpu_core(&cpu_data[cpu]); in loongson3_disable_clock()
[all …]
/kernel/linux/linux-5.10/arch/mips/mm/
Dcontext.c36 if (!((asid += cpu_asid_inc()) & cpu_asid_mask(&cpu_data[cpu]))) { in get_new_mmu_context()
76 mmid = xchg_relaxed(&cpu_data[cpu].asid_cache, 0); in flush_context()
88 __set_bit(mmid & cpu_asid_mask(&cpu_data[cpu]), mmid_map); in flush_context()
210 old_active_mmid = READ_ONCE(cpu_data[cpu].asid_cache); in check_switch_mmu_context()
213 !cmpxchg_relaxed(&cpu_data[cpu].asid_cache, old_active_mmid, ctx)) { in check_switch_mmu_context()
220 WRITE_ONCE(cpu_data[cpu].asid_cache, ctx); in check_switch_mmu_context()
/kernel/linux/linux-5.10/arch/alpha/kernel/
Dsmp.c54 struct cpuinfo_alpha cpu_data[NR_CPUS]; variable
55 EXPORT_SYMBOL(cpu_data);
82 cpu_data[cpuid].loops_per_jiffy = loops_per_jiffy; in smp_store_cpu_info()
83 cpu_data[cpuid].last_asn = ASN_FIRST_VERSION; in smp_store_cpu_info()
84 cpu_data[cpuid].need_new_asn = 0; in smp_store_cpu_info()
85 cpu_data[cpuid].asn_lock = 0; in smp_store_cpu_info()
94 cpu_data[cpuid].prof_counter = 1; in smp_setup_percpu_timer()
95 cpu_data[cpuid].prof_multiplier = 1; in smp_setup_percpu_timer()
491 bogosum += cpu_data[cpu].loops_per_jiffy; in smp_cpus_done()
564 cpu_data[this_cpu].ipi_count++; in handle_ipi()
[all …]
/kernel/linux/linux-5.10/arch/sh/include/asm/
Dprocessor.h90 extern struct sh_cpuinfo cpu_data[];
91 #define boot_cpu_data cpu_data[0]
92 #define current_cpu_data cpu_data[smp_processor_id()]
93 #define raw_current_cpu_data cpu_data[raw_smp_processor_id()]
/kernel/linux/linux-5.10/tools/testing/selftests/bpf/prog_tests/
Dperf_buffer.c19 int cpu_data = *(int *)data, duration = 0; in on_sample() local
22 if (cpu_data != cpu) in on_sample()
23 CHECK(cpu_data != cpu, "check_cpu_data", in on_sample()
24 "cpu_data %d != cpu %d\n", cpu_data, cpu); in on_sample()
/kernel/linux/linux-5.10/arch/s390/pci/
Dpci_irq.c162 struct cpu_irq_data *cpu_data; in zpci_handle_fallback_irq() local
178 cpu_data = &per_cpu(irq_data, cpu); in zpci_handle_fallback_irq()
179 if (atomic_inc_return(&cpu_data->scheduled) > 1) in zpci_handle_fallback_irq()
182 cpu_data->csd.func = zpci_handle_remote_irq; in zpci_handle_fallback_irq()
183 cpu_data->csd.info = &cpu_data->scheduled; in zpci_handle_fallback_irq()
184 cpu_data->csd.flags = 0; in zpci_handle_fallback_irq()
185 smp_call_function_single_async(cpu, &cpu_data->csd); in zpci_handle_fallback_irq()
/kernel/linux/linux-5.10/kernel/sched/
Dcore_ctl.c55 struct cpu_data { struct
65 static DEFINE_PER_CPU(struct cpu_data, cpu_state); argument
160 struct cpu_data *c; in show_global_state()
512 static bool is_active(const struct cpu_data *state) in is_active()
527 struct cpu_data *c; in eval_need()
659 struct cpu_data *c; in core_ctl_check()
693 static void move_cpu_lru(struct cpu_data *cpu_data) in move_cpu_lru() argument
698 list_del(&cpu_data->sib); in move_cpu_lru()
699 list_add_tail(&cpu_data->sib, &cpu_data->cluster->lru); in move_cpu_lru()
705 struct cpu_data *c, *tmp; in try_to_isolate()
[all …]
/kernel/linux/linux-5.10/arch/x86/kernel/
Dsmpboot.c198 cpu_data(cpuid).loops_per_jiffy = loops_per_jiffy; in smp_callin()
296 struct cpuinfo_x86 *c = &cpu_data(cpu); in topology_phys_to_logical_pkg()
312 int proc_id = cpu_data(cur_cpu).phys_proc_id; in topology_phys_to_logical_die()
315 struct cpuinfo_x86 *c = &cpu_data(cpu); in topology_phys_to_logical_die()
345 cpu_data(cpu).logical_proc_id = new; in topology_update_package_map()
368 cpu_data(cpu).logical_die_id = new; in topology_update_die_map()
375 struct cpuinfo_x86 *c = &cpu_data(id); in smp_store_boot_cpu_info()
390 struct cpuinfo_x86 *c = &cpu_data(id); in smp_store_cpu_info()
574 struct cpuinfo_x86 *c = &cpu_data(cpu); in set_cpu_sibling_map()
590 o = &cpu_data(i); in set_cpu_sibling_map()
[all …]
/kernel/linux/linux-5.10/arch/arm64/kernel/
Dcpuinfo.c33 DEFINE_PER_CPU(struct cpuinfo_arm64, cpu_data);
145 struct cpuinfo_arm64 *cpuinfo = &per_cpu(cpu_data, i); in c_show()
276 struct cpuinfo_arm64 *info = &per_cpu(cpu_data, cpu); in cpuid_cpu_online()
296 struct cpuinfo_arm64 *info = &per_cpu(cpu_data, cpu); in cpuid_cpu_offline()
314 struct cpuinfo_arm64 *info = &per_cpu(cpu_data, cpu); in cpuinfo_regs_init()
413 struct cpuinfo_arm64 *info = this_cpu_ptr(&cpu_data); in cpuinfo_store_cpu()
420 struct cpuinfo_arm64 *info = &per_cpu(cpu_data, 0); in cpuinfo_store_boot_cpu()

123456