/kernel/power/ |
D | energy_model.c | 129 debugfs_create_file("cpus", 0444, d, dev->em_pd->cpus, in em_debug_create_pd() 399 struct em_data_callback *cb, cpumask_t *cpus, in em_create_pd() argument 408 num_cpus = cpumask_weight(cpus); in em_create_pd() 420 cpumask_copy(em_span_cpus(pd), cpus); in em_create_pd() 440 for_each_cpu(cpu, cpus) { in em_create_pd() 559 struct em_data_callback *cb, cpumask_t *cpus, in em_dev_register_perf_domain() argument 581 if (!cpus) { in em_dev_register_perf_domain() 587 for_each_cpu(cpu, cpus) { in em_dev_register_perf_domain() 601 cpumask_pr_args(cpus)); in em_dev_register_perf_domain() 626 ret = em_create_pd(dev, nr_states, cb, cpus, flags); in em_dev_register_perf_domain()
|
D | power.h | 14 int cpus; member
|
/kernel/ |
D | stop_machine.c | 589 const struct cpumask *cpus) in stop_machine_cpuslocked() argument 595 .active_cpus = cpus, in stop_machine_cpuslocked() 624 int stop_machine(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus) in stop_machine() argument 630 ret = stop_machine_cpuslocked(fn, data, cpus); in stop_machine() 680 const struct cpumask *cpus) in stop_machine_from_inactive_cpu() argument 683 .active_cpus = cpus }; in stop_machine_from_inactive_cpu()
|
/kernel/sched/ |
D | deadline.c | 113 int cpus; in dl_bw_cpus() local 121 cpus = 0; in dl_bw_cpus() 124 cpus++; in dl_bw_cpus() 126 return cpus; in dl_bw_cpus() 213 void __dl_sub(struct dl_bw *dl_b, u64 tsk_bw, int cpus) in __dl_sub() argument 216 __dl_update(dl_b, (s32)tsk_bw / cpus); in __dl_sub() 220 void __dl_add(struct dl_bw *dl_b, u64 tsk_bw, int cpus) in __dl_add() argument 223 __dl_update(dl_b, -((s32)tsk_bw / cpus)); in __dl_add() 2767 int cpu, cpus, ret = 0; in sched_dl_global_validate() local 2782 cpus = dl_bw_cpus(cpu); in sched_dl_global_validate() [all …]
|
D | topology.c | 1314 unsigned long cpus[]; member 1325 #define cpu_capacity_span(asym_data) to_cpumask((asym_data)->cpus) 2054 int sched_numa_find_closest(const struct cpumask *cpus, int cpu) in sched_numa_find_closest() argument 2066 cpu = cpumask_any_and(cpus, masks[i][j]); in sched_numa_find_closest() 2079 const struct cpumask *cpus; member 2091 if (cpumask_weight_and(k->cpus, cur_hop[k->node]) <= k->cpu) in hop_cmp() 2100 k->w = cpumask_weight_and(k->cpus, prev_hop[k->node]); in hop_cmp() 2115 int sched_numa_find_nth_cpu(const struct cpumask *cpus, int cpu, int node) in sched_numa_find_nth_cpu() argument 2117 struct __cmp_key k = { .cpus = cpus, .cpu = cpu }; in sched_numa_find_nth_cpu() 2122 return cpumask_nth_and(cpu, cpus, cpu_online_mask); in sched_numa_find_nth_cpu() [all …]
|
D | fair.c | 236 unsigned int cpus = min_t(unsigned int, num_online_cpus(), 8); in get_update_sysctl_factor() local 244 factor = cpus; in get_update_sysctl_factor() 248 factor = 1 + ilog2(cpus); in get_update_sysctl_factor() 7250 static int select_idle_core(struct task_struct *p, int core, struct cpumask *cpus, int *idle_cpu) in select_idle_core() argument 7259 if (sched_idle_cpu(cpu) && cpumask_test_cpu(cpu, cpus)) { in select_idle_core() 7267 if (*idle_cpu == -1 && cpumask_test_cpu(cpu, cpus)) in select_idle_core() 7274 cpumask_andnot(cpus, cpus, cpu_smt_mask(core)); in select_idle_core() 7312 static inline int select_idle_core(struct task_struct *p, int core, struct cpumask *cpus, int *idle… in select_idle_core() argument 7331 struct cpumask *cpus = this_cpu_cpumask_var_ptr(select_rq_mask); in select_idle_cpu() local 7339 cpumask_and(cpus, sched_domain_span(sd), p->cpus_ptr); in select_idle_cpu() [all …]
|
D | cpufreq_schedutil.c | 433 for_each_cpu(j, policy->cpus) { in sugov_next_freq_shared() 780 for_each_cpu(cpu, policy->cpus) { in sugov_start() 795 for_each_cpu(cpu, policy->cpus) { in sugov_start() 808 for_each_cpu(cpu, policy->cpus) in sugov_stop()
|
D | cpufreq.c | 72 return cpumask_test_cpu(smp_processor_id(), policy->cpus) || in cpufreq_this_cpu_can_update()
|
D | sched.h | 1793 extern int sched_numa_find_closest(const struct cpumask *cpus, int cpu); 1799 static inline int sched_numa_find_closest(const struct cpumask *cpus, int cpu) in sched_numa_find_closest() argument 3266 #define perf_domain_span(pd) (to_cpumask(((pd)->em_pd->cpus)))
|
/kernel/debug/kdb/ |
D | kdb_cmds | 26 defcmd dumpcpu "" "Same as dumpall but only tasks on cpus"
|
/kernel/trace/ |
D | ring_buffer.c | 512 int cpus; member 1865 buffer->cpus = nr_cpu_ids; in __ring_buffer_alloc()
|