Searched refs:cpus (Results 1 – 10 of 10) sorted by relevance
/kernel/ |
D | stop_machine.c | 550 int __stop_machine(int (*fn)(void *), void *data, const struct cpumask *cpus) in __stop_machine() argument 556 .active_cpus = cpus, in __stop_machine() 583 int stop_machine(int (*fn)(void *), void *data, const struct cpumask *cpus) in stop_machine() argument 589 ret = __stop_machine(fn, data, cpus); in stop_machine() 618 const struct cpumask *cpus) in stop_machine_from_inactive_cpu() argument 621 .active_cpus = cpus }; in stop_machine_from_inactive_cpu()
|
D | smp.c | 651 cpumask_var_t cpus; in on_each_cpu_cond() local 656 if (likely(zalloc_cpumask_var(&cpus, (gfp_flags|__GFP_NOWARN)))) { in on_each_cpu_cond() 660 cpumask_set_cpu(cpu, cpus); in on_each_cpu_cond() 661 on_each_cpu_mask(cpus, func, info, wait); in on_each_cpu_cond() 663 free_cpumask_var(cpus); in on_each_cpu_cond()
|
D | watchdog.c | 246 cpumask_t cpus = watchdog_cpus; in watchdog_next_cpu() local 249 next_cpu = cpumask_next(cpu, &cpus); in watchdog_next_cpu() 251 next_cpu = cpumask_first(&cpus); in watchdog_next_cpu()
|
/kernel/sched/ |
D | cpufreq_sched.c | 189 for_each_cpu(cpu_tmp, policy->cpus) { in update_fdomain_capacity_request() 277 for_each_cpu(cpu, policy->cpus) in cpufreq_sched_policy_init() 348 for_each_cpu(cpu, policy->cpus) in cpufreq_sched_start() 373 for_each_cpu(cpu, policy->cpus) in cpufreq_sched_stop()
|
D | fair.c | 160 unsigned int cpus = min_t(int, num_online_cpus(), 8); in get_update_sysctl_factor() local 168 factor = cpus; in get_update_sysctl_factor() 172 factor = 1 + ilog2(cpus); in get_update_sysctl_factor() 1071 int smt, cpu, cpus = 0; in update_numa_stats() local 1082 cpus++; in update_numa_stats() 1093 if (!cpus) in update_numa_stats() 1097 smt = DIV_ROUND_UP(SCHED_CAPACITY_SCALE * cpus, ns->compute_capacity); in update_numa_stats() 1098 capacity = cpus / smt; /* cores */ in update_numa_stats() 6306 struct cpumask *cpus; member 6479 for_each_cpu_and(cpu, env->dst_grpmask, env->cpus) { in can_migrate_task() [all …]
|
D | walt.c | 992 const struct cpumask *cpus = policy->related_cpus; in cpufreq_notifier_policy() local 1053 cpus = cpu_possible_mask; in cpufreq_notifier_policy() 1062 for_each_cpu(i, cpus) { in cpufreq_notifier_policy()
|
D | core.c | 2287 int cpus = 0; in dl_bw_cpus() local 2292 cpus++; in dl_bw_cpus() 2294 return cpus; in dl_bw_cpus() 2321 bool __dl_overflow(struct dl_bw *dl_b, int cpus, u64 old_bw, u64 new_bw) in __dl_overflow() argument 2324 dl_b->bw * cpus < dl_b->total_bw - old_bw + new_bw; in __dl_overflow() 2346 int cpus, err = -1; in dl_overflow() local 2357 cpus = dl_bw_cpus(task_cpu(p)); in dl_overflow() 2359 !__dl_overflow(dl_b, cpus, 0, new_bw)) { in dl_overflow() 2363 !__dl_overflow(dl_b, cpus, p->dl.dl_bw, new_bw)) { in dl_overflow() 5621 int cpus; in sched_cpu_inactive() local [all …]
|
/kernel/debug/kdb/ |
D | kdb_cmds | 26 defcmd dumpcpu "" "Same as dumpall but only tasks on cpus"
|
/kernel/power/ |
D | power.h | 11 int cpus; member
|
/kernel/trace/ |
D | ring_buffer.c | 504 int cpus; member 1374 buffer->cpus = nr_cpu_ids; in __ring_buffer_alloc()
|