Searched refs:new_cpu (Results 1 – 6 of 6) sorted by relevance
/kernel/sched/ |
D | cpudeadline.c | 176 int old_idx, new_cpu; in cpudl_clear() local 191 new_cpu = cp->elements[cp->size - 1].cpu; in cpudl_clear() 193 cp->elements[old_idx].cpu = new_cpu; in cpudl_clear() 195 cp->elements[new_cpu].idx = old_idx; in cpudl_clear()
|
D | fair.c | 2941 static void update_scan_period(struct task_struct *p, int new_cpu) in update_scan_period() argument 2944 int dst_nid = cpu_to_node(new_cpu); in update_scan_period() 2988 static inline void update_scan_period(struct task_struct *p, int new_cpu) in update_scan_period() argument 6286 int new_cpu = cpu; in find_idlest_cpu() local 6314 new_cpu = find_idlest_group_cpu(group, p, cpu); in find_idlest_cpu() 6315 if (new_cpu == cpu) { in find_idlest_cpu() 6322 cpu = new_cpu; in find_idlest_cpu() 6333 return new_cpu; in find_idlest_cpu() 7021 int new_cpu = INT_MAX; in find_energy_efficient_cpu() local 7024 trace_android_rvh_find_energy_efficient_cpu(p, prev_cpu, sync, &new_cpu); in find_energy_efficient_cpu() [all …]
|
D | core.c | 2285 struct task_struct *p, int new_cpu) in move_queued_task() argument 2296 trace_android_rvh_migrate_queued_task(rq, rf, p, new_cpu, &detached); in move_queued_task() 2301 set_task_cpu(p, new_cpu); in move_queued_task() 2305 rq = cpu_rq(new_cpu); in move_queued_task() 2308 BUG_ON(task_cpu(p) != new_cpu); in move_queued_task() 3099 void set_task_cpu(struct task_struct *p, unsigned int new_cpu) in set_task_cpu() argument 3136 WARN_ON_ONCE(!cpu_online(new_cpu)); in set_task_cpu() 3141 trace_sched_migrate_task(p, new_cpu); in set_task_cpu() 3143 if (task_cpu(p) != new_cpu) { in set_task_cpu() 3145 p->sched_class->migrate_task_rq(p, new_cpu); in set_task_cpu() [all …]
|
D | sched.h | 2192 void (*migrate_task_rq)(struct task_struct *p, int new_cpu);
|
D | deadline.c | 1732 static void migrate_task_rq_dl(struct task_struct *p, int new_cpu __maybe_unused) in migrate_task_rq_dl()
|
/kernel/ |
D | workqueue.c | 1415 int new_cpu; in wq_select_unbound_cpu() local 1428 new_cpu = __this_cpu_read(wq_rr_cpu_last); in wq_select_unbound_cpu() 1429 new_cpu = cpumask_next_and(new_cpu, wq_unbound_cpumask, cpu_online_mask); in wq_select_unbound_cpu() 1430 if (unlikely(new_cpu >= nr_cpu_ids)) { in wq_select_unbound_cpu() 1431 new_cpu = cpumask_first_and(wq_unbound_cpumask, cpu_online_mask); in wq_select_unbound_cpu() 1432 if (unlikely(new_cpu >= nr_cpu_ids)) in wq_select_unbound_cpu() 1435 __this_cpu_write(wq_rr_cpu_last, new_cpu); in wq_select_unbound_cpu() 1437 return new_cpu; in wq_select_unbound_cpu()
|