Lines Matching refs:dest_cpu
2561 int dest_cpu; member
2587 struct task_struct *p, int dest_cpu) in __migrate_task() argument
2590 if (!is_cpu_allowed(p, dest_cpu)) in __migrate_task()
2594 rq = move_queued_task(rq, rf, p, dest_cpu); in __migrate_task()
2653 rq = __migrate_task(rq, &rf, p, arg->dest_cpu); in migration_cpu_stop()
2655 p->wake_cpu = arg->dest_cpu; in migration_cpu_stop()
2953 int dest_cpu, unsigned int flags) in affine_move_task() argument
3000 .dest_cpu = dest_cpu, in affine_move_task()
3016 pending->arg.dest_cpu = dest_cpu; in affine_move_task()
3064 rq = move_queued_task(rq, rf, p, dest_cpu); in affine_move_task()
3109 unsigned int dest_cpu; in __set_cpus_allowed_ptr_locked() local
3159 dest_cpu = cpumask_any_and_distribute(cpu_valid_mask, new_mask); in __set_cpus_allowed_ptr_locked()
3160 trace_android_rvh_set_cpus_allowed_by_task(cpu_valid_mask, new_mask, p, &dest_cpu); in __set_cpus_allowed_ptr_locked()
3162 if (dest_cpu >= nr_cpu_ids) { in __set_cpus_allowed_ptr_locked()
3172 ret = affine_move_task(rq, p, rf, dest_cpu, flags); in __set_cpus_allowed_ptr_locked()
3566 int dest_cpu = -1; in select_fallback_rq() local
3568 trace_android_rvh_select_fallback_rq(cpu, p, &dest_cpu); in select_fallback_rq()
3569 if (dest_cpu >= 0) in select_fallback_rq()
3570 return dest_cpu; in select_fallback_rq()
3581 for_each_cpu(dest_cpu, nodemask) { in select_fallback_rq()
3582 if (is_cpu_allowed(p, dest_cpu)) in select_fallback_rq()
3583 return dest_cpu; in select_fallback_rq()
3589 for_each_cpu(dest_cpu, p->cpus_ptr) { in select_fallback_rq()
3590 if (!is_cpu_allowed(p, dest_cpu)) in select_fallback_rq()
3633 return dest_cpu; in select_fallback_rq()
5518 int dest_cpu; in sched_exec() local
5526 dest_cpu = p->sched_class->select_task_rq(p, task_cpu(p), WF_EXEC); in sched_exec()
5527 if (dest_cpu == smp_processor_id()) in sched_exec()
5530 if (likely(cpu_active(dest_cpu))) { in sched_exec()
5531 struct migration_arg arg = { p, dest_cpu }; in sched_exec()