Lines Matching refs:this_cpu
6173 wake_affine_idle(int this_cpu, int prev_cpu, int sync) in wake_affine_idle() argument
6187 if (available_idle_cpu(this_cpu) && cpus_share_cache(this_cpu, prev_cpu)) in wake_affine_idle()
6188 return available_idle_cpu(prev_cpu) ? prev_cpu : this_cpu; in wake_affine_idle()
6190 if (sync && cpu_rq(this_cpu)->nr_running == 1) in wake_affine_idle()
6191 return this_cpu; in wake_affine_idle()
6201 int this_cpu, int prev_cpu, int sync) in wake_affine_weight() argument
6206 this_eff_load = cpu_load(cpu_rq(this_cpu)); in wake_affine_weight()
6212 return this_cpu; in wake_affine_weight()
6228 prev_eff_load *= capacity_of(this_cpu); in wake_affine_weight()
6239 return this_eff_load < prev_eff_load ? this_cpu : nr_cpumask_bits; in wake_affine_weight()
6243 int this_cpu, int prev_cpu, int sync) in wake_affine() argument
6248 target = wake_affine_idle(this_cpu, prev_cpu, sync); in wake_affine()
6251 target = wake_affine_weight(sd, p, this_cpu, prev_cpu, sync); in wake_affine()
6254 if (target != this_cpu) in wake_affine()
6263 find_idlest_group(struct sched_domain *sd, struct task_struct *p, int this_cpu);
6269 find_idlest_group_cpu(struct sched_group *group, struct task_struct *p, int this_cpu) in find_idlest_group_cpu() argument
6274 int least_loaded_cpu = this_cpu; in find_idlest_group_cpu()
9465 find_idlest_group(struct sched_domain *sd, struct task_struct *p, int this_cpu) in find_idlest_group() argument
9485 if (!sched_group_cookie_match(cpu_rq(this_cpu), p, group)) in find_idlest_group()
9488 local_group = cpumask_test_cpu(this_cpu, in find_idlest_group()
9581 if (cpu_to_node(this_cpu) == p->numa_preferred_nid) in find_idlest_group()
10336 static int load_balance(int this_cpu, struct rq *this_rq, in load_balance() argument
10349 .dst_cpu = this_cpu, in load_balance()
10521 if (!cpumask_test_cpu(this_cpu, busiest->curr->cpus_ptr)) { in load_balance()
10536 busiest->push_cpu = this_cpu; in load_balance()
11165 int this_cpu = this_rq->cpu; in _nohz_idle_balance() local
11191 for_each_cpu_wrap(balance_cpu, nohz.idle_cpus_mask, this_cpu+1) { in _nohz_idle_balance()
11288 int this_cpu = this_rq->cpu; in nohz_newidle_balance() local
11294 if (!housekeeping_cpu(this_cpu, HK_FLAG_SCHED)) in nohz_newidle_balance()
11310 atomic_or(NOHZ_NEWILB_KICK, nohz_flags(this_cpu)); in nohz_newidle_balance()
11336 int this_cpu = this_rq->cpu; in newidle_balance() local
11364 if (!cpu_active(this_cpu)) in newidle_balance()
11389 update_blocked_averages(this_cpu); in newidle_balance()
11391 for_each_domain(this_cpu, sd) { in newidle_balance()
11401 t0 = sched_clock_cpu(this_cpu); in newidle_balance()
11403 pulled_task = load_balance(this_cpu, this_rq, in newidle_balance()
11407 domain_cost = sched_clock_cpu(this_cpu) - t0; in newidle_balance()