Lines Matching refs:idle
4866 cfs_b->idle = 0; in __assign_cfs_rq_runtime()
5126 if (rq->curr == rq->idle && rq->cfs.nr_running) in unthrottle_cfs_rq()
5192 if (cfs_b->idle && !throttled) in do_sched_cfs_period_timer()
5199 cfs_b->idle = 1; in do_sched_cfs_period_timer()
5224 cfs_b->idle = 0; in do_sched_cfs_period_timer()
5426 int idle = 0; in sched_cfs_period_timer() local
5435 idle = do_sched_cfs_period_timer(cfs_b, overrun, flags); in sched_cfs_period_timer()
5467 if (idle) in sched_cfs_period_timer()
5471 return idle ? HRTIMER_NORESTART : HRTIMER_RESTART; in sched_cfs_period_timer()
6151 struct cpuidle_state *idle = idle_get_state(rq); in find_idlest_group_cpu() local
6152 if (idle && idle->exit_latency < min_exit_latency) { in find_idlest_group_cpu()
6158 min_exit_latency = idle->exit_latency; in find_idlest_group_cpu()
6161 } else if ((!idle || idle->exit_latency == min_exit_latency) && in find_idlest_group_cpu()
6308 bool idle = true; in select_idle_core() local
6312 idle = false; in select_idle_core()
6318 if (idle) in select_idle_core()
6855 struct cpuidle_state *idle; in find_energy_efficient_cpu() local
6969 idle = idle_get_state(cpu_rq(cpu)); in find_energy_efficient_cpu()
6970 if (idle && idle->exit_latency > min_exit_lat && in find_energy_efficient_cpu()
6974 if (idle) in find_energy_efficient_cpu()
6975 min_exit_lat = idle->exit_latency; in find_energy_efficient_cpu()
7340 if (unlikely(!se->on_rq || curr == rq->idle)) in check_preempt_wakeup()
7358 goto idle; in pick_next_task_fair()
7397 goto idle; in pick_next_task_fair()
7473 idle: in pick_next_task_fair()
7757 enum cpu_idle_type idle; member
7848 if (env->idle == CPU_IDLE) in migrate_degrades_locality()
7915 if (env->idle == CPU_NEWLY_IDLE || (env->flags & LBF_DST_PINNED)) in can_migrate_task()
7951 schedstat_inc(env->sd->lb_hot_gained[env->idle]); in can_migrate_task()
8009 schedstat_inc(env->sd->lb_gained[env->idle]); in detach_one_task()
8040 if (env->idle != CPU_NOT_IDLE && env->src_rq->nr_running <= 1) in detach_tasks()
8120 if (env->idle == CPU_NEWLY_IDLE) in detach_tasks()
8141 schedstat_add(env->sd->lb_gained[env->idle], detached); in detach_tasks()
8813 env->idle != CPU_NOT_IDLE && in update_sg_lb_stats()
9012 if (rq->curr != rq->idle && rq->curr != p) in idle_cpu_without()
9306 if (env->idle == CPU_NEWLY_IDLE && READ_ONCE(nohz.has_blocked)) in update_sd_lb_stats()
9319 if (env->idle != CPU_NEWLY_IDLE || in update_sd_lb_stats()
9458 if (env->idle != CPU_NOT_IDLE && env->imbalance == 0) { in calculate_imbalance()
9670 if (env->idle == CPU_NOT_IDLE) in find_busiest_group()
9856 return env->idle != CPU_NOT_IDLE && (env->sd->flags & SD_ASYM_PACKING) && in asym_active_balance()
9874 if ((env->idle != CPU_NOT_IDLE) && in voluntary_active_balance()
9915 if (env->idle == CPU_NEWLY_IDLE) in should_we_balance()
9936 struct sched_domain *sd, enum cpu_idle_type idle, in load_balance() argument
9951 .idle = idle, in load_balance()
9960 schedstat_inc(sd->lb_count[idle]); in load_balance()
9970 schedstat_inc(sd->lb_nobusyg[idle]); in load_balance()
9976 schedstat_inc(sd->lb_nobusyq[idle]); in load_balance()
9982 schedstat_add(sd->lb_imbalance[idle], env.imbalance); in load_balance()
10099 schedstat_inc(sd->lb_failed[idle]); in load_balance()
10106 if (idle != CPU_NEWLY_IDLE) in load_balance()
10185 schedstat_inc(sd->lb_balanced[idle]); in load_balance()
10198 if (env.idle == CPU_NEWLY_IDLE) in load_balance()
10302 .idle = CPU_IDLE, in active_load_balance_cpu_stop()
10355 static void rebalance_domains(struct rq *rq, enum cpu_idle_type idle) in rebalance_domains() argument
10359 int busy = idle != CPU_IDLE && !sched_idle_cpu(cpu); in rebalance_domains()
10406 if (load_balance(cpu, rq, sd, idle, &continue_balancing)) { in rebalance_domains()
10412 idle = idle_cpu(cpu) ? CPU_IDLE : CPU_NOT_IDLE; in rebalance_domains()
10413 busy = idle != CPU_IDLE && !sched_idle_cpu(cpu); in rebalance_domains()
10453 if ((idle == CPU_IDLE) && time_after(nohz.next_balance, rq->next_balance)) in rebalance_domains()
10772 enum cpu_idle_type idle) in _nohz_idle_balance() argument
10850 if (idle != CPU_NEWLY_IDLE) { in _nohz_idle_balance()
10876 static bool nohz_idle_balance(struct rq *this_rq, enum cpu_idle_type idle) in nohz_idle_balance() argument
10885 if (idle != CPU_IDLE) in nohz_idle_balance()
10888 _nohz_idle_balance(this_rq, flags, idle); in nohz_idle_balance()
10928 static inline bool nohz_idle_balance(struct rq *this_rq, enum cpu_idle_type idle) in nohz_idle_balance() argument
11068 enum cpu_idle_type idle = this_rq->idle_balance ? in run_rebalance_domains()
11079 if (nohz_idle_balance(this_rq, idle)) in run_rebalance_domains()
11084 rebalance_domains(this_rq, idle); in run_rebalance_domains()