Searched refs:cfs (Results 1 – 5 of 5) sorted by relevance
536 for (cfs_rq = &rq->cfs, pos = NULL; cfs_rq; cfs_rq = pos)743 nr_running = rq_of(cfs_rq)->cfs.h_nr_running; in sched_slice()1804 ns->nr_running += rq->cfs.h_nr_running; in update_numa_stats()3506 if (&rq->cfs == cfs_rq) { in cfs_rq_util_change()5016 rq_of(cfs_rq)->cfs.load.weight >= 2*se->load.weight) { in set_next_entity()5528 if (rq->curr == rq->idle && rq->cfs.nr_running) in unthrottle_cfs_rq()6033 if (rq->cfs.h_nr_running > 1) { in hrtick_start_fair()6102 return unlikely(rq->nr_running == rq->cfs.idle_h_nr_running && in sched_idle_rq()6144 util_est_enqueue(&rq->cfs, p); in enqueue_task_fair()6235 util_est_dequeue(&rq->cfs, p); in dequeue_task_fair()[all …]
1009 struct cfs_rq cfs; member1204 return container_of(cfs_rq, struct rq, cfs); in rq_of()1471 return &task_rq(p)->cfs; in task_cfs_rq()1479 return &rq->cfs; in cfs_rq_of()2320 return rq->cfs.nr_running > 0; in sched_fair_runnable()3012 cfs_rq = &cpu_rq(cpu)->cfs; in cpu_util_cfs()
165 u32 util_sum = rq->cfs.avg.util_sum; in update_idle_rq_clock_pelt()
608 rq0_min_vruntime = cpu_rq(0)->cfs.min_vruntime; in print_cfs_rq()
5560 struct sched_entity *curr = (&task_rq(p)->cfs)->curr; in prefetch_curr_exec_start()6036 rq->nr_running == rq->cfs.h_nr_running)) { in __pick_next_task()9959 init_cfs_rq(&rq->cfs); in sched_init()9984 init_tg_cfs_entry(&root_task_group, &rq->cfs, NULL, i, NULL); in sched_init()