Lines Matching defs:cfs_rq
355 static inline bool list_add_leaf_cfs_rq(struct cfs_rq *cfs_rq) in list_add_leaf_cfs_rq()
423 static inline void list_del_leaf_cfs_rq(struct cfs_rq *cfs_rq) in list_del_leaf_cfs_rq()
449 #define for_each_leaf_cfs_rq_safe(rq, cfs_rq, pos) \ argument
505 static int cfs_rq_is_idle(struct cfs_rq *cfs_rq) in cfs_rq_is_idle()
522 static inline bool list_add_leaf_cfs_rq(struct cfs_rq *cfs_rq) in list_add_leaf_cfs_rq()
527 static inline void list_del_leaf_cfs_rq(struct cfs_rq *cfs_rq) in list_del_leaf_cfs_rq()
535 #define for_each_leaf_cfs_rq_safe(rq, cfs_rq, pos) \ argument
553 static int cfs_rq_is_idle(struct cfs_rq *cfs_rq) in cfs_rq_is_idle()
599 static void update_min_vruntime(struct cfs_rq *cfs_rq) in update_min_vruntime()
635 static void __enqueue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se) in __enqueue_entity()
641 static void __dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se) in __dequeue_entity()
647 struct sched_entity *__pick_first_entity(struct cfs_rq *cfs_rq) in __pick_first_entity()
668 struct sched_entity *__pick_last_entity(struct cfs_rq *cfs_rq) in __pick_last_entity()
735 static u64 sched_slice(struct cfs_rq *cfs_rq, struct sched_entity *se) in sched_slice()
781 static u64 sched_vslice(struct cfs_rq *cfs_rq, struct sched_entity *se) in sched_vslice()
841 struct cfs_rq *cfs_rq = cfs_rq_of(se); in post_init_entity_util_avg() local
886 static void update_tg_load_avg(struct cfs_rq *cfs_rq) in update_tg_load_avg()
894 static void update_curr(struct cfs_rq *cfs_rq) in update_curr()
940 update_stats_wait_start_fair(struct cfs_rq *cfs_rq, struct sched_entity *se) in update_stats_wait_start_fair()
957 update_stats_wait_end_fair(struct cfs_rq *cfs_rq, struct sched_entity *se) in update_stats_wait_end_fair()
983 update_stats_enqueue_sleeper_fair(struct cfs_rq *cfs_rq, struct sched_entity *se) in update_stats_enqueue_sleeper_fair()
1003 update_stats_enqueue_fair(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in update_stats_enqueue_fair()
1020 update_stats_dequeue_fair(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in update_stats_dequeue_fair()
1052 update_stats_curr_start(struct cfs_rq *cfs_rq, struct sched_entity *se) in update_stats_curr_start()
3215 account_entity_enqueue(struct cfs_rq *cfs_rq, struct sched_entity *se) in account_entity_enqueue()
3232 account_entity_dequeue(struct cfs_rq *cfs_rq, struct sched_entity *se) in account_entity_dequeue()
3296 enqueue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) in enqueue_load_avg()
3303 dequeue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) in dequeue_load_avg()
3313 enqueue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) { } in enqueue_load_avg()
3315 dequeue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) { } in dequeue_load_avg()
3318 static void reweight_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, in reweight_entity()
3348 struct cfs_rq *cfs_rq = cfs_rq_of(se); in reweight_task() local
3434 static long calc_group_shares(struct cfs_rq *cfs_rq) in calc_group_shares()
3502 static inline void cfs_rq_util_change(struct cfs_rq *cfs_rq, int flags) in cfs_rq_util_change()
3549 static inline u64 cfs_rq_last_update_time(struct cfs_rq *cfs_rq) in cfs_rq_last_update_time()
3563 static inline bool child_cfs_rq_on_list(struct cfs_rq *cfs_rq) in child_cfs_rq_on_list()
3581 static inline bool cfs_rq_is_decayed(struct cfs_rq *cfs_rq) in cfs_rq_is_decayed()
3609 static inline void update_tg_load_avg(struct cfs_rq *cfs_rq) in update_tg_load_avg()
3724 update_tg_cfs_util(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq) in update_tg_cfs_util()
3756 update_tg_cfs_runnable(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq) in update_tg_cfs_runnable()
3786 update_tg_cfs_load(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq) in update_tg_cfs_load()
3853 static inline void add_tg_cfs_propagate(struct cfs_rq *cfs_rq, long runnable_sum) in add_tg_cfs_propagate()
3862 struct cfs_rq *cfs_rq, *gcfs_rq; in propagate_entity_load_avg() local
3919 static inline void update_tg_load_avg(struct cfs_rq *cfs_rq) {} in update_tg_load_avg()
3926 static inline void add_tg_cfs_propagate(struct cfs_rq *cfs_rq, long runnable_sum) {} in add_tg_cfs_propagate()
3934 struct cfs_rq *cfs_rq; in migrate_se_pelt_lag() local
4029 update_cfs_rq_load_avg(u64 now, struct cfs_rq *cfs_rq) in update_cfs_rq_load_avg()
4100 static void attach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) in attach_entity_load_avg()
4157 static void detach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) in detach_entity_load_avg()
4190 static inline void update_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in update_load_avg()
4240 struct cfs_rq *cfs_rq = cfs_rq_of(se); in sync_entity_load_avg() local
4253 struct cfs_rq *cfs_rq = cfs_rq_of(se); in remove_entity_load_avg() local
4274 static inline unsigned long cfs_rq_runnable_avg(struct cfs_rq *cfs_rq) in cfs_rq_runnable_avg()
4279 static inline unsigned long cfs_rq_load_avg(struct cfs_rq *cfs_rq) in cfs_rq_load_avg()
4303 static inline void util_est_enqueue(struct cfs_rq *cfs_rq, in util_est_enqueue()
4319 static inline void util_est_dequeue(struct cfs_rq *cfs_rq, in util_est_dequeue()
4350 static inline void util_est_update(struct cfs_rq *cfs_rq, in util_est_update()
4605 static inline bool cfs_rq_is_decayed(struct cfs_rq *cfs_rq) in cfs_rq_is_decayed()
4615 static inline void update_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se, int not_used1) in update_load_avg()
4623 attach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) {} in attach_entity_load_avg()
4625 detach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) {} in detach_entity_load_avg()
4633 util_est_enqueue(struct cfs_rq *cfs_rq, struct task_struct *p) {} in util_est_enqueue()
4636 util_est_dequeue(struct cfs_rq *cfs_rq, struct task_struct *p) {} in util_est_dequeue()
4639 util_est_update(struct cfs_rq *cfs_rq, struct task_struct *p, in util_est_update()
4645 static void check_spread(struct cfs_rq *cfs_rq, struct sched_entity *se) in check_spread()
4660 struct cfs_rq *cfs_rq; in entity_is_long_sleeper() local
4682 place_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int initial) in place_entity()
4775 enqueue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in enqueue_entity()
4835 struct cfs_rq *cfs_rq = cfs_rq_of(se); in __clear_buddies_last() local
4846 struct cfs_rq *cfs_rq = cfs_rq_of(se); in __clear_buddies_next() local
4857 struct cfs_rq *cfs_rq = cfs_rq_of(se); in __clear_buddies_skip() local
4865 static void clear_buddies(struct cfs_rq *cfs_rq, struct sched_entity *se) in clear_buddies()
4880 dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in dequeue_entity()
4944 check_preempt_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr) in check_preempt_tick()
4991 void set_next_entity(struct cfs_rq *cfs_rq, struct sched_entity *se) in set_next_entity()
5040 pick_next_entity(struct cfs_rq *cfs_rq, struct sched_entity *curr) in pick_next_entity()
5095 static void put_prev_entity(struct cfs_rq *cfs_rq, struct sched_entity *prev) in put_prev_entity()
5120 entity_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr, int queued) in entity_tick()
5235 struct cfs_rq *cfs_rq, u64 target_runtime) in __assign_cfs_rq_runtime()
5262 static int assign_cfs_rq_runtime(struct cfs_rq *cfs_rq) in assign_cfs_rq_runtime()
5274 static void __account_cfs_rq_runtime(struct cfs_rq *cfs_rq, u64 delta_exec) in __account_cfs_rq_runtime()
5293 void account_cfs_rq_runtime(struct cfs_rq *cfs_rq, u64 delta_exec) in account_cfs_rq_runtime()
5301 static inline int cfs_rq_throttled(struct cfs_rq *cfs_rq) in cfs_rq_throttled()
5307 static inline int throttled_hierarchy(struct cfs_rq *cfs_rq) in throttled_hierarchy()
5332 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in tg_unthrottle_up() local
5350 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in tg_throttle_down() local
5362 static bool throttle_cfs_rq(struct cfs_rq *cfs_rq) in throttle_cfs_rq()
5449 void unthrottle_cfs_rq(struct cfs_rq *cfs_rq) in unthrottle_cfs_rq()
5534 struct cfs_rq *cfs_rq; in distribute_cfs_runtime() local
5685 static void __return_cfs_rq_runtime(struct cfs_rq *cfs_rq) in __return_cfs_rq_runtime()
5708 static __always_inline void return_cfs_rq_runtime(struct cfs_rq *cfs_rq) in return_cfs_rq_runtime()
5753 static void check_enqueue_throttle(struct cfs_rq *cfs_rq) in check_enqueue_throttle()
5774 struct cfs_rq *pcfs_rq, *cfs_rq; in sync_throttle() local
5790 static bool check_cfs_rq_runtime(struct cfs_rq *cfs_rq) in check_cfs_rq_runtime()
5891 static void init_cfs_rq_runtime(struct cfs_rq *cfs_rq) in init_cfs_rq_runtime()
5936 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in update_runtime_enabled() local
5954 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in unthrottle_offline_cfs_rqs() local
5983 static void account_cfs_rq_runtime(struct cfs_rq *cfs_rq, u64 delta_exec) {} in account_cfs_rq_runtime()
5984 static bool check_cfs_rq_runtime(struct cfs_rq *cfs_rq) { return false; } in check_cfs_rq_runtime()
5985 static void check_enqueue_throttle(struct cfs_rq *cfs_rq) {} in check_enqueue_throttle()
5987 static __always_inline void return_cfs_rq_runtime(struct cfs_rq *cfs_rq) {} in return_cfs_rq_runtime()
5989 static inline int cfs_rq_throttled(struct cfs_rq *cfs_rq) in cfs_rq_throttled()
5994 static inline int throttled_hierarchy(struct cfs_rq *cfs_rq) in throttled_hierarchy()
6008 static void init_cfs_rq_runtime(struct cfs_rq *cfs_rq) {} in init_cfs_rq_runtime()
6029 struct cfs_rq *cfs_rq = cfs_rq_of(se); in hrtick_start_fair() local
6111 static bool sched_idle_cfs_rq(struct cfs_rq *cfs_rq) in sched_idle_cfs_rq()
6132 struct cfs_rq *cfs_rq; in enqueue_task_fair() local
6229 struct cfs_rq *cfs_rq; in dequeue_task_fair() local
6337 struct cfs_rq *cfs_rq; in cpu_load_without() local
6360 struct cfs_rq *cfs_rq; in cpu_runnable_without() local
7055 struct cfs_rq *cfs_rq = &cpu_rq(cpu)->cfs; in cpu_util_next() local
7599 struct cfs_rq *cfs_rq = cfs_rq_of(se); in migrate_task_rq_fair() local
7725 struct cfs_rq *cfs_rq = task_cfs_rq(curr); in check_preempt_wakeup() local
7834 struct cfs_rq *cfs_rq; in pick_task_fair() local
7866 struct cfs_rq *cfs_rq = &rq->cfs; in pick_next_task_fair() local
8023 struct cfs_rq *cfs_rq; in put_prev_task_fair() local
8039 struct cfs_rq *cfs_rq = task_cfs_rq(curr); in yield_task_fair() local
8734 static inline bool cfs_rq_has_blocked(struct cfs_rq *cfs_rq) in cfs_rq_has_blocked()
8775 static inline bool cfs_rq_has_blocked(struct cfs_rq *cfs_rq) { return false; } in cfs_rq_has_blocked()
8811 struct cfs_rq *cfs_rq, *pos; in __update_blocked_fair() local
8859 static void update_cfs_rq_h_load(struct cfs_rq *cfs_rq) in update_cfs_rq_h_load()
8894 struct cfs_rq *cfs_rq = task_cfs_rq(p); in task_h_load() local
8903 struct cfs_rq *cfs_rq = &rq->cfs; in __update_blocked_fair() local
11906 struct cfs_rq *cfs_rq = cfs_rq_of(se); in se_fi_update() local
11988 struct cfs_rq *cfs_rq; in task_tick_fair() local
12012 struct cfs_rq *cfs_rq; in task_fork_fair() local
12101 struct cfs_rq *cfs_rq = cfs_rq_of(se); in propagate_entity_cfs_rq() local
12130 struct cfs_rq *cfs_rq = cfs_rq_of(se); in detach_entity_cfs_rq() local
12152 struct cfs_rq *cfs_rq = cfs_rq_of(se); in attach_entity_cfs_rq() local
12164 struct cfs_rq *cfs_rq = cfs_rq_of(se); in detach_task_cfs_rq() local
12181 struct cfs_rq *cfs_rq = cfs_rq_of(se); in attach_task_cfs_rq() local
12231 struct cfs_rq *cfs_rq = cfs_rq_of(se); in set_next_task_fair() local
12239 void init_cfs_rq(struct cfs_rq *cfs_rq) in init_cfs_rq()
12286 struct cfs_rq *cfs_rq; in alloc_fair_sched_group() local
12369 void init_tg_cfs_entry(struct task_group *tg, struct cfs_rq *cfs_rq, in init_tg_cfs_entry()
12499 struct cfs_rq *cfs_rq = cfs_rq_of(se); in sched_group_set_idle() local
12608 struct cfs_rq *cfs_rq, *pos; in print_cfs_stats() local