Searched refs:cfs_rq (Results 1 – 6 of 6) sorted by relevance
/kernel/sched/ |
D | fair.c | 355 static inline bool list_add_leaf_cfs_rq(struct cfs_rq *cfs_rq) in list_add_leaf_cfs_rq() argument 357 struct rq *rq = rq_of(cfs_rq); in list_add_leaf_cfs_rq() 360 if (cfs_rq->on_list) in list_add_leaf_cfs_rq() 363 cfs_rq->on_list = 1; in list_add_leaf_cfs_rq() 374 if (cfs_rq->tg->parent && in list_add_leaf_cfs_rq() 375 cfs_rq->tg->parent->cfs_rq[cpu]->on_list) { in list_add_leaf_cfs_rq() 382 list_add_tail_rcu(&cfs_rq->leaf_cfs_rq_list, in list_add_leaf_cfs_rq() 383 &(cfs_rq->tg->parent->cfs_rq[cpu]->leaf_cfs_rq_list)); in list_add_leaf_cfs_rq() 393 if (!cfs_rq->tg->parent) { in list_add_leaf_cfs_rq() 398 list_add_tail_rcu(&cfs_rq->leaf_cfs_rq_list, in list_add_leaf_cfs_rq() [all …]
|
D | pelt.h | 5 int __update_load_avg_se(u64 now, struct cfs_rq *cfs_rq, struct sched_entity *se); 6 int __update_load_avg_cfs_rq(u64 now, struct cfs_rq *cfs_rq); 185 static inline void update_idle_cfs_rq_clock_pelt(struct cfs_rq *cfs_rq) in update_idle_cfs_rq_clock_pelt() argument 189 if (unlikely(cfs_rq->throttle_count)) in update_idle_cfs_rq_clock_pelt() 192 throttled = cfs_rq->throttled_clock_pelt_time; in update_idle_cfs_rq_clock_pelt() 194 u64_u32_store(cfs_rq->throttled_pelt_idle, throttled); in update_idle_cfs_rq_clock_pelt() 198 static inline u64 cfs_rq_clock_pelt(struct cfs_rq *cfs_rq) in cfs_rq_clock_pelt() argument 200 if (unlikely(cfs_rq->throttle_count)) in cfs_rq_clock_pelt() 201 return cfs_rq->throttled_clock_pelt - cfs_rq->throttled_clock_pelt_time; in cfs_rq_clock_pelt() 203 return rq_clock_pelt(rq_of(cfs_rq)) - cfs_rq->throttled_clock_pelt_time; in cfs_rq_clock_pelt() [all …]
|
D | pelt.c | 313 int __update_load_avg_se(u64 now, struct cfs_rq *cfs_rq, struct sched_entity *se) in __update_load_avg_se() argument 316 cfs_rq->curr == se)) { in __update_load_avg_se() 327 int __update_load_avg_cfs_rq(u64 now, struct cfs_rq *cfs_rq) in __update_load_avg_cfs_rq() argument 329 if (___update_load_sum(now, &cfs_rq->avg, in __update_load_avg_cfs_rq() 330 scale_load_down(cfs_rq->load.weight), in __update_load_avg_cfs_rq() 331 cfs_rq->h_nr_running, in __update_load_avg_cfs_rq() 332 cfs_rq->curr != NULL)) { in __update_load_avg_cfs_rq() 334 ___update_load_avg(&cfs_rq->avg, 1); in __update_load_avg_cfs_rq() 335 trace_pelt_cfs_tp(cfs_rq); in __update_load_avg_cfs_rq()
|
D | debug.c | 583 void print_cfs_rq(struct seq_file *m, int cpu, struct cfs_rq *cfs_rq) in print_cfs_rq() argument 593 SEQ_printf_task_group_path(m, cfs_rq->tg, "cfs_rq[%d]:%s\n", cpu); in print_cfs_rq() 599 SPLIT_NS(cfs_rq->exec_clock)); in print_cfs_rq() 602 if (rb_first_cached(&cfs_rq->tasks_timeline)) in print_cfs_rq() 603 MIN_vruntime = (__pick_first_entity(cfs_rq))->vruntime; in print_cfs_rq() 604 last = __pick_last_entity(cfs_rq); in print_cfs_rq() 607 min_vruntime = cfs_rq->min_vruntime; in print_cfs_rq() 623 cfs_rq->nr_spread_over); in print_cfs_rq() 624 SEQ_printf(m, " .%-30s: %d\n", "nr_running", cfs_rq->nr_running); in print_cfs_rq() 625 SEQ_printf(m, " .%-30s: %d\n", "h_nr_running", cfs_rq->h_nr_running); in print_cfs_rq() [all …]
|
D | sched.h | 340 struct cfs_rq; 379 struct cfs_rq **cfs_rq; member 471 extern void init_tg_cfs_entry(struct task_group *tg, struct cfs_rq *cfs_rq, 478 extern void unthrottle_cfs_rq(struct cfs_rq *cfs_rq); 504 struct cfs_rq *prev, struct cfs_rq *next); 507 struct cfs_rq *prev, struct cfs_rq *next) { } in set_task_rq_fair() 561 struct cfs_rq { struct 1009 struct cfs_rq cfs; 1195 static inline struct rq *rq_of(struct cfs_rq *cfs_rq) in rq_of() argument 1197 return cfs_rq->rq; in rq_of() [all …]
|
D | core.c | 4535 p->se.cfs_rq = NULL; in __sched_fork() 5558 struct sched_entity *curr = (&p->se)->cfs_rq->curr; in prefetch_curr_exec_start() 9915 root_task_group.cfs_rq = (struct cfs_rq **)ptr; in sched_init() 10883 struct cfs_rq *cfs_rq = tg->cfs_rq[i]; in tg_set_cfs_bandwidth() local 10884 struct rq *rq = cfs_rq->rq; in tg_set_cfs_bandwidth() 10888 cfs_rq->runtime_enabled = runtime_enabled; in tg_set_cfs_bandwidth() 10889 cfs_rq->runtime_remaining = 0; in tg_set_cfs_bandwidth() 10891 if (cfs_rq->throttled) in tg_set_cfs_bandwidth() 10892 unthrottle_cfs_rq(cfs_rq); in tg_set_cfs_bandwidth()
|