Lines Matching refs:rt_rq
286 struct rt_rq **rt_rq; member
323 static DEFINE_PER_CPU(struct rt_rq, init_rt_rq) ____cacheline_aligned_in_smp;
387 p->rt.rt_rq = task_group(p)->rt_rq[cpu]; in set_task_rq()
466 struct rt_rq { struct
563 struct rt_rq rt;
8265 static void init_rt_rq(struct rt_rq *rt_rq, struct rq *rq) in init_rt_rq() argument
8270 array = &rt_rq->active; in init_rt_rq()
8279 rt_rq->highest_prio = MAX_RT_PRIO; in init_rt_rq()
8282 rt_rq->rt_nr_migratory = 0; in init_rt_rq()
8283 rt_rq->overloaded = 0; in init_rt_rq()
8286 rt_rq->rt_time = 0; in init_rt_rq()
8287 rt_rq->rt_throttled = 0; in init_rt_rq()
8288 rt_rq->rt_runtime = 0; in init_rt_rq()
8289 spin_lock_init(&rt_rq->rt_runtime_lock); in init_rt_rq()
8292 rt_rq->rt_nr_boosted = 0; in init_rt_rq()
8293 rt_rq->rq = rq; in init_rt_rq()
8327 static void init_tg_rt_entry(struct task_group *tg, struct rt_rq *rt_rq, in init_tg_rt_entry() argument
8333 tg->rt_rq[cpu] = rt_rq; in init_tg_rt_entry()
8334 init_rt_rq(rt_rq, rq); in init_tg_rt_entry()
8335 rt_rq->tg = tg; in init_tg_rt_entry()
8336 rt_rq->rt_se = rt_se; in init_tg_rt_entry()
8337 rt_rq->rt_runtime = tg->rt_bandwidth.rt_runtime; in init_tg_rt_entry()
8339 list_add(&rt_rq->leaf_rt_rq_list, &rq->leaf_rt_rq_list); in init_tg_rt_entry()
8346 rt_se->rt_rq = &rq->rt; in init_tg_rt_entry()
8348 rt_se->rt_rq = parent->my_q; in init_tg_rt_entry()
8350 rt_se->my_q = rt_rq; in init_tg_rt_entry()
8396 init_task_group.rt_rq = (struct rt_rq **)ptr; in sched_init()
8403 root_task_group.rt_rq = (struct rt_rq **)ptr; in sched_init()
8804 if (tg->rt_rq) in free_rt_sched_group()
8805 kfree(tg->rt_rq[i]); in free_rt_sched_group()
8810 kfree(tg->rt_rq); in free_rt_sched_group()
8817 struct rt_rq *rt_rq; in alloc_rt_sched_group() local
8822 tg->rt_rq = kzalloc(sizeof(rt_rq) * nr_cpu_ids, GFP_KERNEL); in alloc_rt_sched_group()
8823 if (!tg->rt_rq) in alloc_rt_sched_group()
8835 rt_rq = kzalloc_node(sizeof(struct rt_rq), in alloc_rt_sched_group()
8837 if (!rt_rq) in alloc_rt_sched_group()
8845 init_tg_rt_entry(tg, rt_rq, rt_se, i, 0, parent->rt_se[i]); in alloc_rt_sched_group()
8856 list_add_rcu(&tg->rt_rq[cpu]->leaf_rt_rq_list, in register_rt_sched_group()
8862 list_del_rcu(&tg->rt_rq[cpu]->leaf_rt_rq_list); in unregister_rt_sched_group()
9210 struct rt_rq *rt_rq = tg->rt_rq[i]; in tg_set_bandwidth() local
9212 spin_lock(&rt_rq->rt_runtime_lock); in tg_set_bandwidth()
9213 rt_rq->rt_runtime = rt_runtime; in tg_set_bandwidth()
9214 spin_unlock(&rt_rq->rt_runtime_lock); in tg_set_bandwidth()
9316 struct rt_rq *rt_rq = &cpu_rq(i)->rt; in sched_rt_global_constraints() local
9318 spin_lock(&rt_rq->rt_runtime_lock); in sched_rt_global_constraints()
9319 rt_rq->rt_runtime = global_rt_runtime(); in sched_rt_global_constraints()
9320 spin_unlock(&rt_rq->rt_runtime_lock); in sched_rt_global_constraints()