Home
last modified time | relevance | path

Searched refs:tg (Results 1 – 5 of 5) sorted by relevance

/kernel/
Dsched.c304 user->tg->uid = user->uid; in set_tg_uid()
363 struct task_group *tg; in task_group() local
367 tg = __task_cred(p)->user->tg; in task_group()
370 tg = container_of(task_subsys_state(p, cpu_cgroup_subsys_id), in task_group()
373 tg = &init_task_group; in task_group()
375 return tg; in task_group()
436 struct task_group *tg; /* group that "owns" this runqueue */ member
487 struct task_group *tg; member
1451 static int tg_nop(struct task_group *tg, void *data) in tg_nop() argument
1483 update_group_shares_cpu(struct task_group *tg, int cpu, in update_group_shares_cpu() argument
[all …]
Duser.c60 .tg = &init_task_group,
97 sched_destroy_group(up->tg); in sched_destroy_user()
104 up->tg = sched_create_group(&root_task_group); in sched_create_user()
105 if (IS_ERR(up->tg)) in sched_create_user()
143 return sprintf(buf, "%lu\n", sched_group_shares(up->tg)); in cpu_shares_show()
156 rc = sched_group_set_shares(up->tg, shares); in cpu_shares_store()
172 return sprintf(buf, "%ld\n", sched_group_rt_runtime(up->tg)); in cpu_rt_runtime_show()
185 rc = sched_group_set_rt_runtime(up->tg, rt_runtime); in cpu_rt_runtime_store()
199 return sprintf(buf, "%lu\n", sched_group_rt_period(up->tg)); in cpu_rt_period_show()
212 rc = sched_group_set_rt_period(up->tg, rt_period); in cpu_rt_period_store()
[all …]
Dsched_debug.c58 struct task_group *tg) in print_cfs_group_stats() argument
60 struct sched_entity *se = tg->se[cpu]; in print_cfs_group_stats()
150 static void task_group_path(struct task_group *tg, char *buf, int buflen) in task_group_path() argument
153 if (!tg->css.cgroup) { in task_group_path()
157 cgroup_path(tg->css.cgroup, buf, buflen); in task_group_path()
171 struct task_group *tg = cfs_rq->tg; in print_cfs_rq() local
173 task_group_path(tg, path, sizeof(path)); in print_cfs_rq()
178 uid_t uid = cfs_rq->tg->uid; in print_cfs_rq()
217 print_cfs_group_stats(m, cpu, cfs_rq->tg); in print_cfs_rq()
225 struct task_group *tg = rt_rq->tg; in print_rt_rq() local
[all …]
Dsched_fair.c124 return cfs_rq->tg->cfs_rq[this_cpu]; in cpu_cfs_rq()
1126 static long effective_load(struct task_group *tg, int cpu, in effective_load() argument
1129 struct sched_entity *se = tg->se[cpu]; in effective_load()
1131 if (!tg->parent) in effective_load()
1153 S = se->my_q->tg->shares; in effective_load()
1180 static inline unsigned long effective_load(struct task_group *tg, int cpu, in effective_load() argument
1195 struct task_group *tg; in wake_affine() local
1214 tg = task_group(current); in wake_affine()
1217 tl += effective_load(tg, this_cpu, -weight, -weight); in wake_affine()
1218 load += effective_load(tg, prev_cpu, 0, -weight); in wake_affine()
[all …]
Dsched_rt.c68 if (!rt_rq->tg) in sched_rt_runtime()
76 return ktime_to_ns(rt_rq->tg->rt_bandwidth.rt_period); in sched_rt_period()
161 return &rt_rq->tg->rt_bandwidth; in sched_rt_bandwidth()
582 if (rt_rq->tg) in inc_rt_tasks()
583 start_rt_bandwidth(&rt_rq->tg->rt_bandwidth); in inc_rt_tasks()