Home
last modified time | relevance | path

Searched refs:tasks (Results 1 – 14 of 14) sorted by relevance

/kernel/sched/
Dpsi.c226 static bool test_state(unsigned int *tasks, enum psi_states state) in test_state() argument
230 return tasks[NR_IOWAIT]; in test_state()
232 return tasks[NR_IOWAIT] && !tasks[NR_RUNNING]; in test_state()
234 return tasks[NR_MEMSTALL]; in test_state()
236 return tasks[NR_MEMSTALL] && !tasks[NR_RUNNING]; in test_state()
238 return tasks[NR_RUNNING] > tasks[NR_ONCPU]; in test_state()
240 return tasks[NR_IOWAIT] || tasks[NR_MEMSTALL] || in test_state()
241 tasks[NR_RUNNING]; in test_state()
769 if (groupc->tasks[t]) { in psi_group_change()
770 groupc->tasks[t]--; in psi_group_change()
[all …]
Dfair.c7770 struct list_head tasks; member
8025 struct list_head *tasks = &env->src_rq->cfs_tasks; in detach_tasks() local
8035 while (!list_empty(tasks)) { in detach_tasks()
8043 p = list_last_entry(tasks, struct task_struct, se.group_node); in detach_tasks()
8110 list_add(&p->se.group_node, &env->tasks); in detach_tasks()
8133 list_move(&p->se.group_node, tasks); in detach_tasks()
8178 struct list_head *tasks = &env->tasks; in attach_tasks() local
8185 while (!list_empty(tasks)) { in attach_tasks()
8186 p = list_first_entry(tasks, struct task_struct, se.group_node); in attach_tasks()
9955 .tasks = LIST_HEAD_INIT(env.tasks), in load_balance()
Dcore.c1036 if (!bucket[bucket_id].tasks) in uclamp_rq_max_value()
1194 bucket->tasks++; in uclamp_rq_inc_id()
1203 if (bucket->tasks == 1 || uc_se->value > bucket->value) in uclamp_rq_inc_id()
1258 SCHED_WARN_ON(!bucket->tasks); in uclamp_rq_dec_id()
1259 if (likely(bucket->tasks)) in uclamp_rq_dec_id()
1260 bucket->tasks--; in uclamp_rq_dec_id()
1270 if (likely(bucket->tasks)) in uclamp_rq_dec_id()
Dsched.h877 unsigned long tasks : BITS_PER_LONG - bits_per(SCHED_CAPACITY_SCALE); member
/kernel/
Dpid.c50 .tasks = {
263 INIT_HLIST_HEAD(&pid->tasks[type]); in alloc_pid()
334 hlist_add_head_rcu(&task->pid_links[type], &pid->tasks[type]); in attach_pid()
372 struct hlist_head *head1 = &pid1->tasks[PIDTYPE_PID]; in exchange_tids()
373 struct hlist_head *head2 = &pid2->tasks[PIDTYPE_PID]; in exchange_tids()
401 first = rcu_dereference_check(hlist_first_rcu(&pid->tasks[type]), in pid_task()
Dfork.c2367 list_add_tail_rcu(&p->tasks, &init_task.tasks); in copy_process()
Dexit.c130 list_del_rcu(&p->tasks); in __unhash_process()
/kernel/debug/kdb/
Dkdb_cmds26 defcmd dumpcpu "" "Same as dumpall but only tasks on cpus"
/kernel/cgroup/
Dcgroup.c754 .tasks = LIST_HEAD_INIT(init_css_set.tasks),
794 return !list_empty(&cset->tasks) || !list_empty(&cset->mg_tasks); in css_set_populated()
924 &to_cset->tasks); in css_set_move_task()
1232 INIT_LIST_HEAD(&cset->tasks); in find_css_set()
2583 list_splice_tail_init(&cset->mg_tasks, &cset->tasks); in cgroup_migrate_execute()
3031 list_for_each_entry_safe(task, ntask, &ext_src_set->cset.tasks, cg_list) in cgroup_update_dfl_csses()
4578 if (!list_empty(&cset->tasks)) { in css_task_iter_advance_css_set()
4579 it->cur_tasks_head = &cset->tasks; in css_task_iter_advance_css_set()
4647 if (it->task_pos == &it->cur_cset->tasks) { in css_task_iter_advance()
5819 BUG_ON(!list_empty(&init_task.tasks)); in cgroup_init_subsys()
Ddebug.c164 list_for_each_entry(task, &cset->tasks, cg_list) { in cgroup_css_links_read()
/kernel/time/
DKconfig115 the CPU is running tasks. Typically this requires running a single
/kernel/rcu/
DKconfig246 to CPUs executing in userspace or idle during tasks trace
/kernel/power/
DKconfig18 done, no tasks are frozen for suspend to RAM/standby.
/kernel/trace/
DKconfig340 if a system is reliable for Real Time tasks.
703 goes into an infinite loop (calling cond_sched() to let other tasks