Searched refs:tasks (Results 1 – 13 of 13) sorted by relevance
/kernel/sched/ |
D | psi.c | 216 static bool test_state(unsigned int *tasks, enum psi_states state) in test_state() argument 220 return tasks[NR_IOWAIT]; in test_state() 222 return tasks[NR_IOWAIT] && !tasks[NR_RUNNING]; in test_state() 224 return tasks[NR_MEMSTALL]; in test_state() 226 return tasks[NR_MEMSTALL] && !tasks[NR_RUNNING]; in test_state() 228 return tasks[NR_RUNNING] > 1; in test_state() 230 return tasks[NR_IOWAIT] || tasks[NR_MEMSTALL] || in test_state() 231 tasks[NR_RUNNING]; in test_state() 697 if (groupc->tasks[t] == 0 && !psi_bug) { in psi_group_change() 699 cpu, t, groupc->tasks[0], in psi_group_change() [all …]
|
D | fair.c | 7370 struct list_head tasks; member 7608 struct list_head *tasks = &env->src_rq->cfs_tasks; in detach_tasks() local 7618 while (!list_empty(tasks)) { in detach_tasks() 7626 p = list_last_entry(tasks, struct task_struct, se.group_node); in detach_tasks() 7660 list_add(&p->se.group_node, &env->tasks); in detach_tasks() 7684 list_move(&p->se.group_node, tasks); in detach_tasks() 7729 struct list_head *tasks = &env->tasks; in attach_tasks() local 7736 while (!list_empty(tasks)) { in attach_tasks() 7737 p = list_first_entry(tasks, struct task_struct, se.group_node); in attach_tasks() 9158 .tasks = LIST_HEAD_INIT(env.tasks), in load_balance()
|
D | core.c | 895 if (!bucket[bucket_id].tasks) in uclamp_rq_max_value() 988 bucket->tasks++; in uclamp_rq_inc_id() 997 if (bucket->tasks == 1 || uc_se->value > bucket->value) in uclamp_rq_inc_id() 1052 SCHED_WARN_ON(!bucket->tasks); in uclamp_rq_dec_id() 1053 if (likely(bucket->tasks)) in uclamp_rq_dec_id() 1054 bucket->tasks--; in uclamp_rq_dec_id() 1064 if (likely(bucket->tasks)) in uclamp_rq_dec_id()
|
D | sched.h | 821 unsigned long tasks : BITS_PER_LONG - bits_per(SCHED_CAPACITY_SCALE); member
|
/kernel/ |
D | pid.c | 48 .tasks = { 216 INIT_HLIST_HEAD(&pid->tasks[type]); in alloc_pid() 286 hlist_add_head_rcu(&task->pid_links[type], &pid->tasks[type]); in attach_pid() 302 if (!hlist_empty(&pid->tasks[tmp])) in __change_pid() 334 first = rcu_dereference_check(hlist_first_rcu(&pid->tasks[type]), in pid_task()
|
D | exit.c | 128 list_del_rcu(&p->tasks); in __unhash_process() 1520 (!wo->wo_pid || hlist_empty(&wo->wo_pid->tasks[wo->wo_type]))) in do_wait()
|
D | fork.c | 2229 list_add_tail_rcu(&p->tasks, &init_task.tasks); in copy_process()
|
/kernel/debug/kdb/ |
D | kdb_cmds | 26 defcmd dumpcpu "" "Same as dumpall but only tasks on cpus"
|
/kernel/cgroup/ |
D | cgroup.c | 742 .tasks = LIST_HEAD_INIT(init_css_set.tasks), 782 return !list_empty(&cset->tasks) || !list_empty(&cset->mg_tasks); in css_set_populated() 913 &to_cset->tasks); in css_set_move_task() 1221 INIT_LIST_HEAD(&cset->tasks); in find_css_set() 1979 list_add_tail(&p->cg_list, &cset->tasks); in cgroup_enable_task_cg_lists() 2634 list_splice_tail_init(&cset->mg_tasks, &cset->tasks); in cgroup_migrate_execute() 3081 list_for_each_entry_safe(task, ntask, &ext_src_set->cset.tasks, cg_list) in cgroup_update_dfl_csses() 4611 if (!list_empty(&cset->tasks)) { in css_task_iter_advance_css_set() 4612 it->task_pos = cset->tasks.next; in css_task_iter_advance_css_set() 4613 it->cur_tasks_head = &cset->tasks; in css_task_iter_advance_css_set() [all …]
|
D | debug.c | 164 list_for_each_entry(task, &cset->tasks, cg_list) { in cgroup_css_links_read()
|
/kernel/time/ |
D | Kconfig | 106 the CPU is running tasks. Typically this requires running a single
|
/kernel/power/ |
D | Kconfig | 18 done, no tasks are frozen for suspend to RAM/standby.
|
/kernel/trace/ |
D | Kconfig | 257 if a system is reliable for Real Time tasks. 680 goes into an infinite loop (calling cond_sched() to let other tasks
|