Searched refs:task_util (Results 1 – 3 of 3) sorted by relevance
| /kernel/linux/linux-5.10/include/trace/events/ |
| D | eas_sched.h | 6 int isolated, int idle, unsigned long task_util, 9 TP_ARGS(task, cpu, target_cpu, isolated, idle, task_util, cpu_util, cpu_cap), 19 __field(unsigned long, task_util) 32 __entry->task_util = task_util; 40 __entry->idle, __entry->task_util,
|
| /kernel/linux/linux-5.10/Documentation/scheduler/ |
| D | sched-capacity.rst | 140 task_util(p) = duty_cycle(p) 339 task_util(p) < capacity(task_cpu(p)) 348 clamp task_util() in the previous criterion. 358 clamp(task_util(p), task_uclamp_min(p), task_uclamp_max(p)) < capacity(cpu) 402 then it might become CPU-bound, IOW ``task_util(p) > capacity(task_cpu(p))``;
|
| /kernel/linux/linux-5.10/kernel/sched/ |
| D | fair.c | 3940 static inline unsigned long task_util(struct task_struct *p) in task_util() function 3962 return max(task_util(p), _task_util_est(p)); in task_util_est() 4065 ue.enqueued = task_util(p); in util_est_update() 4090 if (task_util(p) > capacity_orig_of(cpu_of(rq_of(cfs_rq)))) in util_est_update() 6469 unsigned long task_util, best_cap = 0; in select_idle_capacity() local 6476 task_util = uclamp_task_util(p); in select_idle_capacity() 6486 if (fits_capacity(task_util, cpu_cap)) in select_idle_capacity() 6498 static inline bool asym_fits_capacity(unsigned long task_util, int cpu) in asym_fits_capacity() argument 6501 return fits_capacity(task_util, capacity_of(cpu)); in asym_fits_capacity() 6512 unsigned long task_util; in select_idle_sibling() local [all …]
|