| /kernel/sched/ |
| D | deadline.c | 38 struct rq *rq = task_rq(p); in dl_rq_of_se() local 92 static inline int dl_overloaded(struct rq *rq) in dl_overloaded() 97 static inline void dl_set_overload(struct rq *rq) in dl_set_overload() 113 static inline void dl_clear_overload(struct rq *rq) in dl_clear_overload() 159 static void enqueue_pushable_dl_task(struct rq *rq, struct task_struct *p) in enqueue_pushable_dl_task() 190 static void dequeue_pushable_dl_task(struct rq *rq, struct task_struct *p) in dequeue_pushable_dl_task() 212 static inline int has_pushable_dl_tasks(struct rq *rq) in has_pushable_dl_tasks() 219 static inline bool need_pull_dl_task(struct rq *rq, struct task_struct *prev) in need_pull_dl_task() 230 static inline void queue_push_tasks(struct rq *rq) in queue_push_tasks() 238 static inline void queue_pull_task(struct rq *rq) in queue_pull_task() [all …]
|
| D | sched.h | 45 static inline void check_for_migration(struct rq *rq, struct task_struct *p) { } in check_for_migration() 51 static inline void update_idle_core(struct rq *rq) { } in update_idle_core() 436 struct rq *rq; /* cpu runqueue to which this cfs_rq is attached */ member 507 struct rq *rq; member 628 struct rq { struct 630 raw_spinlock_t lock; 636 unsigned int nr_running; 638 unsigned int nr_numa_running; 639 unsigned int nr_preferred_running; 642 unsigned long cpu_load[CPU_LOAD_IDX_MAX]; [all …]
|
| D | stop_task.c | 22 check_preempt_curr_stop(struct rq *rq, struct task_struct *p, int flags) in check_preempt_curr_stop() 28 pick_next_task_stop(struct rq *rq, struct task_struct *prev, struct pin_cookie cookie) in pick_next_task_stop() 43 enqueue_task_stop(struct rq *rq, struct task_struct *p, int flags) in enqueue_task_stop() 50 dequeue_task_stop(struct rq *rq, struct task_struct *p, int flags) in dequeue_task_stop() 56 static void yield_task_stop(struct rq *rq) in yield_task_stop() 61 static void put_prev_task_stop(struct rq *rq, struct task_struct *prev) in put_prev_task_stop() 80 static void task_tick_stop(struct rq *rq, struct task_struct *curr, int queued) in task_tick_stop() 84 static void set_curr_task_stop(struct rq *rq) in set_curr_task_stop() 91 static void switched_to_stop(struct rq *rq, struct task_struct *p) in switched_to_stop() 97 prio_changed_stop(struct rq *rq, struct task_struct *p, int oldprio) in prio_changed_stop() [all …]
|
| D | rt.c | 161 struct rq *rq = cpu_rq(cpu); in init_tg_rt_entry() local 247 struct rq *rq = rq_of_rt_se(rt_se); in rt_rq_of_se() local 264 static inline bool need_pull_rt_task(struct rq *rq, struct task_struct *prev) in need_pull_rt_task() 270 static inline int rt_overloaded(struct rq *rq) in rt_overloaded() 275 static inline void rt_set_overload(struct rq *rq) in rt_set_overload() 294 static inline void rt_clear_overload(struct rq *rq) in rt_clear_overload() 351 static inline int has_pushable_tasks(struct rq *rq) in has_pushable_tasks() 362 static inline void queue_push_tasks(struct rq *rq) in queue_push_tasks() 370 static inline void queue_pull_task(struct rq *rq) in queue_pull_task() 375 static void enqueue_pushable_task(struct rq *rq, struct task_struct *p) in enqueue_pushable_task() [all …]
|
| D | stats.h | 8 rq_sched_info_arrive(struct rq *rq, unsigned long long delta) in rq_sched_info_arrive() 20 rq_sched_info_depart(struct rq *rq, unsigned long long delta) in rq_sched_info_depart() 27 rq_sched_info_dequeued(struct rq *rq, unsigned long long delta) in rq_sched_info_dequeued() 41 rq_sched_info_arrive(struct rq *rq, unsigned long long delta) in rq_sched_info_arrive() 44 rq_sched_info_dequeued(struct rq *rq, unsigned long long delta) in rq_sched_info_dequeued() 47 rq_sched_info_depart(struct rq *rq, unsigned long long delta) in rq_sched_info_depart() 69 static inline void sched_info_dequeued(struct rq *rq, struct task_struct *t) in sched_info_dequeued() 87 static void sched_info_arrive(struct rq *rq, struct task_struct *t) in sched_info_arrive() 106 static inline void sched_info_queued(struct rq *rq, struct task_struct *t) in sched_info_queued() 121 static inline void sched_info_depart(struct rq *rq, struct task_struct *t) in sched_info_depart() [all …]
|
| D | idle_task.c | 21 static void check_preempt_curr_idle(struct rq *rq, struct task_struct *p, int flags) in check_preempt_curr_idle() 27 pick_next_task_idle(struct rq *rq, struct task_struct *prev, struct pin_cookie cookie) in pick_next_task_idle() 40 dequeue_task_idle(struct rq *rq, struct task_struct *p, int flags) in dequeue_task_idle() 48 static void put_prev_task_idle(struct rq *rq, struct task_struct *prev) in put_prev_task_idle() 53 static void task_tick_idle(struct rq *rq, struct task_struct *curr, int queued) in task_tick_idle() 57 static void set_curr_task_idle(struct rq *rq) in set_curr_task_idle() 61 static void switched_to_idle(struct rq *rq, struct task_struct *p) in switched_to_idle() 67 prio_changed_idle(struct rq *rq, struct task_struct *p, int oldprio) in prio_changed_idle() 72 static unsigned int get_rr_interval_idle(struct rq *rq, struct task_struct *task) in get_rr_interval_idle() 77 static void update_curr_idle(struct rq *rq) in update_curr_idle()
|
| D | walt.c | 66 static inline void fixup_cum_window_demand(struct rq *rq, s64 delta) in fixup_cum_window_demand() 74 walt_inc_cumulative_runnable_avg(struct rq *rq, in walt_inc_cumulative_runnable_avg() 91 walt_dec_cumulative_runnable_avg(struct rq *rq, in walt_dec_cumulative_runnable_avg() 107 fixup_cumulative_runnable_avg(struct rq *rq, in fixup_cumulative_runnable_avg() 204 update_window_start(struct rq *rq, u64 wallclock) in update_window_start() 230 static u64 scale_exec_time(u64 delta, struct rq *rq) in scale_exec_time() 237 static int cpu_is_waiting_on_io(struct rq *rq) in cpu_is_waiting_on_io() 248 struct rq *rq = cpu_rq(cpu); in walt_account_irqtime() local 289 struct rq *rq = cpu_rq(cpu); in walt_irqload() local 310 static int account_busy_for_cpu_time(struct rq *rq, struct task_struct *p, in account_busy_for_cpu_time() [all …]
|
| D | walt.h | 42 static inline void walt_update_task_ravg(struct task_struct *p, struct rq *rq, in walt_update_task_ravg() 44 static inline void walt_inc_cumulative_runnable_avg(struct rq *rq, struct task_struct *p) { } in walt_inc_cumulative_runnable_avg() 45 static inline void walt_dec_cumulative_runnable_avg(struct rq *rq, struct task_struct *p) { } in walt_dec_cumulative_runnable_avg() 46 static inline void walt_inc_cfs_cumulative_runnable_avg(struct cfs_rq *rq, in walt_inc_cfs_cumulative_runnable_avg() 48 static inline void walt_dec_cfs_cumulative_runnable_avg(struct cfs_rq *rq, in walt_dec_cfs_cumulative_runnable_avg() 53 static inline void walt_set_window_start(struct rq *rq) { } in walt_set_window_start()
|
| D | core.c | 101 void update_rq_clock(struct rq *rq) in update_rq_clock() 168 unlock_rq_of(struct rq *rq, struct task_struct *p, struct rq_flags *flags) in unlock_rq_of() 179 struct rq *rq; in this_rq_lock() local 194 struct rq *rq; in __task_rq_lock() local 219 struct rq *rq; in task_rq_lock() local 258 static void hrtick_clear(struct rq *rq) in hrtick_clear() 270 struct rq *rq = container_of(timer, struct rq, hrtick_timer); in hrtick() local 284 static void __hrtick_restart(struct rq *rq) in __hrtick_restart() 296 struct rq *rq = arg; in __hrtick_start() local 309 void hrtick_start(struct rq *rq, u64 delay) in hrtick_start() [all …]
|
| D | fair.c | 305 struct rq *rq = rq_of(cfs_rq); in list_add_leaf_cfs_rq() local 373 #define for_each_leaf_cfs_rq(rq, cfs_rq) \ argument 448 struct rq *rq = task_rq(p); in cfs_rq_of() local 467 #define for_each_leaf_cfs_rq(rq, cfs_rq) \ argument 881 static void update_curr_fair(struct rq *rq) in update_curr_fair() 1128 static void account_numa_enqueue(struct rq *rq, struct task_struct *p) in account_numa_enqueue() 1134 static void account_numa_dequeue(struct rq *rq, struct task_struct *p) in account_numa_dequeue() 1418 struct rq *rq = cpu_rq(cpu); in update_numa_stats() local 2566 void task_tick_numa(struct rq *rq, struct task_struct *curr) in task_tick_numa() 2598 static void task_tick_numa(struct rq *rq, struct task_struct *curr) in task_tick_numa() [all …]
|
| D | cputime.c | 272 struct rq *rq = this_rq(); in account_idle_time() local 336 struct rq *rq; in read_sum_exec_runtime() local 415 struct rq *rq, int ticks) in irqtime_account_process_tick() 453 struct rq *rq = this_rq(); in irqtime_account_idle_ticks() local 460 struct rq *rq, int nr_ticks) {} in irqtime_account_process_tick() 531 struct rq *rq = this_rq(); in account_process_tick() local
|
| D | debug.c | 487 print_task(struct seq_file *m, struct rq *rq, struct task_struct *p) in print_task() 515 static void print_rq(struct seq_file *m, struct rq *rq, int rq_cpu) in print_rq() 540 struct rq *rq = cpu_rq(cpu); in print_cfs_rq() local 648 struct rq *rq = cpu_rq(cpu); in print_cpu() local
|
| D | stats.c | 45 struct rq *rq; in show_schedstat() local
|
| D | tune.c | 379 struct rq *rq; in schedtune_can_attach() local 499 struct rq *rq; in schedtune_exit_task() local
|
| D | cpufreq_schedutil.c | 206 struct rq *rq = cpu_rq(cpu); in sugov_get_util() local
|
| /kernel/trace/ |
| D | blktrace.c | 722 static void blk_add_trace_rq(struct request_queue *q, struct request *rq, in blk_add_trace_rq() 742 struct request_queue *q, struct request *rq) in blk_add_trace_rq_abort() 748 struct request_queue *q, struct request *rq) in blk_add_trace_rq_insert() 754 struct request_queue *q, struct request *rq) in blk_add_trace_rq_issue() 761 struct request *rq) in blk_add_trace_rq_requeue() 768 struct request *rq, in blk_add_trace_rq_complete() 812 struct request *rq, in blk_add_trace_bio_backmerge() 820 struct request *rq, in blk_add_trace_bio_frontmerge() 952 struct request *rq, dev_t dev, in blk_add_trace_rq_remap() 982 struct request *rq, in blk_add_driver_data() [all …]
|