/kernel/ |
D | sched_rt.c | 8 static inline int rt_overloaded(struct rq *rq) in rt_overloaded() argument 10 return atomic_read(&rq->rd->rto_count); in rt_overloaded() 13 static inline void rt_set_overload(struct rq *rq) in rt_set_overload() argument 15 if (!rq->online) in rt_set_overload() 18 cpumask_set_cpu(rq->cpu, rq->rd->rto_mask); in rt_set_overload() 27 atomic_inc(&rq->rd->rto_count); in rt_set_overload() 30 static inline void rt_clear_overload(struct rq *rq) in rt_clear_overload() argument 32 if (!rq->online) in rt_clear_overload() 36 atomic_dec(&rq->rd->rto_count); in rt_clear_overload() 37 cpumask_clear_cpu(rq->cpu, rq->rd->rto_mask); in rt_clear_overload() [all …]
|
D | sched_idletask.c | 17 static void check_preempt_curr_idle(struct rq *rq, struct task_struct *p, int sync) in check_preempt_curr_idle() argument 19 resched_task(rq->idle); in check_preempt_curr_idle() 22 static struct task_struct *pick_next_task_idle(struct rq *rq) in pick_next_task_idle() argument 24 schedstat_inc(rq, sched_goidle); in pick_next_task_idle() 26 return rq->idle; in pick_next_task_idle() 34 dequeue_task_idle(struct rq *rq, struct task_struct *p, int sleep) in dequeue_task_idle() argument 36 spin_unlock_irq(&rq->lock); in dequeue_task_idle() 39 spin_lock_irq(&rq->lock); in dequeue_task_idle() 42 static void put_prev_task_idle(struct rq *rq, struct task_struct *prev) in put_prev_task_idle() argument 48 load_balance_idle(struct rq *this_rq, int this_cpu, struct rq *busiest, in load_balance_idle() [all …]
|
D | sched_stats.h | 21 struct rq *rq = cpu_rq(cpu); in show_schedstat() local 30 cpu, rq->yld_both_empty, in show_schedstat() 31 rq->yld_act_empty, rq->yld_exp_empty, rq->yld_count, in show_schedstat() 32 rq->sched_switch, rq->sched_count, rq->sched_goidle, in show_schedstat() 33 rq->ttwu_count, rq->ttwu_local, in show_schedstat() 34 rq->rq_cpu_time, in show_schedstat() 35 rq->rq_sched_info.run_delay, rq->rq_sched_info.pcount); in show_schedstat() 112 rq_sched_info_arrive(struct rq *rq, unsigned long long delta) in rq_sched_info_arrive() argument 114 if (rq) { in rq_sched_info_arrive() 115 rq->rq_sched_info.run_delay += delta; in rq_sched_info_arrive() [all …]
|
D | sched.c | 129 static void double_rq_lock(struct rq *rq1, struct rq *rq2); 425 struct rq *rq; /* cpu runqueue to which this cfs_rq is attached */ member 485 struct rq *rq; member 541 struct rq { struct 640 static DEFINE_PER_CPU_SHARED_ALIGNED(struct rq, runqueues); argument 642 static inline void check_preempt_curr(struct rq *rq, struct task_struct *p, int sync) in check_preempt_curr() argument 644 rq->curr->sched_class->check_preempt_curr(rq, p, sync); in check_preempt_curr() 647 static inline int cpu_of(struct rq *rq) in cpu_of() argument 650 return rq->cpu; in cpu_of() 671 static inline void update_rq_clock(struct rq *rq) in update_rq_clock() argument [all …]
|
D | sched_fair.c | 90 static inline struct rq *rq_of(struct cfs_rq *cfs_rq) in rq_of() 92 return cfs_rq->rq; in rq_of() 128 #define for_each_leaf_cfs_rq(rq, cfs_rq) \ argument 129 list_for_each_entry_rcu(cfs_rq, &rq->leaf_cfs_rq_list, leaf_cfs_rq_list) 191 static inline struct rq *rq_of(struct cfs_rq *cfs_rq) in rq_of() 193 return container_of(cfs_rq, struct rq, cfs); in rq_of() 209 struct rq *rq = task_rq(p); in cfs_rq_of() local 211 return &rq->cfs; in cfs_rq_of() 225 #define for_each_leaf_cfs_rq(rq, cfs_rq) \ argument 226 for (cfs_rq = &rq->cfs; cfs_rq; cfs_rq = NULL) [all …]
|
D | sched_debug.c | 91 print_task(struct seq_file *m, struct rq *rq, struct task_struct *p) in print_task() argument 93 if (rq->curr == p) in print_task() 124 static void print_rq(struct seq_file *m, struct rq *rq, int rq_cpu) in print_rq() argument 142 print_task(m, rq, p); in print_rq() 165 struct rq *rq = &per_cpu(runqueues, cpu); in print_cfs_rq() local 187 spin_lock_irqsave(&rq->lock, flags); in print_cfs_rq() 195 spin_unlock_irqrestore(&rq->lock, flags); in print_cfs_rq() 251 struct rq *rq = &per_cpu(runqueues, cpu); in print_cpu() local 265 SEQ_printf(m, " .%-30s: %Ld\n", #x, (long long)(rq->x)) in print_cpu() 267 SEQ_printf(m, " .%-30s: %Ld.%06ld\n", #x, SPLIT_NS(rq->x)) in print_cpu() [all …]
|
/kernel/trace/ |
D | trace_sched_wakeup.c | 108 probe_wakeup_sched_switch(struct rq *rq, struct task_struct *prev, in probe_wakeup_sched_switch() argument 214 probe_wakeup(struct rq *rq, struct task_struct *p, int success) in probe_wakeup() argument
|
D | trace_sched_switch.c | 23 probe_sched_switch(struct rq *__rq, struct task_struct *prev, in probe_sched_switch() 52 probe_sched_wakeup(struct rq *__rq, struct task_struct *wakee, int success) in probe_sched_wakeup()
|