Home
last modified time | relevance | path

Searched refs:rq (Results 1 – 8 of 8) sorted by relevance

/kernel/
Dsched_rt.c8 static inline int rt_overloaded(struct rq *rq) in rt_overloaded() argument
10 return atomic_read(&rq->rd->rto_count); in rt_overloaded()
13 static inline void rt_set_overload(struct rq *rq) in rt_set_overload() argument
15 if (!rq->online) in rt_set_overload()
18 cpumask_set_cpu(rq->cpu, rq->rd->rto_mask); in rt_set_overload()
27 atomic_inc(&rq->rd->rto_count); in rt_set_overload()
30 static inline void rt_clear_overload(struct rq *rq) in rt_clear_overload() argument
32 if (!rq->online) in rt_clear_overload()
36 atomic_dec(&rq->rd->rto_count); in rt_clear_overload()
37 cpumask_clear_cpu(rq->cpu, rq->rd->rto_mask); in rt_clear_overload()
[all …]
Dsched_idletask.c17 static void check_preempt_curr_idle(struct rq *rq, struct task_struct *p, int sync) in check_preempt_curr_idle() argument
19 resched_task(rq->idle); in check_preempt_curr_idle()
22 static struct task_struct *pick_next_task_idle(struct rq *rq) in pick_next_task_idle() argument
24 schedstat_inc(rq, sched_goidle); in pick_next_task_idle()
26 return rq->idle; in pick_next_task_idle()
34 dequeue_task_idle(struct rq *rq, struct task_struct *p, int sleep) in dequeue_task_idle() argument
36 spin_unlock_irq(&rq->lock); in dequeue_task_idle()
39 spin_lock_irq(&rq->lock); in dequeue_task_idle()
42 static void put_prev_task_idle(struct rq *rq, struct task_struct *prev) in put_prev_task_idle() argument
48 load_balance_idle(struct rq *this_rq, int this_cpu, struct rq *busiest, in load_balance_idle()
[all …]
Dsched_stats.h21 struct rq *rq = cpu_rq(cpu); in show_schedstat() local
30 cpu, rq->yld_both_empty, in show_schedstat()
31 rq->yld_act_empty, rq->yld_exp_empty, rq->yld_count, in show_schedstat()
32 rq->sched_switch, rq->sched_count, rq->sched_goidle, in show_schedstat()
33 rq->ttwu_count, rq->ttwu_local, in show_schedstat()
34 rq->rq_cpu_time, in show_schedstat()
35 rq->rq_sched_info.run_delay, rq->rq_sched_info.pcount); in show_schedstat()
112 rq_sched_info_arrive(struct rq *rq, unsigned long long delta) in rq_sched_info_arrive() argument
114 if (rq) { in rq_sched_info_arrive()
115 rq->rq_sched_info.run_delay += delta; in rq_sched_info_arrive()
[all …]
Dsched.c129 static void double_rq_lock(struct rq *rq1, struct rq *rq2);
425 struct rq *rq; /* cpu runqueue to which this cfs_rq is attached */ member
485 struct rq *rq; member
541 struct rq { struct
640 static DEFINE_PER_CPU_SHARED_ALIGNED(struct rq, runqueues); argument
642 static inline void check_preempt_curr(struct rq *rq, struct task_struct *p, int sync) in check_preempt_curr() argument
644 rq->curr->sched_class->check_preempt_curr(rq, p, sync); in check_preempt_curr()
647 static inline int cpu_of(struct rq *rq) in cpu_of() argument
650 return rq->cpu; in cpu_of()
671 static inline void update_rq_clock(struct rq *rq) in update_rq_clock() argument
[all …]
Dsched_fair.c90 static inline struct rq *rq_of(struct cfs_rq *cfs_rq) in rq_of()
92 return cfs_rq->rq; in rq_of()
128 #define for_each_leaf_cfs_rq(rq, cfs_rq) \ argument
129 list_for_each_entry_rcu(cfs_rq, &rq->leaf_cfs_rq_list, leaf_cfs_rq_list)
191 static inline struct rq *rq_of(struct cfs_rq *cfs_rq) in rq_of()
193 return container_of(cfs_rq, struct rq, cfs); in rq_of()
209 struct rq *rq = task_rq(p); in cfs_rq_of() local
211 return &rq->cfs; in cfs_rq_of()
225 #define for_each_leaf_cfs_rq(rq, cfs_rq) \ argument
226 for (cfs_rq = &rq->cfs; cfs_rq; cfs_rq = NULL)
[all …]
Dsched_debug.c91 print_task(struct seq_file *m, struct rq *rq, struct task_struct *p) in print_task() argument
93 if (rq->curr == p) in print_task()
124 static void print_rq(struct seq_file *m, struct rq *rq, int rq_cpu) in print_rq() argument
142 print_task(m, rq, p); in print_rq()
165 struct rq *rq = &per_cpu(runqueues, cpu); in print_cfs_rq() local
187 spin_lock_irqsave(&rq->lock, flags); in print_cfs_rq()
195 spin_unlock_irqrestore(&rq->lock, flags); in print_cfs_rq()
251 struct rq *rq = &per_cpu(runqueues, cpu); in print_cpu() local
265 SEQ_printf(m, " .%-30s: %Ld\n", #x, (long long)(rq->x)) in print_cpu()
267 SEQ_printf(m, " .%-30s: %Ld.%06ld\n", #x, SPLIT_NS(rq->x)) in print_cpu()
[all …]
/kernel/trace/
Dtrace_sched_wakeup.c108 probe_wakeup_sched_switch(struct rq *rq, struct task_struct *prev, in probe_wakeup_sched_switch() argument
214 probe_wakeup(struct rq *rq, struct task_struct *p, int success) in probe_wakeup() argument
Dtrace_sched_switch.c23 probe_sched_switch(struct rq *__rq, struct task_struct *prev, in probe_sched_switch()
52 probe_sched_wakeup(struct rq *__rq, struct task_struct *wakee, int success) in probe_sched_wakeup()