Lines Matching refs:rd
286 return atomic_read(&rq->rd->rto_count); in rt_overloaded()
294 cpumask_set_cpu(rq->cpu, rq->rd->rto_mask); in rt_set_overload()
305 atomic_inc(&rq->rd->rto_count); in rt_set_overload()
314 atomic_dec(&rq->rd->rto_count); in rt_clear_overload()
315 cpumask_clear_cpu(rq->cpu, rq->rd->rto_mask); in rt_clear_overload()
598 return this_rq()->rd->span; in sched_rt_period_mask()
697 struct root_domain *rd = rq_of_rt_rq(rt_rq)->rd; in do_balance_runtime() local
701 weight = cpumask_weight(rd->span); in do_balance_runtime()
705 for_each_cpu(i, rd->span) { in do_balance_runtime()
748 struct root_domain *rd = rq->rd; in __disable_runtime() local
782 for_each_cpu(i, rd->span) { in __disable_runtime()
1118 cpupri_set(&rq->rd->cpupri, rq->cpu, prio); in inc_rt_prio_smp()
1134 cpupri_set(&rq->rd->cpupri, rq->cpu, rt_rq->highest_prio.curr); in dec_rt_prio_smp()
1615 !cpupri_find(&rq->rd->cpupri, rq->curr, NULL)) in check_preempt_equal_prio()
1623 cpupri_find(&rq->rd->cpupri, p, NULL)) in check_preempt_equal_prio()
1835 ret = cpupri_find_fitness(&task_rq(task)->rd->cpupri, in find_lowest_rq()
1840 ret = cpupri_find(&task_rq(task)->rd->cpupri, in find_lowest_rq()
2167 static int rto_next_cpu(struct root_domain *rd) in rto_next_cpu() argument
2188 cpu = cpumask_next(rd->rto_cpu, rd->rto_mask); in rto_next_cpu()
2191 trace_android_rvh_rto_next_cpu(rd->rto_cpu, rd->rto_mask, &cpu); in rto_next_cpu()
2193 rd->rto_cpu = cpu; in rto_next_cpu()
2198 rd->rto_cpu = -1; in rto_next_cpu()
2206 next = atomic_read_acquire(&rd->rto_loop_next); in rto_next_cpu()
2208 if (rd->rto_loop == next) in rto_next_cpu()
2211 rd->rto_loop = next; in rto_next_cpu()
2232 atomic_inc(&rq->rd->rto_loop_next); in tell_cpu_to_push()
2235 if (!rto_start_trylock(&rq->rd->rto_loop_start)) in tell_cpu_to_push()
2238 raw_spin_lock(&rq->rd->rto_lock); in tell_cpu_to_push()
2246 if (rq->rd->rto_cpu < 0) in tell_cpu_to_push()
2247 cpu = rto_next_cpu(rq->rd); in tell_cpu_to_push()
2249 raw_spin_unlock(&rq->rd->rto_lock); in tell_cpu_to_push()
2251 rto_start_unlock(&rq->rd->rto_loop_start); in tell_cpu_to_push()
2255 sched_get_rd(rq->rd); in tell_cpu_to_push()
2256 irq_work_queue_on(&rq->rd->rto_push_work, cpu); in tell_cpu_to_push()
2263 struct root_domain *rd = in rto_push_irq_work_func() local
2281 raw_spin_lock(&rd->rto_lock); in rto_push_irq_work_func()
2284 cpu = rto_next_cpu(rd); in rto_push_irq_work_func()
2286 raw_spin_unlock(&rd->rto_lock); in rto_push_irq_work_func()
2289 sched_put_rd(rd); in rto_push_irq_work_func()
2294 irq_work_queue_on(&rd->rto_push_work, cpu); in rto_push_irq_work_func()
2317 cpumask_test_cpu(this_rq->cpu, this_rq->rd->rto_mask)) in pull_rt_task()
2327 for_each_cpu(cpu, this_rq->rd->rto_mask) { in pull_rt_task()
2434 cpupri_set(&rq->rd->cpupri, rq->cpu, rq->rt.highest_prio.curr); in rq_online_rt()
2445 cpupri_set(&rq->rd->cpupri, rq->cpu, CPUPRI_INVALID); in rq_offline_rt()