• Home
  • Raw
  • Download

Lines Matching refs:rd

335 	return atomic_read(&rq->rd->rto_count);  in rt_overloaded()
343 cpumask_set_cpu(rq->cpu, rq->rd->rto_mask); in rt_set_overload()
354 atomic_inc(&rq->rd->rto_count); in rt_set_overload()
363 atomic_dec(&rq->rd->rto_count); in rt_clear_overload()
364 cpumask_clear_cpu(rq->cpu, rq->rd->rto_mask); in rt_clear_overload()
638 return this_rq()->rd->span; in sched_rt_period_mask()
737 struct root_domain *rd = rq_of_rt_rq(rt_rq)->rd; in do_balance_runtime() local
741 weight = cpumask_weight(rd->span); in do_balance_runtime()
745 for_each_cpu(i, rd->span) { in do_balance_runtime()
788 struct root_domain *rd = rq->rd; in __disable_runtime() local
822 for_each_cpu(i, rd->span) { in __disable_runtime()
1156 cpupri_set(&rq->rd->cpupri, rq->cpu, prio); in inc_rt_prio_smp()
1172 cpupri_set(&rq->rd->cpupri, rq->cpu, rt_rq->highest_prio.curr); in dec_rt_prio_smp()
1764 !cpupri_find(&rq->rd->cpupri, rq->curr, NULL)) in check_preempt_equal_prio()
1772 cpupri_find(&rq->rd->cpupri, p, NULL)) in check_preempt_equal_prio()
1998 ret = cpupri_find_fitness(&task_rq(task)->rd->cpupri, in find_lowest_rq()
2003 ret = cpupri_find(&task_rq(task)->rd->cpupri, in find_lowest_rq()
2330 static int rto_next_cpu(struct root_domain *rd) in rto_next_cpu() argument
2351 cpu = cpumask_next(rd->rto_cpu, rd->rto_mask); in rto_next_cpu()
2354 trace_android_rvh_rto_next_cpu(rd->rto_cpu, rd->rto_mask, &cpu); in rto_next_cpu()
2356 rd->rto_cpu = cpu; in rto_next_cpu()
2361 rd->rto_cpu = -1; in rto_next_cpu()
2369 next = atomic_read_acquire(&rd->rto_loop_next); in rto_next_cpu()
2371 if (rd->rto_loop == next) in rto_next_cpu()
2374 rd->rto_loop = next; in rto_next_cpu()
2395 atomic_inc(&rq->rd->rto_loop_next); in tell_cpu_to_push()
2398 if (!rto_start_trylock(&rq->rd->rto_loop_start)) in tell_cpu_to_push()
2401 raw_spin_lock(&rq->rd->rto_lock); in tell_cpu_to_push()
2409 if (rq->rd->rto_cpu < 0) in tell_cpu_to_push()
2410 cpu = rto_next_cpu(rq->rd); in tell_cpu_to_push()
2412 raw_spin_unlock(&rq->rd->rto_lock); in tell_cpu_to_push()
2414 rto_start_unlock(&rq->rd->rto_loop_start); in tell_cpu_to_push()
2418 sched_get_rd(rq->rd); in tell_cpu_to_push()
2419 irq_work_queue_on(&rq->rd->rto_push_work, cpu); in tell_cpu_to_push()
2426 struct root_domain *rd = in rto_push_irq_work_func() local
2444 raw_spin_lock(&rd->rto_lock); in rto_push_irq_work_func()
2447 cpu = rto_next_cpu(rd); in rto_push_irq_work_func()
2449 raw_spin_unlock(&rd->rto_lock); in rto_push_irq_work_func()
2452 sched_put_rd(rd); in rto_push_irq_work_func()
2457 irq_work_queue_on(&rd->rto_push_work, cpu); in rto_push_irq_work_func()
2480 cpumask_test_cpu(this_rq->cpu, this_rq->rd->rto_mask)) in pull_rt_task()
2490 for_each_cpu(cpu, this_rq->rd->rto_mask) { in pull_rt_task()
2597 cpupri_set(&rq->rd->cpupri, rq->cpu, rq->rt.highest_prio.curr); in rq_online_rt()
2608 cpupri_set(&rq->rd->cpupri, rq->cpu, CPUPRI_INVALID); in rq_offline_rt()