• Home
  • Raw
  • Download

Lines Matching refs:rd

107 	return &cpu_rq(i)->rd->dl_bw;  in dl_bw_of()
112 struct root_domain *rd = cpu_rq(i)->rd; in dl_bw_cpus() local
118 if (cpumask_subset(rd->span, cpu_active_mask)) in dl_bw_cpus()
119 return cpumask_weight(rd->span); in dl_bw_cpus()
123 for_each_cpu_and(i, rd->span, cpu_active_mask) in dl_bw_cpus()
153 return __dl_bw_capacity(cpu_rq(i)->rd->span); in dl_bw_capacity()
159 struct root_domain *rd = cpu_rq(cpu)->rd; in dl_bw_visited() local
161 if (rd->visit_gen == gen) in dl_bw_visited()
164 rd->visit_gen = gen; in dl_bw_visited()
171 struct root_domain *rd = container_of(dl_b, struct root_domain, dl_bw); in __dl_update() local
176 for_each_cpu_and(i, rd->span, cpu_active_mask) { in __dl_update()
535 return atomic_read(&rq->rd->dlo_count); in dl_overloaded()
543 cpumask_set_cpu(rq->cpu, rq->rd->dlo_mask); in dl_set_overload()
551 atomic_inc(&rq->rd->dlo_count); in dl_set_overload()
559 atomic_dec(&rq->rd->dlo_count); in dl_clear_overload()
560 cpumask_clear_cpu(rq->cpu, rq->rd->dlo_mask); in dl_clear_overload()
724 dl_b = &rq->rd->dl_bw; in dl_task_offline_migration()
726 __dl_sub(dl_b, p->dl.dl_bw, cpumask_weight(rq->rd->span)); in dl_task_offline_migration()
729 dl_b = &later_rq->rd->dl_bw; in dl_task_offline_migration()
731 __dl_add(dl_b, p->dl.dl_bw, cpumask_weight(later_rq->rd->span)); in dl_task_offline_migration()
1472 cpupri_set(&rq->rd->cpupri, rq->cpu, CPUPRI_HIGHER); in inc_dl_deadline()
1474 cpudl_set(&rq->rd->cpudl, rq->cpu, deadline); in inc_dl_deadline()
1489 cpudl_clear(&rq->rd->cpudl, rq->cpu); in dec_dl_deadline()
1490 cpupri_set(&rq->rd->cpupri, rq->cpu, rq->rt.highest_prio.curr); in dec_dl_deadline()
1496 cpudl_set(&rq->rd->cpudl, rq->cpu, entry->deadline); in dec_dl_deadline()
1917 !cpudl_find(&rq->rd->cpudl, rq->curr, NULL)) in check_preempt_equal_dl()
1925 cpudl_find(&rq->rd->cpudl, p, NULL)) in check_preempt_equal_dl()
2151 if (!cpudl_find(&task_rq(task)->rd->cpudl, task, later_mask)) in find_later_rq()
2401 for_each_cpu(cpu, this_rq->rd->dlo_mask) { in pull_dl_task()
2502 src_rd = rq->rd; in set_cpus_allowed_dl()
2532 cpudl_set_freecpu(&rq->rd->cpudl, rq->cpu); in rq_online_dl()
2534 cpudl_set(&rq->rd->cpudl, rq->cpu, rq->dl.earliest_dl.curr); in rq_online_dl()
2543 cpudl_clear(&rq->rd->cpudl, rq->cpu); in rq_offline_dl()
2544 cpudl_clear_freecpu(&rq->rd->cpudl, rq->cpu); in rq_offline_dl()
2570 dl_b = &rq->rd->dl_bw; in dl_add_task_root_domain()
2573 __dl_add(dl_b, p->dl.dl_bw, cpumask_weight(rq->rd->span)); in dl_add_task_root_domain()
2580 void dl_clear_root_domain(struct root_domain *rd) in dl_clear_root_domain() argument
2584 raw_spin_lock_irqsave(&rd->dl_bw.lock, flags); in dl_clear_root_domain()
2585 rd->dl_bw.total_bw = 0; in dl_clear_root_domain()
2586 raw_spin_unlock_irqrestore(&rd->dl_bw.lock, flags); in dl_clear_root_domain()