Home
last modified time | relevance | path

Searched refs:policy (Results 1 – 14 of 14) sorted by relevance

/kernel/sched/
Dcpufreq_schedutil.c25 struct cpufreq_policy *policy; member
90 if (!cpufreq_this_cpu_can_update(sg_policy->policy)) in sugov_should_update_freq()
157 struct cpufreq_policy *policy = sg_policy->policy; in get_next_freq() local
159 policy->cpuinfo.max_freq : policy->cur; in get_next_freq()
164 trace_android_vh_map_util_freq_new(util, freq, max, &next_freq, policy, in get_next_freq()
175 return cpufreq_driver_resolve_freq(policy, freq); in get_next_freq()
382 if (sg_policy->policy->fast_switch_enabled) { in sugov_update_single_freq()
383 cpufreq_driver_fast_switch(sg_policy->policy, next_f); in sugov_update_single_freq()
426 struct cpufreq_policy *policy = sg_policy->policy; in sugov_next_freq_shared() local
430 for_each_cpu(j, policy->cpus) { in sugov_next_freq_shared()
[all …]
Dcpufreq.c73 bool cpufreq_this_cpu_can_update(struct cpufreq_policy *policy) in cpufreq_this_cpu_can_update() argument
75 return cpumask_test_cpu(smp_processor_id(), policy->cpus) || in cpufreq_this_cpu_can_update()
76 (policy->dvfs_possible_from_any_cpu && in cpufreq_this_cpu_can_update()
Dcore.c2062 static inline int __normal_prio(int policy, int rt_prio, int nice) in __normal_prio() argument
2066 if (dl_policy(policy)) in __normal_prio()
2068 else if (rt_policy(policy)) in __normal_prio()
2085 return __normal_prio(p->policy, p->rt_priority, PRIO_TO_NICE(p->static_prio)); in normal_prio()
4513 p->policy = SCHED_NORMAL; in sched_fork()
7412 int policy = attr->sched_policy; in __setscheduler_params() local
7414 if (policy == SETPARAM_POLICY) in __setscheduler_params()
7415 policy = p->policy; in __setscheduler_params()
7417 p->policy = policy; in __setscheduler_params()
7419 if (dl_policy(policy)) in __setscheduler_params()
[all …]
Dsched.h164 static inline int idle_policy(int policy) in idle_policy() argument
166 return policy == SCHED_IDLE; in idle_policy()
168 static inline int fair_policy(int policy) in fair_policy() argument
170 return policy == SCHED_NORMAL || policy == SCHED_BATCH; in fair_policy()
173 static inline int rt_policy(int policy) in rt_policy() argument
175 return policy == SCHED_FIFO || policy == SCHED_RR; in rt_policy()
178 static inline int dl_policy(int policy) in dl_policy() argument
180 return policy == SCHED_DEADLINE; in dl_policy()
182 static inline bool valid_policy(int policy) in valid_policy() argument
184 return idle_policy(policy) || fair_policy(policy) || in valid_policy()
[all …]
Didle.c369 WARN_ON_ONCE(current->policy != SCHED_FIFO); in play_idle_precise()
Ddeadline.c2698 int sched_dl_overflow(struct task_struct *p, int policy, in sched_dl_overflow() argument
2703 u64 new_bw = dl_policy(policy) ? to_ratio(period, runtime) : 0; in sched_dl_overflow()
2724 if (dl_policy(policy) && !task_has_dl_policy(p) && in sched_dl_overflow()
2730 } else if (dl_policy(policy) && task_has_dl_policy(p) && in sched_dl_overflow()
2743 } else if (!dl_policy(policy) && task_has_dl_policy(p)) { in sched_dl_overflow()
Drt.c1248 return (tsk->policy == SCHED_RR) ? 1 : 0; in rt_se_rr_nr_running()
2600 if (p->policy != SCHED_RR) in task_tick_rt()
2626 if (task->policy == SCHED_RR) in get_rr_interval_rt()
Ddebug.c1031 P(policy); in proc_sched_show_task()
Dfair.c7480 if (unlikely(p->policy != SCHED_NORMAL) || !sched_feat(WAKEUP_PREEMPTION)) in check_preempt_wakeup()
7756 if (curr->policy != SCHED_BATCH) { in yield_task_fair()
/kernel/
Dtsacct.c52 stats->ac_sched = tsk->policy; in bacct_add_tsk()
Dtaskstats.c648 .policy = taskstats_cmd_get_policy,
656 .policy = cgroupstats_cmd_get_policy,
Dcpu.c1351 unsigned int policy = current->policy; in switch_to_rt_policy() local
1353 if (policy == SCHED_NORMAL) in switch_to_rt_policy()
/kernel/trace/
Dtrace.h170 unsigned long policy; member
Dtrace.c1789 max_data->policy = tsk->policy; in __update_max_tr()
4323 data->policy, data->rt_priority); in print_trace_header()