Searched refs:policy (Results 1 – 14 of 14) sorted by relevance
/kernel/sched/ |
D | cpufreq_schedutil.c | 25 struct cpufreq_policy *policy; member 90 if (!cpufreq_this_cpu_can_update(sg_policy->policy)) in sugov_should_update_freq() 157 struct cpufreq_policy *policy = sg_policy->policy; in get_next_freq() local 159 policy->cpuinfo.max_freq : policy->cur; in get_next_freq() 164 trace_android_vh_map_util_freq_new(util, freq, max, &next_freq, policy, in get_next_freq() 175 return cpufreq_driver_resolve_freq(policy, freq); in get_next_freq() 382 if (sg_policy->policy->fast_switch_enabled) { in sugov_update_single_freq() 383 cpufreq_driver_fast_switch(sg_policy->policy, next_f); in sugov_update_single_freq() 426 struct cpufreq_policy *policy = sg_policy->policy; in sugov_next_freq_shared() local 430 for_each_cpu(j, policy->cpus) { in sugov_next_freq_shared() [all …]
|
D | cpufreq.c | 73 bool cpufreq_this_cpu_can_update(struct cpufreq_policy *policy) in cpufreq_this_cpu_can_update() argument 75 return cpumask_test_cpu(smp_processor_id(), policy->cpus) || in cpufreq_this_cpu_can_update() 76 (policy->dvfs_possible_from_any_cpu && in cpufreq_this_cpu_can_update()
|
D | core.c | 2062 static inline int __normal_prio(int policy, int rt_prio, int nice) in __normal_prio() argument 2066 if (dl_policy(policy)) in __normal_prio() 2068 else if (rt_policy(policy)) in __normal_prio() 2085 return __normal_prio(p->policy, p->rt_priority, PRIO_TO_NICE(p->static_prio)); in normal_prio() 4513 p->policy = SCHED_NORMAL; in sched_fork() 7412 int policy = attr->sched_policy; in __setscheduler_params() local 7414 if (policy == SETPARAM_POLICY) in __setscheduler_params() 7415 policy = p->policy; in __setscheduler_params() 7417 p->policy = policy; in __setscheduler_params() 7419 if (dl_policy(policy)) in __setscheduler_params() [all …]
|
D | sched.h | 164 static inline int idle_policy(int policy) in idle_policy() argument 166 return policy == SCHED_IDLE; in idle_policy() 168 static inline int fair_policy(int policy) in fair_policy() argument 170 return policy == SCHED_NORMAL || policy == SCHED_BATCH; in fair_policy() 173 static inline int rt_policy(int policy) in rt_policy() argument 175 return policy == SCHED_FIFO || policy == SCHED_RR; in rt_policy() 178 static inline int dl_policy(int policy) in dl_policy() argument 180 return policy == SCHED_DEADLINE; in dl_policy() 182 static inline bool valid_policy(int policy) in valid_policy() argument 184 return idle_policy(policy) || fair_policy(policy) || in valid_policy() [all …]
|
D | idle.c | 369 WARN_ON_ONCE(current->policy != SCHED_FIFO); in play_idle_precise()
|
D | deadline.c | 2698 int sched_dl_overflow(struct task_struct *p, int policy, in sched_dl_overflow() argument 2703 u64 new_bw = dl_policy(policy) ? to_ratio(period, runtime) : 0; in sched_dl_overflow() 2724 if (dl_policy(policy) && !task_has_dl_policy(p) && in sched_dl_overflow() 2730 } else if (dl_policy(policy) && task_has_dl_policy(p) && in sched_dl_overflow() 2743 } else if (!dl_policy(policy) && task_has_dl_policy(p)) { in sched_dl_overflow()
|
D | rt.c | 1248 return (tsk->policy == SCHED_RR) ? 1 : 0; in rt_se_rr_nr_running() 2600 if (p->policy != SCHED_RR) in task_tick_rt() 2626 if (task->policy == SCHED_RR) in get_rr_interval_rt()
|
D | debug.c | 1031 P(policy); in proc_sched_show_task()
|
D | fair.c | 7480 if (unlikely(p->policy != SCHED_NORMAL) || !sched_feat(WAKEUP_PREEMPTION)) in check_preempt_wakeup() 7756 if (curr->policy != SCHED_BATCH) { in yield_task_fair()
|
/kernel/ |
D | tsacct.c | 52 stats->ac_sched = tsk->policy; in bacct_add_tsk()
|
D | taskstats.c | 648 .policy = taskstats_cmd_get_policy, 656 .policy = cgroupstats_cmd_get_policy,
|
D | cpu.c | 1351 unsigned int policy = current->policy; in switch_to_rt_policy() local 1353 if (policy == SCHED_NORMAL) in switch_to_rt_policy()
|
/kernel/trace/ |
D | trace.h | 170 unsigned long policy; member
|
D | trace.c | 1789 max_data->policy = tsk->policy; in __update_max_tr() 4323 data->policy, data->rt_priority); in print_trace_header()
|