/kernel/time/ |
D | clockevents.c | 471 static void clockevents_config(struct clock_event_device *dev, u32 freq) in clockevents_config() argument 484 do_div(sec, freq); in clockevents_config() 490 clockevents_calc_mult_shift(dev, freq, sec); in clockevents_config() 505 u32 freq, unsigned long min_delta, in clockevents_config_and_register() argument 510 clockevents_config(dev, freq); in clockevents_config_and_register() 515 int __clockevents_update_freq(struct clock_event_device *dev, u32 freq) in __clockevents_update_freq() argument 517 clockevents_config(dev, freq); in __clockevents_update_freq() 540 int clockevents_update_freq(struct clock_event_device *dev, u32 freq) in clockevents_update_freq() argument 546 ret = tick_broadcast_update_freq(dev, freq); in clockevents_update_freq() 548 ret = __clockevents_update_freq(dev, freq); in clockevents_update_freq()
|
D | tick-internal.h | 58 extern int __clockevents_update_freq(struct clock_event_device *dev, u32 freq); 71 extern int tick_broadcast_update_freq(struct clock_event_device *dev, u32 freq); 84 static inline int tick_broadcast_update_freq(struct clock_event_device *dev, u32 freq) { return -EN… in tick_broadcast_update_freq() argument
|
D | clocksource.c | 1111 void __clocksource_update_freq_scale(struct clocksource *cs, u32 scale, u32 freq) in __clocksource_update_freq_scale() argument 1119 if (freq) { in __clocksource_update_freq_scale() 1130 do_div(sec, freq); in __clocksource_update_freq_scale() 1137 clocks_calc_mult_shift(&cs->mult, &cs->shift, freq, in __clocksource_update_freq_scale() 1151 if (scale && freq && !cs->uncertainty_margin) { in __clocksource_update_freq_scale() 1152 cs->uncertainty_margin = NSEC_PER_SEC / (scale * freq); in __clocksource_update_freq_scale() 1165 while (freq && ((cs->mult + cs->maxadj < cs->mult) in __clocksource_update_freq_scale() 1198 int __clocksource_register_scale(struct clocksource *cs, u32 scale, u32 freq) in __clocksource_register_scale() argument 1214 __clocksource_update_freq_scale(cs, scale, freq); in __clocksource_register_scale()
|
D | ntp.c | 169 static inline void pps_set_freq(s64 freq) in pps_set_freq() argument 171 pps_freq = freq; in pps_set_freq() 218 static inline void pps_set_freq(s64 freq) {} in pps_set_freq() argument 722 time_freq = txc->freq * PPM_SCALE; in process_adjtimex_modes() 808 txc->freq = shift_right((time_freq >> PPM_SCALE_INV_SHIFT) * in __do_adjtimex()
|
D | time.c | 296 txc->freq = tx32.freq; in get_old_timex32() 325 tx32.freq = txc->freq; in put_old_timex32()
|
D | tick-broadcast.c | 214 int tick_broadcast_update_freq(struct clock_event_device *dev, u32 freq) in tick_broadcast_update_freq() argument 220 ret = __clockevents_update_freq(dev, freq); in tick_broadcast_update_freq()
|
D | timekeeping.c | 2404 if (LLONG_MIN / PPM_SCALE > txc->freq) in timekeeping_validate_timex() 2406 if (LLONG_MAX / PPM_SCALE < txc->freq) in timekeeping_validate_timex()
|
/kernel/power/ |
D | energy_model.c | 110 unsigned long power, freq, prev_freq = 0, prev_cost = ULONG_MAX; in em_create_perf_table() local 120 for (i = 0, freq = 0; i < nr_states; i++, freq++) { in em_create_perf_table() 126 ret = cb->active_power(dev, &power, &freq); in em_create_perf_table() 137 if (freq <= prev_freq) { in em_create_perf_table() 139 freq); in em_create_perf_table() 154 table[i].frequency = prev_freq = freq; in em_create_perf_table()
|
/kernel/sched/ |
D | cpufreq_schedutil.c | 146 unsigned int freq = arch_scale_freq_invariant() ? in get_next_freq() local 151 trace_android_vh_map_util_freq(util, freq, max, &next_freq, policy, in get_next_freq() 154 freq = next_freq; in get_next_freq() 156 freq = map_util_freq(util, freq, max); in get_next_freq() 158 if (freq == sg_policy->cached_raw_freq && !sg_policy->need_freq_update) in get_next_freq() 161 sg_policy->cached_raw_freq = freq; in get_next_freq() 162 return cpufreq_driver_resolve_freq(policy, freq); in get_next_freq() 473 unsigned int freq; in sugov_work() local 487 freq = sg_policy->next_freq; in sugov_work() 492 __cpufreq_driver_target(sg_policy->policy, freq, CPUFREQ_RELATION_L); in sugov_work()
|
D | debug.c | 724 unsigned int freq = cpu_khz ? : 1; in print_cpu() local 727 cpu, freq / 1000, (freq % 1000)); in print_cpu()
|
/kernel/events/ |
D | core.c | 2339 if (event->attr.freq && event->attr.sample_freq) in event_sched_out() 2589 if (event->attr.freq && event->attr.sample_freq) in event_sched_in() 4134 if (!event->attr.freq || !event->attr.sample_freq) in perf_adjust_freq_unthr_context() 4887 if (event->attr.freq) in unaccount_event() 5601 if (event->attr.freq) { in __perf_event_period() 5643 if (event->attr.freq && value > sysctl_perf_event_sample_rate) in _perf_event_period() 5649 if (!event->attr.freq && (value & (1ULL << 63))) in _perf_event_period() 9309 if (event->attr.freq) { in __perf_event_account_interrupt() 9511 if ((event->attr.sample_type & PERF_SAMPLE_PERIOD) && !event->attr.freq) { in perf_swevent_event() 9517 if (nr == 1 && hwc->sample_period == 1 && !event->attr.freq) in perf_swevent_event() [all …]
|