Home
last modified time | relevance | path

Searched refs:it (Results 1 – 22 of 22) sorted by relevance

/kernel/time/
Dposix-cpu-timers.c108 return pid_task(timer->it.cpu.pid, clock_pid_type(timer->it_clock)); in cpu_timer_task_rcu()
117 u64 delta, incr, expires = timer->it.cpu.node.expires; in bump_cpu_timer()
137 timer->it.cpu.node.expires += incr; in bump_cpu_timer()
141 return timer->it.cpu.node.expires; in bump_cpu_timer()
402 timerqueue_init(&new_timer->it.cpu.node); in posix_cpu_timer_create()
403 new_timer->it.cpu.pid = get_pid(pid); in posix_cpu_timer_create()
416 struct cpu_timer *ctmr = &timer->it.cpu; in posix_cpu_timer_del()
439 if (timer->it.cpu.firing) in posix_cpu_timer_del()
502 struct cpu_timer *ctmr = &timer->it.cpu; in arm_timer()
534 struct cpu_timer *ctmr = &timer->it.cpu; in cpu_timer_fire()
[all …]
Ditimer.c51 struct cpu_itimer *it = &tsk->signal->it[clock_id]; in get_cpu_itimer() local
55 val = it->expires; in get_cpu_itimer()
56 interval = it->incr; in get_cpu_itimer()
173 struct cpu_itimer *it = &tsk->signal->it[clock_id]; in set_cpu_itimer() local
180 oval = it->expires; in set_cpu_itimer()
181 ointerval = it->incr; in set_cpu_itimer()
187 it->expires = nval; in set_cpu_itimer()
188 it->incr = ninterval; in set_cpu_itimer()
Dalarmtimer.c568 it.alarm.alarmtimer); in alarm_handle_timer()
601 struct alarm *alarm = &timr->it.alarm.alarmtimer; in alarm_timer_rearm()
614 struct alarm *alarm = &timr->it.alarm.alarmtimer; in alarm_timer_forward()
626 struct alarm *alarm = &timr->it.alarm.alarmtimer; in alarm_timer_remaining()
637 return alarm_try_to_cancel(&timr->it.alarm.alarmtimer); in alarm_timer_try_to_cancel()
650 hrtimer_cancel_wait_running(&timr->it.alarm.alarmtimer.timer); in alarm_timer_wait_running()
663 struct alarm *alarm = &timr->it.alarm.alarmtimer; in alarm_timer_arm()
671 alarm_start(&timr->it.alarm.alarmtimer, expires); in alarm_timer_arm()
743 alarm_init(&new_timer->it.alarm.alarmtimer, type, alarm_handle_timer); in alarm_timer_create()
Dposix-timers.c295 struct hrtimer *timer = &timr->it.real.timer; in common_hrtimer_rearm()
373 timr = container_of(timer, struct k_itimer, it.real.timer); in posix_timer_fn()
493 hrtimer_init(&new_timer->it.real.timer, new_timer->it_clock, 0); in common_timer_create()
642 struct hrtimer *timer = &timr->it.real.timer; in common_hrtimer_remaining()
649 struct hrtimer *timer = &timr->it.real.timer; in common_hrtimer_forward()
796 struct hrtimer *timer = &timr->it.real.timer; in common_hrtimer_arm()
812 hrtimer_init(&timr->it.real.timer, timr->it_clock, mode); in common_hrtimer_arm()
813 timr->it.real.timer.function = posix_timer_fn; in common_hrtimer_arm()
825 return hrtimer_try_to_cancel(&timr->it.real.timer); in common_hrtimer_try_to_cancel()
830 hrtimer_cancel_wait_running(&timer->it.real.timer); in common_timer_wait_running()
Dtime.c861 int get_itimerspec64(struct itimerspec64 *it, in get_itimerspec64() argument
866 ret = get_timespec64(&it->it_interval, &uit->it_interval); in get_itimerspec64()
870 ret = get_timespec64(&it->it_value, &uit->it_value); in get_itimerspec64()
876 int put_itimerspec64(const struct itimerspec64 *it, in put_itimerspec64() argument
881 ret = put_timespec64(&it->it_interval, &uit->it_interval); in put_itimerspec64()
885 ret = put_timespec64(&it->it_value, &uit->it_value); in put_itimerspec64()
DKconfig86 rate, even when the CPU doesn't need it.
123 transitions: syscalls, exceptions and interrupts. Even when it's
147 for context tracking and the subsystems that rely on it: RCU
164 We keep it around for a little while to enforce backward
/kernel/cgroup/
Dlegacy_freezer.c248 struct css_task_iter it; in update_if_frozen() local
271 css_task_iter_start(css, 0, &it); in update_if_frozen()
273 while ((task = css_task_iter_next(&it))) { in update_if_frozen()
288 css_task_iter_end(&it); in update_if_frozen()
320 struct css_task_iter it; in freeze_cgroup() local
323 css_task_iter_start(&freezer->css, 0, &it); in freeze_cgroup()
324 while ((task = css_task_iter_next(&it))) in freeze_cgroup()
326 css_task_iter_end(&it); in freeze_cgroup()
331 struct css_task_iter it; in unfreeze_cgroup() local
334 css_task_iter_start(&freezer->css, 0, &it); in unfreeze_cgroup()
[all …]
Dcgroup.c234 static void css_task_iter_skip(struct css_task_iter *it,
873 struct css_task_iter *it, *pos; in css_set_skip_task_iters() local
875 list_for_each_entry_safe(it, pos, &cset->task_iters, iters_node) in css_set_skip_task_iters()
876 css_task_iter_skip(it, task); in css_set_skip_task_iters()
1785 struct css_task_iter *it; in rebind_subsystems() local
1815 list_for_each_entry(it, &cset->task_iters, iters_node) in rebind_subsystems()
1816 if (it->cset_head == &scgrp->e_csets[ss->id]) in rebind_subsystems()
1817 it->cset_head = &dcgrp->e_csets[ss->id]; in rebind_subsystems()
4512 static struct css_set *css_task_iter_next_css_set(struct css_task_iter *it) in css_task_iter_next_css_set() argument
4521 if (it->tcset_pos) { in css_task_iter_next_css_set()
[all …]
Dcgroup-v1.c100 struct css_task_iter it; in cgroup_transfer_tasks() local
130 css_task_iter_start(&from->self, 0, &it); in cgroup_transfer_tasks()
133 task = css_task_iter_next(&it); in cgroup_transfer_tasks()
138 css_task_iter_end(&it); in cgroup_transfer_tasks()
334 struct css_task_iter it; in pidlist_array_load() local
351 css_task_iter_start(&cgrp->self, 0, &it); in pidlist_array_load()
352 while ((tsk = css_task_iter_next(&it))) { in pidlist_array_load()
363 css_task_iter_end(&it); in pidlist_array_load()
707 struct css_task_iter it; in cgroupstats_build() local
731 css_task_iter_start(&cgrp->self, 0, &it); in cgroupstats_build()
[all …]
Dfreezer.c179 struct css_task_iter it; in cgroup_do_freeze() local
196 css_task_iter_start(&cgrp->self, 0, &it); in cgroup_do_freeze()
197 while ((task = css_task_iter_next(&it))) { in cgroup_do_freeze()
206 css_task_iter_end(&it); in cgroup_do_freeze()
Dcpuset.c963 struct css_task_iter it; in dl_update_tasks_root_domain() local
969 css_task_iter_start(&cs->css, 0, &it); in dl_update_tasks_root_domain()
971 while ((task = css_task_iter_next(&it))) in dl_update_tasks_root_domain()
974 css_task_iter_end(&it); in dl_update_tasks_root_domain()
1121 struct css_task_iter it; in update_tasks_cpumask() local
1125 css_task_iter_start(&cs->css, 0, &it); in update_tasks_cpumask()
1126 while ((task = css_task_iter_next(&it))) { in update_tasks_cpumask()
1135 css_task_iter_end(&it); in update_tasks_cpumask()
1759 struct css_task_iter it; in update_tasks_nodemask() local
1776 css_task_iter_start(&cs->css, 0, &it); in update_tasks_nodemask()
[all …]
/kernel/sched/
Didle.c349 struct idle_timer *it = container_of(timer, struct idle_timer, timer); in idle_inject_timer_fn() local
351 WRITE_ONCE(it->done, 1); in idle_inject_timer_fn()
359 struct idle_timer it; in play_idle_precise() local
376 it.done = 0; in play_idle_precise()
377 hrtimer_init_on_stack(&it.timer, CLOCK_MONOTONIC, HRTIMER_MODE_REL_HARD); in play_idle_precise()
378 it.timer.function = idle_inject_timer_fn; in play_idle_precise()
379 hrtimer_start(&it.timer, ns_to_ktime(duration_ns), in play_idle_precise()
382 while (!READ_ONCE(it.done)) in play_idle_precise()
Drt.c2594 struct css_task_iter it; in tg_has_rt_tasks() local
2603 css_task_iter_start(&tg->css, 0, &it); in tg_has_rt_tasks()
2604 while (!ret && (task = css_task_iter_next(&it))) in tg_has_rt_tasks()
2606 css_task_iter_end(&it); in tg_has_rt_tasks()
Dcore.c1379 struct css_task_iter it; in uclamp_update_active_tasks() local
1382 css_task_iter_start(css, 0, &it); in uclamp_update_active_tasks()
1383 while ((p = css_task_iter_next(&it))) in uclamp_update_active_tasks()
1385 css_task_iter_end(&it); in uclamp_update_active_tasks()
/kernel/trace/
DKconfig114 # options do not appear when something else selects it. We need the two options
175 tracing is enabled by the administrator. If it's runtime disabled
248 kernel and displays it in /sys/kernel/debug/tracing/stack_trace.
253 then it will not have any overhead while the stack tracer
328 to be scheduled in, starting from the point it has woken up.
339 time, this tracer will detect it. This is useful for testing
356 When the tracer is not running, it has no affect on the system,
357 but when it is running, it can cause the system to be
362 file. Every time a latency is greater than tracing_thresh, it will
420 trace recording, as it needs to add some checks to synchronize
[all …]
/kernel/rcu/
DKconfig.debug16 Enable RCU lockdep checking for list usages. By default it is
19 false-positive splats, we keep it default disabled but once all
80 Say M if you want to build it as a module instead.
124 grace periods, making them as short as it can. This limits
128 But in conjunction with tools like KASAN, it can be helpful
/kernel/power/
DKconfig48 system and powers it off; and restores that checkpoint on reboot.
57 In principle it does not require ACPI or APM, although for example
58 ACPI will be used for the final steps when it is available. One
65 have it detect the saved image, restore memory state from it, and
136 non-zero numbered CPU, it may define ARCH_SUSPEND_NONZERO_CPU. This
200 make it wake up a few seconds later using an RTC wakeup alarm.
204 linked, ensuring that it's available when this test runs.
252 machine, reboot it and then run
276 manpage ("man 8 hdparm") for that), and it doesn't turn off
/kernel/dma/
DKconfig146 default, but it can be enabled by passing cma=size[MG] to the kernel.
157 If 0 percent is selected, CMA is disabled by default, but it can be
186 for larger buffers it just a memory waste. With this parameter you can
/kernel/
DKconfig.preempt31 low priority process to voluntarily preempt itself even if it
48 even if it is in kernel mode executing a system call and would
/kernel/debug/kdb/
Dkdb_cmds5 # be safe. Global breakpoint commands affect each cpu as it is booted.
/kernel/bpf/preload/
DKconfig10 # The dependency on !COMPILE_TEST prevents it from being enabled
/kernel/bpf/
Dcore.c758 unsigned int it = 0; in bpf_get_kallsym() local
766 if (it++ != symnum) in bpf_get_kallsym()