/kernel/time/ |
D | timekeeping_internal.h | 19 static inline u64 clocksource_delta(u64 now, u64 last, u64 mask) in clocksource_delta() argument 21 u64 ret = (now - last) & mask; in clocksource_delta() 30 static inline u64 clocksource_delta(u64 now, u64 last, u64 mask) in clocksource_delta() argument 32 return (now - last) & mask; in clocksource_delta()
|
D | Kconfig | 20 # Clocksources require validation of the clocksource against the last
|
D | timekeeping.c | 244 u64 now, last, mask, max, delta; in timekeeping_get_delta() local 257 last = tkr->cycle_last; in timekeeping_get_delta() 262 delta = clocksource_delta(now, last, mask); in timekeeping_get_delta()
|
/kernel/ |
D | user_namespace.c | 242 u32 first, last, id2; in cmp_map_id() local 254 last = first + el->count - 1; in cmp_map_id() 256 if (key->id >= first && key->id <= last && in cmp_map_id() 257 (id2 >= first && id2 <= last)) in cmp_map_id() 292 u32 first, last, id2; in map_id_range_down_base() local 299 last = first + map->extent[idx].count - 1; in map_id_range_down_base() 300 if (id >= first && id <= last && in map_id_range_down_base() 301 (id2 >= first && id2 <= last)) in map_id_range_down_base() 341 u32 first, last; in map_id_up_base() local 346 last = first + map->extent[idx].count - 1; in map_id_up_base() [all …]
|
D | cpu.c | 76 struct hlist_node *last; member 484 st->last = NULL; in cpuhp_set_state() 516 if (!st->last) { in cpuhp_reset_state() 815 st->result = cpuhp_invoke_callback(cpu, state, bringup, st->node, &st->last); in cpuhp_thread_fun() 823 st->result = cpuhp_invoke_callback(cpu, state, bringup, st->node, &st->last); in cpuhp_thread_fun() 869 st->last = NULL; in cpuhp_invoke_ap_callback() 881 if ((ret = st->result) && st->last) { in cpuhp_invoke_ap_callback() 892 st->node = st->last = NULL; in cpuhp_invoke_ap_callback()
|
D | ucount.c | 291 struct ucounts *last, enum ucount_type type) in do_dec_rlimit_put_ucounts() argument 294 for (iter = ucounts; iter != last; iter = next) { in do_dec_rlimit_put_ucounts()
|
/kernel/bpf/ |
D | bpf_lru_list.c | 175 struct list_head *cur, *last, *next = inactive; in __bpf_lru_list_rotate_inactive() local 182 last = l->next_inactive_rotation->next; in __bpf_lru_list_rotate_inactive() 183 if (last == inactive) in __bpf_lru_list_rotate_inactive() 184 last = last->next; in __bpf_lru_list_rotate_inactive() 197 if (cur == last) in __bpf_lru_list_rotate_inactive()
|
/kernel/sched/ |
D | topology.c | 977 struct sched_group *first = NULL, *last = NULL, *sg; in build_overlap_sched_groups() local 1053 if (last) in build_overlap_sched_groups() 1054 last->next = sg; in build_overlap_sched_groups() 1055 last = sg; in build_overlap_sched_groups() 1056 last->next = first; in build_overlap_sched_groups() 1187 struct sched_group *first = NULL, *last = NULL; in build_sched_groups() local 1210 if (last) in build_sched_groups() 1211 last->next = sg; in build_sched_groups() 1212 last = sg; in build_sched_groups() 1214 last->next = first; in build_sched_groups()
|
D | debug.c | 586 struct sched_entity *last; in print_cfs_rq() local 602 last = __pick_last_entity(cfs_rq); in print_cfs_rq() 603 if (last) in print_cfs_rq() 604 max_vruntime = last->vruntime; in print_cfs_rq()
|
D | fair.c | 619 struct rb_node *last = rb_last(&cfs_rq->tasks_timeline.rb_root); in __pick_last_entity() local 621 if (!last) in __pick_last_entity() 624 return __node_2_se(last); in __pick_last_entity() 4563 if (cfs_rq->last != se) in __clear_buddies_last() 4566 cfs_rq->last = NULL; in __clear_buddies_last() 4594 if (cfs_rq->last == se) in clear_buddies() 4800 } else if (cfs_rq->last && wakeup_preempt_entity(cfs_rq->last, left) < 1) { in pick_next_entity() 4804 se = cfs_rq->last; in pick_next_entity() 7454 cfs_rq_of(se)->last = se; in set_last_buddy() 8067 &p->se == cfs_rq_of(&p->se)->last)) in task_hot()
|
D | sched.h | 564 struct sched_entity *last; member
|
/kernel/locking/ |
D | test-ww_mutex.c | 278 unsigned int n, last = nthreads - 1; in __test_cycle() local 289 if (n == last) in __test_cycle() 295 cycle->a_signal = &cycles[last].b_signal; in __test_cycle()
|
/kernel/power/ |
D | swap.c | 975 struct swap_map_page_list *tmp, *last; in get_swap_reader() local 984 last = handle->maps = NULL; in get_swap_reader() 994 if (last) in get_swap_reader() 995 last->next = tmp; in get_swap_reader() 996 last = tmp; in get_swap_reader()
|
D | Kconfig | 227 This enables code to save the last PM event point across 245 This enables some cheesy code to save the last PM event point in the
|
/kernel/events/ |
D | ring_buffer.c | 713 int last, order; in rb_alloc_aux() local 720 for (last = rb->aux_nr_pages + (1 << page_private(page)); in rb_alloc_aux() 721 last > rb->aux_nr_pages; rb->aux_nr_pages++) in rb_alloc_aux()
|
/kernel/trace/ |
D | Kconfig | 275 and last enabled. 789 last=632 first=3672 max=632 min=632 avg=316 std=446 std^2=199712 790 last=278 first=3672 max=632 min=278 avg=303 std=316 std^2=100337 791 last=277 first=3672 max=632 min=277 avg=296 std=258 std^2=67064 792 last=273 first=3672 max=632 min=273 avg=292 std=224 std^2=50411 793 last=273 first=3672 max=632 min=273 avg=288 std=200 std^2=40389 794 last=281 first=3672 max=632 min=273 avg=287 std=183 std^2=33666
|
D | trace.c | 2014 struct tracer *t, **last; in init_trace_selftests() local 2037 last = &trace_types; in init_trace_selftests() 2040 *last = t->next; in init_trace_selftests() 2043 last = &t->next; in init_trace_selftests() 9796 union trace_eval_map_item **last = &trace_eval_maps; in trace_module_remove_evals() local 9809 last = &map->tail.next; in trace_module_remove_evals() 9815 *last = trace_eval_jmp_to_tail(map)->tail.next; in trace_module_remove_evals()
|
/kernel/cgroup/ |
D | cgroup.c | 4359 unsigned long last = cfile->notified_at; in cgroup_file_notify() local 4360 unsigned long next = last + CGROUP_FILE_NOTIFY_MIN_INTV; in cgroup_file_notify() 4362 if (time_in_range(jiffies, last, next)) { in cgroup_file_notify() 4503 struct cgroup_subsys_state *last, *tmp; in css_rightmost_descendant() local 4508 last = pos; in css_rightmost_descendant() 4511 css_for_each_child(tmp, last) in css_rightmost_descendant() 4515 return last; in css_rightmost_descendant() 4521 struct cgroup_subsys_state *last; in css_leftmost_descendant() local 4524 last = pos; in css_leftmost_descendant() 4528 return last; in css_leftmost_descendant()
|
/kernel/rcu/ |
D | Kconfig | 174 bool "Accelerate last non-dyntick-idle CPU's grace periods"
|