/kernel/time/ |
D | timekeeping_internal.h | 17 static inline u64 clocksource_delta(u64 now, u64 last, u64 mask) in clocksource_delta() argument 19 u64 ret = (now - last) & mask; in clocksource_delta() 28 static inline u64 clocksource_delta(u64 now, u64 last, u64 mask) in clocksource_delta() argument 30 return (now - last) & mask; in clocksource_delta()
|
D | Kconfig | 20 # Clocksources require validation of the clocksource against the last
|
D | timekeeping.c | 221 u64 now, last, mask, max, delta; in timekeeping_get_delta() local 234 last = tkr->cycle_last; in timekeeping_get_delta() 239 delta = clocksource_delta(now, last, mask); in timekeeping_get_delta()
|
/kernel/ |
D | user_namespace.c | 225 u32 first, last, id2; in cmp_map_id() local 237 last = first + el->count - 1; in cmp_map_id() 239 if (key->id >= first && key->id <= last && in cmp_map_id() 240 (id2 >= first && id2 <= last)) in cmp_map_id() 275 u32 first, last, id2; in map_id_range_down_base() local 282 last = first + map->extent[idx].count - 1; in map_id_range_down_base() 283 if (id >= first && id <= last && in map_id_range_down_base() 284 (id2 >= first && id2 <= last)) in map_id_range_down_base() 324 u32 first, last; in map_id_up_base() local 329 last = first + map->extent[idx].count - 1; in map_id_up_base() [all …]
|
D | cpu.c | 66 struct hlist_node *last; member 458 st->last = NULL; in cpuhp_set_state() 476 if (!st->last) { in cpuhp_reset_state() 688 st->result = cpuhp_invoke_callback(cpu, state, bringup, st->node, &st->last); in cpuhp_thread_fun() 696 st->result = cpuhp_invoke_callback(cpu, state, bringup, st->node, &st->last); in cpuhp_thread_fun() 741 st->last = NULL; in cpuhp_invoke_ap_callback() 753 if ((ret = st->result) && st->last) { in cpuhp_invoke_ap_callback() 764 st->node = st->last = NULL; in cpuhp_invoke_ap_callback()
|
/kernel/bpf/ |
D | bpf_lru_list.c | 170 struct list_head *cur, *last, *next = inactive; in __bpf_lru_list_rotate_inactive() local 177 last = l->next_inactive_rotation->next; in __bpf_lru_list_rotate_inactive() 178 if (last == inactive) in __bpf_lru_list_rotate_inactive() 179 last = last->next; in __bpf_lru_list_rotate_inactive() 192 if (cur == last) in __bpf_lru_list_rotate_inactive()
|
/kernel/sched/ |
D | topology.c | 932 struct sched_group *first = NULL, *last = NULL, *sg; in build_overlap_sched_groups() local 973 if (last) in build_overlap_sched_groups() 974 last->next = sg; in build_overlap_sched_groups() 975 last = sg; in build_overlap_sched_groups() 976 last->next = first; in build_overlap_sched_groups() 1107 struct sched_group *first = NULL, *last = NULL; in build_sched_groups() local 1130 if (last) in build_sched_groups() 1131 last->next = sg; in build_sched_groups() 1132 last = sg; in build_sched_groups() 1134 last->next = first; in build_sched_groups()
|
D | debug.c | 488 struct sched_entity *last; in print_cfs_rq() local 504 last = __pick_last_entity(cfs_rq); in print_cfs_rq() 505 if (last) in print_cfs_rq() 506 max_vruntime = last->vruntime; in print_cfs_rq()
|
D | fair.c | 623 struct rb_node *last = rb_last(&cfs_rq->tasks_timeline.rb_root); in __pick_last_entity() local 625 if (!last) in __pick_last_entity() 628 return rb_entry(last, struct sched_entity, run_node); in __pick_last_entity() 4032 if (cfs_rq->last != se) in __clear_buddies_last() 4035 cfs_rq->last = NULL; in __clear_buddies_last() 4063 if (cfs_rq->last == se) in clear_buddies() 4247 if (cfs_rq->last && wakeup_preempt_entity(cfs_rq->last, left) < 1) in pick_next_entity() 4248 se = cfs_rq->last; in pick_next_entity() 6707 cfs_rq_of(se)->last = se; in set_last_buddy() 7210 &p->se == cfs_rq_of(&p->se)->last)) in task_hot()
|
D | sched.h | 513 struct sched_entity *last; member
|
/kernel/trace/ |
D | trace_output.c | 666 int last = __TRACE_LAST_TYPE; in trace_search_list() local 670 return last + 1; in trace_search_list() 678 if (e->type != last + 1) in trace_search_list() 680 last++; in trace_search_list() 684 if ((last + 1) > TRACE_EVENT_TYPE_MAX) in trace_search_list() 688 return last + 1; in trace_search_list()
|
D | Kconfig | 176 and last enabled. 698 last=632 first=3672 max=632 min=632 avg=316 std=446 std^2=199712 699 last=278 first=3672 max=632 min=278 avg=303 std=316 std^2=100337 700 last=277 first=3672 max=632 min=277 avg=296 std=258 std^2=67064 701 last=273 first=3672 max=632 min=273 avg=292 std=224 std^2=50411 702 last=273 first=3672 max=632 min=273 avg=288 std=200 std^2=40389 703 last=281 first=3672 max=632 min=273 avg=287 std=183 std^2=33666
|
D | trace.c | 1734 struct tracer *t, **last; in init_trace_selftests() local 1756 last = &trace_types; in init_trace_selftests() 1759 *last = t->next; in init_trace_selftests() 1762 last = &t->next; in init_trace_selftests() 8713 union trace_eval_map_item **last = &trace_eval_maps; in trace_module_remove_evals() local 8726 last = &map->tail.next; in trace_module_remove_evals() 8732 *last = trace_eval_jmp_to_tail(map)->tail.next; in trace_module_remove_evals()
|
/kernel/locking/ |
D | test-ww_mutex.c | 278 unsigned int n, last = nthreads - 1; in __test_cycle() local 289 if (n == last) in __test_cycle() 295 cycle->a_signal = &cycles[last].b_signal; in __test_cycle()
|
/kernel/events/ |
D | ring_buffer.c | 664 int last, order; in rb_alloc_aux() local 671 for (last = rb->aux_nr_pages + (1 << page_private(page)); in rb_alloc_aux() 672 last > rb->aux_nr_pages; rb->aux_nr_pages++) in rb_alloc_aux()
|
/kernel/power/ |
D | swap.c | 965 struct swap_map_page_list *tmp, *last; in get_swap_reader() local 974 last = handle->maps = NULL; in get_swap_reader() 984 if (last) in get_swap_reader() 985 last->next = tmp; in get_swap_reader() 986 last = tmp; in get_swap_reader()
|
D | Kconfig | 217 This enables code to save the last PM event point across 235 This enables some cheesy code to save the last PM event point in the
|
/kernel/cgroup/ |
D | cgroup.c | 4149 unsigned long last = cfile->notified_at; in cgroup_file_notify() local 4150 unsigned long next = last + CGROUP_FILE_NOTIFY_MIN_INTV; in cgroup_file_notify() 4152 if (time_in_range(jiffies, last, next)) { in cgroup_file_notify() 4291 struct cgroup_subsys_state *last, *tmp; in css_rightmost_descendant() local 4296 last = pos; in css_rightmost_descendant() 4299 css_for_each_child(tmp, last) in css_rightmost_descendant() 4303 return last; in css_rightmost_descendant() 4309 struct cgroup_subsys_state *last; in css_leftmost_descendant() local 4312 last = pos; in css_leftmost_descendant() 4316 return last; in css_leftmost_descendant()
|
/kernel/rcu/ |
D | Kconfig | 145 bool "Accelerate last non-dyntick-idle CPU's grace periods"
|