/kernel/locking/ |
D | osq_lock.c | 44 struct optimistic_spin_node *prev) in osq_wait_next() argument 55 old = prev ? prev->cpu : OSQ_UNLOCKED_VAL; in osq_wait_next() 93 struct optimistic_spin_node *prev, *next; in osq_lock() local 111 prev = decode_cpu(old); in osq_lock() 112 node->prev = prev; in osq_lock() 126 WRITE_ONCE(prev->next, node); in osq_lock() 143 if (need_resched() || vcpu_is_preempted(node_cpu(node->prev))) in osq_lock() 160 if (prev->next == node && in osq_lock() 161 cmpxchg(&prev->next, node, NULL) == node) in osq_lock() 178 prev = READ_ONCE(node->prev); in osq_lock() [all …]
|
D | mcs_spinlock.h | 67 struct mcs_spinlock *prev; in mcs_spin_lock() local 79 prev = xchg(lock, node); in mcs_spin_lock() 80 if (likely(prev == NULL)) { in mcs_spin_lock() 91 WRITE_ONCE(prev->next, node); in mcs_spin_lock()
|
D | qspinlock.c | 273 struct mcs_spinlock *prev) { } in __pv_wait_node() argument 316 struct mcs_spinlock *prev, *next, *node; in queued_spin_lock_slowpath() local 467 prev = decode_tail(old); in queued_spin_lock_slowpath() 470 WRITE_ONCE(prev->next, node); in queued_spin_lock_slowpath() 472 pv_wait_node(node, prev); in queued_spin_lock_slowpath()
|
D | qspinlock_paravirt.h | 267 pv_wait_early(struct pv_node *prev, int loop) in pv_wait_early() argument 272 return READ_ONCE(prev->state) != vcpu_running; in pv_wait_early() 293 static void pv_wait_node(struct mcs_spinlock *node, struct mcs_spinlock *prev) in pv_wait_node() argument 296 struct pv_node *pp = (struct pv_node *)prev; in pv_wait_node()
|
D | lockdep.c | 2029 struct held_lock *prev, in print_bad_irq_dependency() argument 2053 print_lock(prev); in print_bad_irq_dependency() 2055 print_lock_name(hlock_class(prev)); in print_bad_irq_dependency() 2076 hlock_class(prev), hlock_class(next)); in print_bad_irq_dependency() 2223 static int check_irq_usage(struct task_struct *curr, struct held_lock *prev, in check_irq_usage() argument 2238 this.class = hlock_class(prev); in check_irq_usage() 2294 prev, next, in check_irq_usage() 2316 struct held_lock *prev, struct held_lock *next) in check_irq_usage() argument 2332 struct lock_class *prev = hlock_class(prv); in print_deadlock_scenario() local 2338 __print_lock_name(prev); in print_deadlock_scenario() [all …]
|
/kernel/ |
D | cfi.c | 98 static void prepare_next_shadow(const struct cfi_shadow __rcu *prev, in prepare_next_shadow() argument 106 if (!prev) in prepare_next_shadow() 110 if (prev->r.min_page == next->r.min_page) { in prepare_next_shadow() 111 memcpy(next->shadow, prev->shadow, sizeof(next->shadow)); in prepare_next_shadow() 117 if (prev->shadow[i] == SHADOW_INVALID) in prepare_next_shadow() 120 index = ptr_to_shadow(next, shadow_to_page(prev, i)); in prepare_next_shadow() 125 shadow_to_ptr(prev, prev->shadow[i])); in prepare_next_shadow() 188 struct cfi_shadow *prev; in update_shadow() local 198 prev = rcu_dereference_protected(cfi_shadow, 1); in update_shadow() 199 prepare_next_shadow(prev, next); in update_shadow() [all …]
|
D | seccomp.c | 131 struct seccomp_filter *prev; member 271 for (; f; f = f->prev) { in seccomp_run_filters() 322 for (; child; child = child->prev) in is_ancestor() 520 for (walker = current->seccomp.filter; walker; walker = walker->prev) in seccomp_attach_filter() 542 filter->prev = current->seccomp.filter; in seccomp_attach_filter() 579 orig = orig->prev; in __put_seccomp_filter() 1215 for (cur = current->seccomp.filter; cur; cur = cur->prev) { in init_listener() 1475 for (filter = orig; filter; filter = filter->prev) in get_nth_filter() 1484 for (filter = orig; filter && count > 1; filter = filter->prev) in get_nth_filter()
|
D | jump_label.c | 674 struct static_key_mod *jlm, **prev; in jump_label_del_module() local 689 prev = &key->next; in jump_label_del_module() 693 prev = &jlm->next; in jump_label_del_module() 701 if (prev == &key->next) in jump_label_del_module() 704 *prev = jlm->next; in jump_label_del_module()
|
D | sys.c | 2266 struct vm_area_struct **prev, in prctl_update_vma_anon_name() argument 2275 *prev = vma; in prctl_update_vma_anon_name() 2280 *prev = vma_merge(mm, *prev, start, end, vma->vm_flags, vma->anon_vma, in prctl_update_vma_anon_name() 2283 if (*prev) { in prctl_update_vma_anon_name() 2284 vma = *prev; in prctl_update_vma_anon_name() 2288 *prev = vma; in prctl_update_vma_anon_name() 2316 struct vm_area_struct *vma, *prev; in prctl_set_vma_anon_name() local 2325 vma = find_vma_prev(current->mm, start, &prev); in prctl_set_vma_anon_name() 2327 prev = vma; in prctl_set_vma_anon_name() 2349 error = prctl_update_vma_anon_name(vma, &prev, start, tmp, in prctl_set_vma_anon_name() [all …]
|
D | user_namespace.c | 723 struct uid_gid_extent *prev; in mappings_overlap() local 726 prev = &new_map->extent[idx]; in mappings_overlap() 728 prev = &new_map->forward[idx]; in mappings_overlap() 730 prev_upper_first = prev->first; in mappings_overlap() 731 prev_lower_first = prev->lower_first; in mappings_overlap() 732 prev_upper_last = prev_upper_first + prev->count - 1; in mappings_overlap() 733 prev_lower_last = prev_lower_first + prev->count - 1; in mappings_overlap()
|
/kernel/sched/ |
D | cputime.c | 417 void vtime_common_task_switch(struct task_struct *prev) in vtime_common_task_switch() argument 419 if (is_idle_task(prev)) in vtime_common_task_switch() 420 vtime_account_idle(prev); in vtime_common_task_switch() 422 vtime_account_system(prev); in vtime_common_task_switch() 424 vtime_flush(prev); in vtime_common_task_switch() 425 arch_vtime_task_switch(prev); in vtime_common_task_switch() 451 void cputime_adjust(struct task_cputime *curr, struct prev_cputime *prev, in cputime_adjust() argument 598 void cputime_adjust(struct task_cputime *curr, struct prev_cputime *prev, in cputime_adjust() argument 605 raw_spin_lock_irqsave(&prev->lock, flags); in cputime_adjust() 616 if (prev->stime + prev->utime >= rtime) in cputime_adjust() [all …]
|
D | core.c | 3082 static inline void finish_task(struct task_struct *prev) in finish_task() argument 3095 smp_store_release(&prev->on_cpu, 0); in finish_task() 3153 prepare_task_switch(struct rq *rq, struct task_struct *prev, in prepare_task_switch() argument 3156 kcov_prepare_switch(prev); in prepare_task_switch() 3157 sched_info_switch(rq, prev, next); in prepare_task_switch() 3158 perf_event_task_sched_out(prev, next); in prepare_task_switch() 3159 rseq_preempt(prev); in prepare_task_switch() 3160 fire_sched_out_preempt_notifiers(prev, next); in prepare_task_switch() 3184 static struct rq *finish_task_switch(struct task_struct *prev) in finish_task_switch() argument 3220 prev_state = prev->state; in finish_task_switch() [all …]
|
D | stop_task.c | 20 balance_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) in balance_stop() argument 38 pick_next_task_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) in pick_next_task_stop() argument 40 WARN_ON_ONCE(prev || rf); in pick_next_task_stop() 66 static void put_prev_task_stop(struct rq *rq, struct task_struct *prev) in put_prev_task_stop() argument
|
D | stats.h | 226 __sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next) in __sched_info_switch() argument 233 if (prev != rq->idle) in __sched_info_switch() 234 sched_info_depart(rq, prev); in __sched_info_switch() 241 sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next) in sched_info_switch() argument 244 __sched_info_switch(rq, prev, next); in sched_info_switch()
|
D | idle.c | 370 balance_idle(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) in balance_idle() argument 384 static void put_prev_task_idle(struct rq *rq, struct task_struct *prev) in put_prev_task_idle() argument 395 pick_next_task_idle(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) in pick_next_task_idle() argument 399 if (prev) in pick_next_task_idle() 400 put_prev_task(rq, prev); in pick_next_task_idle()
|
D | autogroup.c | 138 struct autogroup *prev; in autogroup_move_group() local 144 prev = p->signal->autogroup; in autogroup_move_group() 145 if (prev == ag) { in autogroup_move_group() 166 autogroup_kref_put(prev); in autogroup_move_group()
|
/kernel/debug/kdb/ |
D | kdb_support.c | 748 struct debug_alloc_header *best, *bestprev, *prev, *h; in debug_kmalloc() local 760 prev = best = bestprev = NULL; in debug_kmalloc() 764 bestprev = prev; in debug_kmalloc() 770 prev = h; in debug_kmalloc() 828 struct debug_alloc_header *prev; in debug_kfree() local 830 prev = (struct debug_alloc_header *)(debug_alloc_pool + in debug_kfree() 833 if (!prev->next || prev->next > h_offset) in debug_kfree() 835 prev = (struct debug_alloc_header *) in debug_kfree() 836 (debug_alloc_pool + prev->next); in debug_kfree() 838 prev_offset = (char *)prev - debug_alloc_pool; in debug_kfree() [all …]
|
/kernel/gcov/ |
D | base.c | 59 struct gcov_info *prev = NULL; in gcov_module_notifier() local 68 gcov_info_unlink(prev, info); in gcov_module_notifier() 72 prev = info; in gcov_module_notifier()
|
D | gcc_3_4.c | 132 void gcov_info_unlink(struct gcov_info *prev, struct gcov_info *info) in gcov_info_unlink() argument 134 if (prev) in gcov_info_unlink() 135 prev->next = info->next; in gcov_info_unlink()
|
/kernel/trace/ |
D | trace_sched_wakeup.c | 376 struct task_struct *prev, in tracing_sched_switch_trace() argument 390 entry->prev_pid = prev->pid; in tracing_sched_switch_trace() 391 entry->prev_prio = prev->prio; in tracing_sched_switch_trace() 392 entry->prev_state = task_state_index(prev); in tracing_sched_switch_trace() 432 struct task_struct *prev, struct task_struct *next) in probe_wakeup_sched_switch() argument 441 tracing_record_cmdline(prev); in probe_wakeup_sched_switch() 477 tracing_sched_switch_trace(wakeup_trace, prev, next, flags, pc); in probe_wakeup_sched_switch()
|
D | ring_buffer.c | 891 struct list_head *list = page->list.prev; in rb_is_reader_page() 923 rb_set_list_to_head(cpu_buffer, head->list.prev); in rb_head_page_activate() 950 struct buffer_page *prev, in rb_head_page_set() argument 957 list = &prev->list; in rb_head_page_set() 973 struct buffer_page *prev, in rb_head_page_set_update() argument 976 return rb_head_page_set(cpu_buffer, head, prev, in rb_head_page_set_update() 982 struct buffer_page *prev, in rb_head_page_set_head() argument 985 return rb_head_page_set(cpu_buffer, head, prev, in rb_head_page_set_head() 991 struct buffer_page *prev, in rb_head_page_set_normal() argument 994 return rb_head_page_set(cpu_buffer, head, prev, in rb_head_page_set_normal() [all …]
|
D | trace_sched_switch.c | 25 struct task_struct *prev, struct task_struct *next) in probe_sched_switch() argument 34 tracing_record_taskinfo_sched_switch(prev, next, flags); in probe_sched_switch()
|
/kernel/livepatch/ |
D | transition.c | 223 struct klp_func *prev; in klp_check_stack_func() local 225 prev = list_next_entry(func, stack_node); in klp_check_stack_func() 226 func_addr = (unsigned long)prev->new_func; in klp_check_stack_func() 227 func_size = prev->new_size; in klp_check_stack_func()
|
/kernel/events/ |
D | uprobes.c | 981 struct map_info *prev = NULL; in build_map_info() local 991 if (!prev && !more) { in build_map_info() 996 prev = kmalloc(sizeof(struct map_info), in build_map_info() 998 if (prev) in build_map_info() 999 prev->next = NULL; in build_map_info() 1001 if (!prev) { in build_map_info() 1009 info = prev; in build_map_info() 1010 prev = prev->next; in build_map_info() 1022 prev = curr; in build_map_info() 1034 info->next = prev; in build_map_info() [all …]
|
/kernel/time/ |
D | tick-sched.c | 285 int prev; in tick_nohz_dep_set_all() local 287 prev = atomic_fetch_or(BIT(bit), dep); in tick_nohz_dep_set_all() 288 if (!prev) in tick_nohz_dep_set_all() 312 int prev; in tick_nohz_dep_set_cpu() local 317 prev = atomic_fetch_or(BIT(bit), &ts->tick_dep_mask); in tick_nohz_dep_set_cpu() 318 if (!prev) { in tick_nohz_dep_set_cpu()
|