Home
last modified time | relevance | path

Searched refs:prev (Results 1 – 25 of 55) sorted by relevance

123

/kernel/locking/
Dosq_lock.c44 struct optimistic_spin_node *prev) in osq_wait_next() argument
55 old = prev ? prev->cpu : OSQ_UNLOCKED_VAL; in osq_wait_next()
93 struct optimistic_spin_node *prev, *next; in osq_lock() local
111 prev = decode_cpu(old); in osq_lock()
112 node->prev = prev; in osq_lock()
126 WRITE_ONCE(prev->next, node); in osq_lock()
143 if (need_resched() || vcpu_is_preempted(node_cpu(node->prev))) in osq_lock()
160 if (prev->next == node && in osq_lock()
161 cmpxchg(&prev->next, node, NULL) == node) in osq_lock()
178 prev = READ_ONCE(node->prev); in osq_lock()
[all …]
Dmcs_spinlock.h67 struct mcs_spinlock *prev; in mcs_spin_lock() local
79 prev = xchg(lock, node); in mcs_spin_lock()
80 if (likely(prev == NULL)) { in mcs_spin_lock()
91 WRITE_ONCE(prev->next, node); in mcs_spin_lock()
Dqspinlock.c273 struct mcs_spinlock *prev) { } in __pv_wait_node() argument
316 struct mcs_spinlock *prev, *next, *node; in queued_spin_lock_slowpath() local
467 prev = decode_tail(old); in queued_spin_lock_slowpath()
470 WRITE_ONCE(prev->next, node); in queued_spin_lock_slowpath()
472 pv_wait_node(node, prev); in queued_spin_lock_slowpath()
Dqspinlock_paravirt.h267 pv_wait_early(struct pv_node *prev, int loop) in pv_wait_early() argument
272 return READ_ONCE(prev->state) != vcpu_running; in pv_wait_early()
293 static void pv_wait_node(struct mcs_spinlock *node, struct mcs_spinlock *prev) in pv_wait_node() argument
296 struct pv_node *pp = (struct pv_node *)prev; in pv_wait_node()
Dlockdep.c2029 struct held_lock *prev, in print_bad_irq_dependency() argument
2053 print_lock(prev); in print_bad_irq_dependency()
2055 print_lock_name(hlock_class(prev)); in print_bad_irq_dependency()
2076 hlock_class(prev), hlock_class(next)); in print_bad_irq_dependency()
2223 static int check_irq_usage(struct task_struct *curr, struct held_lock *prev, in check_irq_usage() argument
2238 this.class = hlock_class(prev); in check_irq_usage()
2294 prev, next, in check_irq_usage()
2316 struct held_lock *prev, struct held_lock *next) in check_irq_usage() argument
2332 struct lock_class *prev = hlock_class(prv); in print_deadlock_scenario() local
2338 __print_lock_name(prev); in print_deadlock_scenario()
[all …]
/kernel/
Dcfi.c98 static void prepare_next_shadow(const struct cfi_shadow __rcu *prev, in prepare_next_shadow() argument
106 if (!prev) in prepare_next_shadow()
110 if (prev->r.min_page == next->r.min_page) { in prepare_next_shadow()
111 memcpy(next->shadow, prev->shadow, sizeof(next->shadow)); in prepare_next_shadow()
117 if (prev->shadow[i] == SHADOW_INVALID) in prepare_next_shadow()
120 index = ptr_to_shadow(next, shadow_to_page(prev, i)); in prepare_next_shadow()
125 shadow_to_ptr(prev, prev->shadow[i])); in prepare_next_shadow()
188 struct cfi_shadow *prev; in update_shadow() local
198 prev = rcu_dereference_protected(cfi_shadow, 1); in update_shadow()
199 prepare_next_shadow(prev, next); in update_shadow()
[all …]
Dseccomp.c131 struct seccomp_filter *prev; member
271 for (; f; f = f->prev) { in seccomp_run_filters()
322 for (; child; child = child->prev) in is_ancestor()
520 for (walker = current->seccomp.filter; walker; walker = walker->prev) in seccomp_attach_filter()
542 filter->prev = current->seccomp.filter; in seccomp_attach_filter()
579 orig = orig->prev; in __put_seccomp_filter()
1215 for (cur = current->seccomp.filter; cur; cur = cur->prev) { in init_listener()
1475 for (filter = orig; filter; filter = filter->prev) in get_nth_filter()
1484 for (filter = orig; filter && count > 1; filter = filter->prev) in get_nth_filter()
Djump_label.c674 struct static_key_mod *jlm, **prev; in jump_label_del_module() local
689 prev = &key->next; in jump_label_del_module()
693 prev = &jlm->next; in jump_label_del_module()
701 if (prev == &key->next) in jump_label_del_module()
704 *prev = jlm->next; in jump_label_del_module()
Dsys.c2266 struct vm_area_struct **prev, in prctl_update_vma_anon_name() argument
2275 *prev = vma; in prctl_update_vma_anon_name()
2280 *prev = vma_merge(mm, *prev, start, end, vma->vm_flags, vma->anon_vma, in prctl_update_vma_anon_name()
2283 if (*prev) { in prctl_update_vma_anon_name()
2284 vma = *prev; in prctl_update_vma_anon_name()
2288 *prev = vma; in prctl_update_vma_anon_name()
2316 struct vm_area_struct *vma, *prev; in prctl_set_vma_anon_name() local
2325 vma = find_vma_prev(current->mm, start, &prev); in prctl_set_vma_anon_name()
2327 prev = vma; in prctl_set_vma_anon_name()
2349 error = prctl_update_vma_anon_name(vma, &prev, start, tmp, in prctl_set_vma_anon_name()
[all …]
Duser_namespace.c723 struct uid_gid_extent *prev; in mappings_overlap() local
726 prev = &new_map->extent[idx]; in mappings_overlap()
728 prev = &new_map->forward[idx]; in mappings_overlap()
730 prev_upper_first = prev->first; in mappings_overlap()
731 prev_lower_first = prev->lower_first; in mappings_overlap()
732 prev_upper_last = prev_upper_first + prev->count - 1; in mappings_overlap()
733 prev_lower_last = prev_lower_first + prev->count - 1; in mappings_overlap()
/kernel/sched/
Dcputime.c417 void vtime_common_task_switch(struct task_struct *prev) in vtime_common_task_switch() argument
419 if (is_idle_task(prev)) in vtime_common_task_switch()
420 vtime_account_idle(prev); in vtime_common_task_switch()
422 vtime_account_system(prev); in vtime_common_task_switch()
424 vtime_flush(prev); in vtime_common_task_switch()
425 arch_vtime_task_switch(prev); in vtime_common_task_switch()
451 void cputime_adjust(struct task_cputime *curr, struct prev_cputime *prev, in cputime_adjust() argument
598 void cputime_adjust(struct task_cputime *curr, struct prev_cputime *prev, in cputime_adjust() argument
605 raw_spin_lock_irqsave(&prev->lock, flags); in cputime_adjust()
616 if (prev->stime + prev->utime >= rtime) in cputime_adjust()
[all …]
Dcore.c3082 static inline void finish_task(struct task_struct *prev) in finish_task() argument
3095 smp_store_release(&prev->on_cpu, 0); in finish_task()
3153 prepare_task_switch(struct rq *rq, struct task_struct *prev, in prepare_task_switch() argument
3156 kcov_prepare_switch(prev); in prepare_task_switch()
3157 sched_info_switch(rq, prev, next); in prepare_task_switch()
3158 perf_event_task_sched_out(prev, next); in prepare_task_switch()
3159 rseq_preempt(prev); in prepare_task_switch()
3160 fire_sched_out_preempt_notifiers(prev, next); in prepare_task_switch()
3184 static struct rq *finish_task_switch(struct task_struct *prev) in finish_task_switch() argument
3220 prev_state = prev->state; in finish_task_switch()
[all …]
Dstop_task.c20 balance_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) in balance_stop() argument
38 pick_next_task_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) in pick_next_task_stop() argument
40 WARN_ON_ONCE(prev || rf); in pick_next_task_stop()
66 static void put_prev_task_stop(struct rq *rq, struct task_struct *prev) in put_prev_task_stop() argument
Dstats.h226 __sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next) in __sched_info_switch() argument
233 if (prev != rq->idle) in __sched_info_switch()
234 sched_info_depart(rq, prev); in __sched_info_switch()
241 sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next) in sched_info_switch() argument
244 __sched_info_switch(rq, prev, next); in sched_info_switch()
Didle.c370 balance_idle(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) in balance_idle() argument
384 static void put_prev_task_idle(struct rq *rq, struct task_struct *prev) in put_prev_task_idle() argument
395 pick_next_task_idle(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) in pick_next_task_idle() argument
399 if (prev) in pick_next_task_idle()
400 put_prev_task(rq, prev); in pick_next_task_idle()
Dautogroup.c138 struct autogroup *prev; in autogroup_move_group() local
144 prev = p->signal->autogroup; in autogroup_move_group()
145 if (prev == ag) { in autogroup_move_group()
166 autogroup_kref_put(prev); in autogroup_move_group()
/kernel/debug/kdb/
Dkdb_support.c748 struct debug_alloc_header *best, *bestprev, *prev, *h; in debug_kmalloc() local
760 prev = best = bestprev = NULL; in debug_kmalloc()
764 bestprev = prev; in debug_kmalloc()
770 prev = h; in debug_kmalloc()
828 struct debug_alloc_header *prev; in debug_kfree() local
830 prev = (struct debug_alloc_header *)(debug_alloc_pool + in debug_kfree()
833 if (!prev->next || prev->next > h_offset) in debug_kfree()
835 prev = (struct debug_alloc_header *) in debug_kfree()
836 (debug_alloc_pool + prev->next); in debug_kfree()
838 prev_offset = (char *)prev - debug_alloc_pool; in debug_kfree()
[all …]
/kernel/gcov/
Dbase.c59 struct gcov_info *prev = NULL; in gcov_module_notifier() local
68 gcov_info_unlink(prev, info); in gcov_module_notifier()
72 prev = info; in gcov_module_notifier()
Dgcc_3_4.c132 void gcov_info_unlink(struct gcov_info *prev, struct gcov_info *info) in gcov_info_unlink() argument
134 if (prev) in gcov_info_unlink()
135 prev->next = info->next; in gcov_info_unlink()
/kernel/trace/
Dtrace_sched_wakeup.c376 struct task_struct *prev, in tracing_sched_switch_trace() argument
390 entry->prev_pid = prev->pid; in tracing_sched_switch_trace()
391 entry->prev_prio = prev->prio; in tracing_sched_switch_trace()
392 entry->prev_state = task_state_index(prev); in tracing_sched_switch_trace()
432 struct task_struct *prev, struct task_struct *next) in probe_wakeup_sched_switch() argument
441 tracing_record_cmdline(prev); in probe_wakeup_sched_switch()
477 tracing_sched_switch_trace(wakeup_trace, prev, next, flags, pc); in probe_wakeup_sched_switch()
Dring_buffer.c891 struct list_head *list = page->list.prev; in rb_is_reader_page()
923 rb_set_list_to_head(cpu_buffer, head->list.prev); in rb_head_page_activate()
950 struct buffer_page *prev, in rb_head_page_set() argument
957 list = &prev->list; in rb_head_page_set()
973 struct buffer_page *prev, in rb_head_page_set_update() argument
976 return rb_head_page_set(cpu_buffer, head, prev, in rb_head_page_set_update()
982 struct buffer_page *prev, in rb_head_page_set_head() argument
985 return rb_head_page_set(cpu_buffer, head, prev, in rb_head_page_set_head()
991 struct buffer_page *prev, in rb_head_page_set_normal() argument
994 return rb_head_page_set(cpu_buffer, head, prev, in rb_head_page_set_normal()
[all …]
Dtrace_sched_switch.c25 struct task_struct *prev, struct task_struct *next) in probe_sched_switch() argument
34 tracing_record_taskinfo_sched_switch(prev, next, flags); in probe_sched_switch()
/kernel/livepatch/
Dtransition.c223 struct klp_func *prev; in klp_check_stack_func() local
225 prev = list_next_entry(func, stack_node); in klp_check_stack_func()
226 func_addr = (unsigned long)prev->new_func; in klp_check_stack_func()
227 func_size = prev->new_size; in klp_check_stack_func()
/kernel/events/
Duprobes.c981 struct map_info *prev = NULL; in build_map_info() local
991 if (!prev && !more) { in build_map_info()
996 prev = kmalloc(sizeof(struct map_info), in build_map_info()
998 if (prev) in build_map_info()
999 prev->next = NULL; in build_map_info()
1001 if (!prev) { in build_map_info()
1009 info = prev; in build_map_info()
1010 prev = prev->next; in build_map_info()
1022 prev = curr; in build_map_info()
1034 info->next = prev; in build_map_info()
[all …]
/kernel/time/
Dtick-sched.c285 int prev; in tick_nohz_dep_set_all() local
287 prev = atomic_fetch_or(BIT(bit), dep); in tick_nohz_dep_set_all()
288 if (!prev) in tick_nohz_dep_set_all()
312 int prev; in tick_nohz_dep_set_cpu() local
317 prev = atomic_fetch_or(BIT(bit), &ts->tick_dep_mask); in tick_nohz_dep_set_cpu()
318 if (!prev) { in tick_nohz_dep_set_cpu()

123