Home
last modified time | relevance | path

Searched refs:next (Results 1 – 25 of 97) sorted by relevance

1234

/kernel/locking/
Dosq_lock.c40 struct optimistic_spin_node *next = NULL; in osq_wait_next() local
72 if (node->next) { in osq_wait_next()
73 next = xchg(&node->next, NULL); in osq_wait_next()
74 if (next) in osq_wait_next()
81 return next; in osq_wait_next()
87 struct optimistic_spin_node *prev, *next; in osq_lock() local
92 node->next = NULL; in osq_lock()
120 WRITE_ONCE(prev->next, node); in osq_lock()
152 if (prev->next == node && in osq_lock()
153 cmpxchg(&prev->next, node, NULL) == node) in osq_lock()
[all …]
Dmcs_spinlock.h18 struct mcs_spinlock *next; member
68 node->next = NULL; in mcs_spin_lock()
88 WRITE_ONCE(prev->next, node); in mcs_spin_lock()
101 struct mcs_spinlock *next = READ_ONCE(node->next); in mcs_spin_unlock() local
103 if (likely(!next)) { in mcs_spin_unlock()
110 while (!(next = READ_ONCE(node->next))) in mcs_spin_unlock()
115 arch_mcs_spin_unlock_contended(&next->locked); in mcs_spin_unlock()
Drwsem-xadd.c126 struct list_head *next; in __rwsem_do_wake() local
129 waiter = list_entry(sem->wait_list.next, struct rwsem_waiter, list); in __rwsem_do_wake()
168 if (waiter->list.next == &sem->wait_list) in __rwsem_do_wake()
171 waiter = list_entry(waiter->list.next, in __rwsem_do_wake()
184 next = sem->wait_list.next; in __rwsem_do_wake()
187 waiter = list_entry(next, struct rwsem_waiter, list); in __rwsem_do_wake()
188 next = waiter->list.next; in __rwsem_do_wake()
203 sem->wait_list.next = next; in __rwsem_do_wake()
204 next->prev = &sem->wait_list; in __rwsem_do_wake()
Dqspinlock.c343 struct mcs_spinlock *prev, *next, *node; in queued_spin_lock_slowpath() local
435 node->next = NULL; in queued_spin_lock_slowpath()
460 WRITE_ONCE(prev->next, node); in queued_spin_lock_slowpath()
507 while (!(next = READ_ONCE(node->next))) in queued_spin_lock_slowpath()
510 arch_mcs_spin_unlock_contended(&next->locked); in queued_spin_lock_slowpath()
511 pv_kick_node(lock, next); in queued_spin_lock_slowpath()
Drwsem-spinlock.c71 waiter = list_entry(sem->wait_list.next, struct rwsem_waiter, list); in __rwsem_do_wake()
84 struct list_head *next = waiter->list.next; in __rwsem_do_wake() local
100 if (next == &sem->wait_list) in __rwsem_do_wake()
102 waiter = list_entry(next, struct rwsem_waiter, list); in __rwsem_do_wake()
119 waiter = list_entry(sem->wait_list.next, struct rwsem_waiter, list); in __rwsem_wake_one_writer()
Dlockdep.c1502 struct held_lock *next, in print_bad_irq_dependency() argument
1522 print_lock(next); in print_bad_irq_dependency()
1529 print_lock_name(hlock_class(next)); in print_bad_irq_dependency()
1548 hlock_class(prev), hlock_class(next)); in print_bad_irq_dependency()
1572 struct held_lock *next, enum lock_usage_bit bit_backwards, in check_usage() argument
1590 that.class = hlock_class(next); in check_usage()
1599 prev, next, in check_usage()
1645 struct held_lock *next, enum lock_usage_bit bit) in check_irq_usage() argument
1653 if (!check_usage(curr, prev, next, bit, in check_irq_usage()
1665 if (!check_usage(curr, prev, next, bit, in check_irq_usage()
[all …]
/kernel/debug/kdb/
Dkdb_support.c693 u32 next; /* offset of next header from start of pool */ member
768 if (!h->next) in debug_kmalloc()
771 h = (struct debug_alloc_header *)(debug_alloc_pool + h->next); in debug_kmalloc()
777 if (best->next == 0 && bestprev == NULL && rem < dah_overhead) in debug_kmalloc()
785 h->next = best->next; in debug_kmalloc()
787 h_offset = best->next; in debug_kmalloc()
792 bestprev->next = h_offset; in debug_kmalloc()
825 h->next = dah_first; in debug_kfree()
833 if (!prev->next || prev->next > h_offset) in debug_kfree()
836 (debug_alloc_pool + prev->next); in debug_kfree()
[all …]
/kernel/
Dtask_work.c35 work->next = head; in task_work_add()
70 pprev = &work->next; in task_work_cancel()
71 else if (cmpxchg(pprev, work, work->next) == work) in task_work_cancel()
90 struct callback_head *work, *head, *next; in task_work_run() local
114 next = work->next; in task_work_run()
116 work = next; in task_work_run()
Djump_label.c265 key->next = NULL; in jump_label_init()
285 struct static_key_mod *next; member
309 for (mod = key->next; mod; mod = mod->next) { in __jump_label_mod_update()
369 key->next = NULL; in jump_label_add_module()
377 jlm->next = key->next; in jump_label_add_module()
378 key->next = jlm; in jump_label_add_module()
405 prev = &key->next; in jump_label_del_module()
406 jlm = key->next; in jump_label_del_module()
409 prev = &jlm->next; in jump_label_del_module()
410 jlm = jlm->next; in jump_label_del_module()
[all …]
Dsoftirq.c454 t->next = NULL; in __tasklet_schedule()
456 __this_cpu_write(tasklet_vec.tail, &(t->next)); in __tasklet_schedule()
467 t->next = NULL; in __tasklet_hi_schedule()
469 __this_cpu_write(tasklet_hi_vec.tail, &(t->next)); in __tasklet_hi_schedule()
479 t->next = __this_cpu_read(tasklet_hi_vec.head); in __tasklet_hi_schedule_first()
498 list = list->next; in tasklet_action()
513 t->next = NULL; in tasklet_action()
515 __this_cpu_write(tasklet_vec.tail, &(t->next)); in tasklet_action()
534 list = list->next; in tasklet_hi_action()
549 t->next = NULL; in tasklet_hi_action()
[all …]
Dresource.c133 .next = r_next,
600 goto next; in __find_resource()
622 next: if (!this || this->end == root->end) in __find_resource()
767 struct resource *first, *next; in __insert_resource() local
785 for (next = first; ; next = next->sibling) { in __insert_resource()
787 if (next->start < new->start || next->end > new->end) in __insert_resource()
788 return next; in __insert_resource()
789 if (!next->sibling) in __insert_resource()
791 if (next->sibling->start > new->end) in __insert_resource()
796 new->sibling = next->sibling; in __insert_resource()
[all …]
Dnotifier.c27 nl = &((*nl)->next); in notifier_chain_register()
29 n->next = *nl; in notifier_chain_register()
42 nl = &((*nl)->next); in notifier_chain_cond_register()
44 n->next = *nl; in notifier_chain_cond_register()
54 rcu_assign_pointer(*nl, n->next); in notifier_chain_unregister()
57 nl = &((*nl)->next); in notifier_chain_unregister()
84 next_nb = rcu_dereference_raw(nb->next); in notifier_call_chain()
Daudit_tree.c472 struct audit_krule *rule, *next; in kill_rules() local
475 list_for_each_entry_safe(rule, next, &tree->rules, rlist) { in kill_rules()
501 p = list_entry(victim->chunks.next, struct node, list); in prune_one()
520 for (p = tree->chunks.next; p != &tree->chunks; p = q) { in trim_marked()
522 q = p->next; in trim_marked()
532 node = list_entry(tree->chunks.next, struct node, list); in trim_marked()
591 while (cursor.next != &tree_list) { in audit_trim_trees()
598 tree = container_of(cursor.next, struct audit_tree, list); in audit_trim_trees()
675 victim = list_entry(prune_list.next, in prune_tree_thread()
805 while (cursor.next != &tree_list) { in audit_tag_tree()
[all …]
/kernel/rcu/
Dtiny.c134 struct rcu_head *next, *list; in __rcu_process_callbacks() local
157 next = list->next; in __rcu_process_callbacks()
158 prefetch(next); in __rcu_process_callbacks()
163 list = next; in __rcu_process_callbacks()
213 head->next = NULL; in __call_rcu()
217 rcp->curtail = &head->next; in __call_rcu()
Dupdate.c542 rhp->next = NULL; in call_rcu_tasks()
547 rcu_tasks_cbs_tail = &rhp->next; in call_rcu_tasks()
651 struct rcu_head *next; in rcu_tasks_kthread() local
779 next = list->next; in rcu_tasks_kthread()
783 list = next; in rcu_tasks_kthread()
/kernel/time/
Dtick-common.c102 ktime_t next = dev->next_event; in tick_handle_periodic() local
123 next = ktime_add(next, tick_period); in tick_handle_periodic()
125 if (!clockevents_program_event(dev, next, false)) in tick_handle_periodic()
157 ktime_t next; in tick_setup_periodic() local
161 next = tick_next_period; in tick_setup_periodic()
167 if (!clockevents_program_event(dev, next, false)) in tick_setup_periodic()
169 next = ktime_add(next, tick_period); in tick_setup_periodic()
Dntp.c513 struct timespec64 next; in sync_cmos_clock() local
549 next.tv_nsec = (NSEC_PER_SEC / 2) - now.tv_nsec - (TICK_NSEC / 2); in sync_cmos_clock()
550 if (next.tv_nsec <= 0) in sync_cmos_clock()
551 next.tv_nsec += NSEC_PER_SEC; in sync_cmos_clock()
554 next.tv_sec = 659; in sync_cmos_clock()
556 next.tv_sec = 0; in sync_cmos_clock()
558 if (next.tv_nsec >= NSEC_PER_SEC) { in sync_cmos_clock()
559 next.tv_sec++; in sync_cmos_clock()
560 next.tv_nsec -= NSEC_PER_SEC; in sync_cmos_clock()
563 &sync_cmos_work, timespec64_to_jiffies(&next)); in sync_cmos_clock()
Dtimer_stats.c57 struct entry *next; member
183 curr = curr->next; in tstat_lookup()
200 curr = curr->next; in tstat_lookup()
207 curr->next = NULL; in tstat_lookup()
213 prev->next = curr; in tstat_lookup()
/kernel/sched/
Dstats.h132 struct task_struct *prev, struct task_struct *next) in __sched_info_switch() argument
142 if (next != rq->idle) in __sched_info_switch()
143 sched_info_arrive(rq, next); in __sched_info_switch()
147 struct task_struct *prev, struct task_struct *next) in sched_info_switch() argument
150 __sched_info_switch(rq, prev, next); in sched_info_switch()
157 #define sched_info_arrive(rq, next) do { } while (0) argument
158 #define sched_info_switch(rq, t, next) do { } while (0) argument
Drt.c87 rt_rq->highest_prio.next = MAX_RT_PRIO; in init_rt_rq()
377 if (p->prio < rq->rt.highest_prio.next) in enqueue_pushable_task()
378 rq->rt.highest_prio.next = p->prio; in enqueue_pushable_task()
389 rq->rt.highest_prio.next = p->prio; in dequeue_pushable_task()
391 rq->rt.highest_prio.next = MAX_RT_PRIO; in dequeue_pushable_task()
456 tg = list_entry_rcu(tg->list.next, in next_task_group()
657 goto next; in do_balance_runtime()
675 next: in do_balance_runtime()
1629 struct sched_rt_entity *next = NULL; in pick_next_rt_entity() local
1637 next = list_entry(queue->next, struct sched_rt_entity, run_list); in pick_next_rt_entity()
[all …]
/kernel/events/
Duprobes.c504 uc->next = uprobe->consumers; in consumer_add()
520 for (con = &uprobe->consumers; *con; con = &(*con)->next) { in consumer_del()
522 *con = uc->next; in consumer_del()
627 for (uc = uprobe->consumers; uc; uc = uc->next) { in filter_chain()
695 struct map_info *next; member
702 struct map_info *next = info->next; in free_map_info() local
704 return next; in free_map_info()
731 prev->next = NULL; in build_map_info()
742 prev = prev->next; in build_map_info()
743 info->next = curr; in build_map_info()
[all …]
/kernel/trace/
Dftrace.c152 while (likely(op = rcu_dereference_raw_notrace((op)->next)) && \
179 ops != &ftrace_list_end; ops = ops->next) in ftrace_nr_registered_ops()
289 } else if (ftrace_ops_list->next == &ftrace_list_end) { in update_ftrace_function()
353 ops->next = *list; in add_ftrace_ops()
371 if (*list == ops && ops->next == &ftrace_list_end) { in remove_ftrace_ops()
376 for (p = list; *p != &ftrace_list_end; p = &(*p)->next) in remove_ftrace_ops()
383 *p = (*p)->next; in remove_ftrace_ops()
513 struct ftrace_profile_page *next; member
555 pg = pg->next; in function_stat_next()
694 pg = pg->next; in ftrace_profile_reset()
[all …]
Dtrace_sched_wakeup.c373 struct task_struct *next, in tracing_sched_switch_trace() argument
389 entry->next_pid = next->pid; in tracing_sched_switch_trace()
390 entry->next_prio = next->prio; in tracing_sched_switch_trace()
391 entry->next_state = next->state; in tracing_sched_switch_trace()
392 entry->next_cpu = task_cpu(next); in tracing_sched_switch_trace()
428 struct task_struct *prev, struct task_struct *next) in probe_wakeup_sched_switch() argument
451 if (next != wakeup_task) in probe_wakeup_sched_switch()
466 if (unlikely(!tracer_enabled || next != wakeup_task)) in probe_wakeup_sched_switch()
473 tracing_sched_switch_trace(wakeup_trace, prev, next, flags, pc); in probe_wakeup_sched_switch()
/kernel/gcov/
Dgcc_4_7.c90 struct gcov_info *next; member
128 return info->next; in gcov_info_next()
137 info->next = gcov_info_head; in gcov_info_link()
149 prev->next = info->next; in gcov_info_unlink()
151 gcov_info_head = info->next; in gcov_info_unlink()
271 dup->next = NULL; in gcov_info_dup()
Dgcc_3_4.c74 struct gcov_info *next; member
113 return info->next; in gcov_info_next()
122 info->next = gcov_info_head; in gcov_info_link()
134 prev->next = info->next; in gcov_info_unlink()
136 gcov_info_head = info->next; in gcov_info_unlink()

1234