/kernel/locking/ |
D | osq_lock.c | 40 struct optimistic_spin_node *next = NULL; in osq_wait_next() local 72 if (node->next) { in osq_wait_next() 73 next = xchg(&node->next, NULL); in osq_wait_next() 74 if (next) in osq_wait_next() 81 return next; in osq_wait_next() 87 struct optimistic_spin_node *prev, *next; in osq_lock() local 92 node->next = NULL; in osq_lock() 120 WRITE_ONCE(prev->next, node); in osq_lock() 152 if (prev->next == node && in osq_lock() 153 cmpxchg(&prev->next, node, NULL) == node) in osq_lock() [all …]
|
D | mcs_spinlock.h | 18 struct mcs_spinlock *next; member 68 node->next = NULL; in mcs_spin_lock() 88 WRITE_ONCE(prev->next, node); in mcs_spin_lock() 101 struct mcs_spinlock *next = READ_ONCE(node->next); in mcs_spin_unlock() local 103 if (likely(!next)) { in mcs_spin_unlock() 110 while (!(next = READ_ONCE(node->next))) in mcs_spin_unlock() 115 arch_mcs_spin_unlock_contended(&next->locked); in mcs_spin_unlock()
|
D | rwsem-xadd.c | 126 struct list_head *next; in __rwsem_do_wake() local 129 waiter = list_entry(sem->wait_list.next, struct rwsem_waiter, list); in __rwsem_do_wake() 168 if (waiter->list.next == &sem->wait_list) in __rwsem_do_wake() 171 waiter = list_entry(waiter->list.next, in __rwsem_do_wake() 184 next = sem->wait_list.next; in __rwsem_do_wake() 187 waiter = list_entry(next, struct rwsem_waiter, list); in __rwsem_do_wake() 188 next = waiter->list.next; in __rwsem_do_wake() 203 sem->wait_list.next = next; in __rwsem_do_wake() 204 next->prev = &sem->wait_list; in __rwsem_do_wake()
|
D | qspinlock.c | 343 struct mcs_spinlock *prev, *next, *node; in queued_spin_lock_slowpath() local 435 node->next = NULL; in queued_spin_lock_slowpath() 460 WRITE_ONCE(prev->next, node); in queued_spin_lock_slowpath() 507 while (!(next = READ_ONCE(node->next))) in queued_spin_lock_slowpath() 510 arch_mcs_spin_unlock_contended(&next->locked); in queued_spin_lock_slowpath() 511 pv_kick_node(lock, next); in queued_spin_lock_slowpath()
|
D | rwsem-spinlock.c | 71 waiter = list_entry(sem->wait_list.next, struct rwsem_waiter, list); in __rwsem_do_wake() 84 struct list_head *next = waiter->list.next; in __rwsem_do_wake() local 100 if (next == &sem->wait_list) in __rwsem_do_wake() 102 waiter = list_entry(next, struct rwsem_waiter, list); in __rwsem_do_wake() 119 waiter = list_entry(sem->wait_list.next, struct rwsem_waiter, list); in __rwsem_wake_one_writer()
|
D | lockdep.c | 1502 struct held_lock *next, in print_bad_irq_dependency() argument 1522 print_lock(next); in print_bad_irq_dependency() 1529 print_lock_name(hlock_class(next)); in print_bad_irq_dependency() 1548 hlock_class(prev), hlock_class(next)); in print_bad_irq_dependency() 1572 struct held_lock *next, enum lock_usage_bit bit_backwards, in check_usage() argument 1590 that.class = hlock_class(next); in check_usage() 1599 prev, next, in check_usage() 1645 struct held_lock *next, enum lock_usage_bit bit) in check_irq_usage() argument 1653 if (!check_usage(curr, prev, next, bit, in check_irq_usage() 1665 if (!check_usage(curr, prev, next, bit, in check_irq_usage() [all …]
|
/kernel/debug/kdb/ |
D | kdb_support.c | 693 u32 next; /* offset of next header from start of pool */ member 768 if (!h->next) in debug_kmalloc() 771 h = (struct debug_alloc_header *)(debug_alloc_pool + h->next); in debug_kmalloc() 777 if (best->next == 0 && bestprev == NULL && rem < dah_overhead) in debug_kmalloc() 785 h->next = best->next; in debug_kmalloc() 787 h_offset = best->next; in debug_kmalloc() 792 bestprev->next = h_offset; in debug_kmalloc() 825 h->next = dah_first; in debug_kfree() 833 if (!prev->next || prev->next > h_offset) in debug_kfree() 836 (debug_alloc_pool + prev->next); in debug_kfree() [all …]
|
/kernel/ |
D | task_work.c | 35 work->next = head; in task_work_add() 70 pprev = &work->next; in task_work_cancel() 71 else if (cmpxchg(pprev, work, work->next) == work) in task_work_cancel() 90 struct callback_head *work, *head, *next; in task_work_run() local 114 next = work->next; in task_work_run() 116 work = next; in task_work_run()
|
D | jump_label.c | 265 key->next = NULL; in jump_label_init() 285 struct static_key_mod *next; member 309 for (mod = key->next; mod; mod = mod->next) { in __jump_label_mod_update() 369 key->next = NULL; in jump_label_add_module() 377 jlm->next = key->next; in jump_label_add_module() 378 key->next = jlm; in jump_label_add_module() 405 prev = &key->next; in jump_label_del_module() 406 jlm = key->next; in jump_label_del_module() 409 prev = &jlm->next; in jump_label_del_module() 410 jlm = jlm->next; in jump_label_del_module() [all …]
|
D | softirq.c | 454 t->next = NULL; in __tasklet_schedule() 456 __this_cpu_write(tasklet_vec.tail, &(t->next)); in __tasklet_schedule() 467 t->next = NULL; in __tasklet_hi_schedule() 469 __this_cpu_write(tasklet_hi_vec.tail, &(t->next)); in __tasklet_hi_schedule() 479 t->next = __this_cpu_read(tasklet_hi_vec.head); in __tasklet_hi_schedule_first() 498 list = list->next; in tasklet_action() 513 t->next = NULL; in tasklet_action() 515 __this_cpu_write(tasklet_vec.tail, &(t->next)); in tasklet_action() 534 list = list->next; in tasklet_hi_action() 549 t->next = NULL; in tasklet_hi_action() [all …]
|
D | resource.c | 133 .next = r_next, 600 goto next; in __find_resource() 622 next: if (!this || this->end == root->end) in __find_resource() 767 struct resource *first, *next; in __insert_resource() local 785 for (next = first; ; next = next->sibling) { in __insert_resource() 787 if (next->start < new->start || next->end > new->end) in __insert_resource() 788 return next; in __insert_resource() 789 if (!next->sibling) in __insert_resource() 791 if (next->sibling->start > new->end) in __insert_resource() 796 new->sibling = next->sibling; in __insert_resource() [all …]
|
D | notifier.c | 27 nl = &((*nl)->next); in notifier_chain_register() 29 n->next = *nl; in notifier_chain_register() 42 nl = &((*nl)->next); in notifier_chain_cond_register() 44 n->next = *nl; in notifier_chain_cond_register() 54 rcu_assign_pointer(*nl, n->next); in notifier_chain_unregister() 57 nl = &((*nl)->next); in notifier_chain_unregister() 84 next_nb = rcu_dereference_raw(nb->next); in notifier_call_chain()
|
D | audit_tree.c | 472 struct audit_krule *rule, *next; in kill_rules() local 475 list_for_each_entry_safe(rule, next, &tree->rules, rlist) { in kill_rules() 501 p = list_entry(victim->chunks.next, struct node, list); in prune_one() 520 for (p = tree->chunks.next; p != &tree->chunks; p = q) { in trim_marked() 522 q = p->next; in trim_marked() 532 node = list_entry(tree->chunks.next, struct node, list); in trim_marked() 591 while (cursor.next != &tree_list) { in audit_trim_trees() 598 tree = container_of(cursor.next, struct audit_tree, list); in audit_trim_trees() 675 victim = list_entry(prune_list.next, in prune_tree_thread() 805 while (cursor.next != &tree_list) { in audit_tag_tree() [all …]
|
/kernel/rcu/ |
D | tiny.c | 134 struct rcu_head *next, *list; in __rcu_process_callbacks() local 157 next = list->next; in __rcu_process_callbacks() 158 prefetch(next); in __rcu_process_callbacks() 163 list = next; in __rcu_process_callbacks() 213 head->next = NULL; in __call_rcu() 217 rcp->curtail = &head->next; in __call_rcu()
|
D | update.c | 542 rhp->next = NULL; in call_rcu_tasks() 547 rcu_tasks_cbs_tail = &rhp->next; in call_rcu_tasks() 651 struct rcu_head *next; in rcu_tasks_kthread() local 779 next = list->next; in rcu_tasks_kthread() 783 list = next; in rcu_tasks_kthread()
|
/kernel/time/ |
D | tick-common.c | 102 ktime_t next = dev->next_event; in tick_handle_periodic() local 123 next = ktime_add(next, tick_period); in tick_handle_periodic() 125 if (!clockevents_program_event(dev, next, false)) in tick_handle_periodic() 157 ktime_t next; in tick_setup_periodic() local 161 next = tick_next_period; in tick_setup_periodic() 167 if (!clockevents_program_event(dev, next, false)) in tick_setup_periodic() 169 next = ktime_add(next, tick_period); in tick_setup_periodic()
|
D | ntp.c | 513 struct timespec64 next; in sync_cmos_clock() local 549 next.tv_nsec = (NSEC_PER_SEC / 2) - now.tv_nsec - (TICK_NSEC / 2); in sync_cmos_clock() 550 if (next.tv_nsec <= 0) in sync_cmos_clock() 551 next.tv_nsec += NSEC_PER_SEC; in sync_cmos_clock() 554 next.tv_sec = 659; in sync_cmos_clock() 556 next.tv_sec = 0; in sync_cmos_clock() 558 if (next.tv_nsec >= NSEC_PER_SEC) { in sync_cmos_clock() 559 next.tv_sec++; in sync_cmos_clock() 560 next.tv_nsec -= NSEC_PER_SEC; in sync_cmos_clock() 563 &sync_cmos_work, timespec64_to_jiffies(&next)); in sync_cmos_clock()
|
D | timer_stats.c | 57 struct entry *next; member 183 curr = curr->next; in tstat_lookup() 200 curr = curr->next; in tstat_lookup() 207 curr->next = NULL; in tstat_lookup() 213 prev->next = curr; in tstat_lookup()
|
/kernel/sched/ |
D | stats.h | 132 struct task_struct *prev, struct task_struct *next) in __sched_info_switch() argument 142 if (next != rq->idle) in __sched_info_switch() 143 sched_info_arrive(rq, next); in __sched_info_switch() 147 struct task_struct *prev, struct task_struct *next) in sched_info_switch() argument 150 __sched_info_switch(rq, prev, next); in sched_info_switch() 157 #define sched_info_arrive(rq, next) do { } while (0) argument 158 #define sched_info_switch(rq, t, next) do { } while (0) argument
|
D | rt.c | 87 rt_rq->highest_prio.next = MAX_RT_PRIO; in init_rt_rq() 377 if (p->prio < rq->rt.highest_prio.next) in enqueue_pushable_task() 378 rq->rt.highest_prio.next = p->prio; in enqueue_pushable_task() 389 rq->rt.highest_prio.next = p->prio; in dequeue_pushable_task() 391 rq->rt.highest_prio.next = MAX_RT_PRIO; in dequeue_pushable_task() 456 tg = list_entry_rcu(tg->list.next, in next_task_group() 657 goto next; in do_balance_runtime() 675 next: in do_balance_runtime() 1629 struct sched_rt_entity *next = NULL; in pick_next_rt_entity() local 1637 next = list_entry(queue->next, struct sched_rt_entity, run_list); in pick_next_rt_entity() [all …]
|
/kernel/events/ |
D | uprobes.c | 504 uc->next = uprobe->consumers; in consumer_add() 520 for (con = &uprobe->consumers; *con; con = &(*con)->next) { in consumer_del() 522 *con = uc->next; in consumer_del() 627 for (uc = uprobe->consumers; uc; uc = uc->next) { in filter_chain() 695 struct map_info *next; member 702 struct map_info *next = info->next; in free_map_info() local 704 return next; in free_map_info() 731 prev->next = NULL; in build_map_info() 742 prev = prev->next; in build_map_info() 743 info->next = curr; in build_map_info() [all …]
|
/kernel/trace/ |
D | ftrace.c | 152 while (likely(op = rcu_dereference_raw_notrace((op)->next)) && \ 179 ops != &ftrace_list_end; ops = ops->next) in ftrace_nr_registered_ops() 289 } else if (ftrace_ops_list->next == &ftrace_list_end) { in update_ftrace_function() 353 ops->next = *list; in add_ftrace_ops() 371 if (*list == ops && ops->next == &ftrace_list_end) { in remove_ftrace_ops() 376 for (p = list; *p != &ftrace_list_end; p = &(*p)->next) in remove_ftrace_ops() 383 *p = (*p)->next; in remove_ftrace_ops() 513 struct ftrace_profile_page *next; member 555 pg = pg->next; in function_stat_next() 694 pg = pg->next; in ftrace_profile_reset() [all …]
|
D | trace_sched_wakeup.c | 373 struct task_struct *next, in tracing_sched_switch_trace() argument 389 entry->next_pid = next->pid; in tracing_sched_switch_trace() 390 entry->next_prio = next->prio; in tracing_sched_switch_trace() 391 entry->next_state = next->state; in tracing_sched_switch_trace() 392 entry->next_cpu = task_cpu(next); in tracing_sched_switch_trace() 428 struct task_struct *prev, struct task_struct *next) in probe_wakeup_sched_switch() argument 451 if (next != wakeup_task) in probe_wakeup_sched_switch() 466 if (unlikely(!tracer_enabled || next != wakeup_task)) in probe_wakeup_sched_switch() 473 tracing_sched_switch_trace(wakeup_trace, prev, next, flags, pc); in probe_wakeup_sched_switch()
|
/kernel/gcov/ |
D | gcc_4_7.c | 90 struct gcov_info *next; member 128 return info->next; in gcov_info_next() 137 info->next = gcov_info_head; in gcov_info_link() 149 prev->next = info->next; in gcov_info_unlink() 151 gcov_info_head = info->next; in gcov_info_unlink() 271 dup->next = NULL; in gcov_info_dup()
|
D | gcc_3_4.c | 74 struct gcov_info *next; member 113 return info->next; in gcov_info_next() 122 info->next = gcov_info_head; in gcov_info_link() 134 prev->next = info->next; in gcov_info_unlink() 136 gcov_info_head = info->next; in gcov_info_unlink()
|