/kernel/ |
D | audit_tree.c | 19 struct list_head list; member 34 struct list_head list; member 103 INIT_LIST_HEAD(&tree->list); in alloc_tree() 202 INIT_LIST_HEAD(&chunk->owners[i].list); in alloc_chunk() 228 struct list_head *list; in insert_hash() local 237 list = chunk_hash(chunk->key); in insert_hash() 238 list_add_rcu(&chunk->hash, list); in insert_hash() 245 struct list_head *list = chunk_hash(key); in audit_tree_lookup() local 248 list_for_each_entry_rcu(p, list, hash) { in audit_tree_lookup() 313 list_replace_init(&old->owners[j].list, &new->owners[i].list); in replace_chunk() [all …]
|
D | kprobes.c | 85 struct list_head list; member 142 list_for_each_entry_rcu(kip, &c->pages, list) { in __get_insn_slot() 180 INIT_LIST_HEAD(&kip->list); in __get_insn_slot() 186 list_add_rcu(&kip->list, &c->pages); in __get_insn_slot() 209 if (!list_is_singular(&kip->list)) { in collect_one_slot() 217 list_del_rcu(&kip->list); in collect_one_slot() 234 list_for_each_entry_safe(kip, next, &c->pages, list) { in collect_garbage_slots() 256 list_for_each_entry_rcu(kip, &c->pages, list) { in __free_insn_slot() 294 list_for_each_entry_rcu(kip, &c->pages, list) { in __is_insn_slot_addr() 313 list_for_each_entry_rcu(kip, &c->pages, list) { in kprobe_cache_get_kallsym() [all …]
|
D | auditfilter.c | 165 int __init audit_register_class(int class, unsigned *list) in audit_register_class() argument 170 while (*list != ~0U) { in audit_register_class() 171 unsigned n = *list++; in audit_register_class() 895 struct list_head *list; in audit_find_rule() local 900 *p = list = &audit_inode_hash[h]; in audit_find_rule() 904 list = &audit_inode_hash[h]; in audit_find_rule() 905 list_for_each_entry(e, list, list) in audit_find_rule() 913 *p = list = &audit_filter_list[entry->rule.listnr]; in audit_find_rule() 916 list_for_each_entry(e, list, list) in audit_find_rule() 935 struct list_head *list; in audit_add_rule() local [all …]
|
D | taskstats.c | 48 struct list_head list; member 55 struct list_head list; member 122 list_for_each_entry(s, &listeners->list, list) { in send_cpu_listeners() 124 if (!list_is_last(&s->list, &listeners->list)) { in send_cpu_listeners() 146 list_for_each_entry_safe(s, tmp, &listeners->list, list) { in send_cpu_listeners() 148 list_del(&s->list); in send_cpu_listeners() 303 list_for_each_entry(s2, &listeners->list, list) { in add_del_listener() 307 list_add(&s->list, &listeners->list); in add_del_listener() 321 list_for_each_entry_safe(s, tmp, &listeners->list, list) { in add_del_listener() 323 list_del(&s->list); in add_del_listener() [all …]
|
D | padata.c | 250 if (list_empty(&reorder->list)) { in padata_find_next() 255 padata = list_entry(reorder->list.next, struct padata_priv, list); in padata_find_next() 267 list_del_init(&padata->list); in padata_find_next() 312 list_add_tail(&padata->list, &squeue->serial.list); in padata_reorder() 331 if (!list_empty(&reorder->list) && padata_find_next(pd, false)) in padata_reorder() 357 list_replace_init(&squeue->serial.list, &local_list); in padata_serial_worker() 366 struct padata_priv, list); in padata_serial_worker() 368 list_del_init(&padata->list); in padata_serial_worker() 397 list_for_each_prev(pos, &reorder->list) { in padata_do_serial() 398 cur = list_entry(pos, struct padata_priv, list); in padata_do_serial() [all …]
|
D | fail_function.c | 27 struct list_head list; member 70 INIT_LIST_HEAD(&attr->list); in fei_attr_new() 87 list_for_each_entry(attr, &fei_attr_list, list) { in fei_attr_lookup() 99 list_for_each_entry(attr, &fei_attr_list, list) { in fei_attr_is_valid() 201 struct fei_attr *attr = list_entry(v, struct fei_attr, list); in fei_seq_show() 223 list_del(&attr->list); in fei_attr_remove() 231 list_for_each_entry_safe(attr, n, &fei_attr_list, list) { in fei_attr_remove_all() 303 list_add_tail(&attr->list, &fei_attr_list); in fei_write()
|
D | audit_watch.c | 280 list_del_rcu(&oentry->list); in audit_update_watch() 284 list_del(&oentry->rule.list); in audit_update_watch() 298 list_add_rcu(&nentry->list, &audit_inode_hash[h]); in audit_update_watch() 299 list_replace(&oentry->rule.list, in audit_update_watch() 300 &nentry->rule.list); in audit_update_watch() 335 list_del(&r->list); in audit_remove_parent_watches() 336 list_del_rcu(&e->list); in audit_remove_parent_watches() 399 int audit_add_watch(struct audit_krule *krule, struct list_head **list) in audit_add_watch() argument 439 *list = &audit_inode_hash[h]; in audit_add_watch()
|
/kernel/bpf/ |
D | bpf_lru_list.c | 77 if (&node->list == l->next_inactive_rotation) in __bpf_lru_node_move_to_free() 83 list_move(&node->list, free_list); in __bpf_lru_node_move_to_free() 98 list_move(&node->list, &l->lists[tgt_type]); in __bpf_lru_node_move_in() 123 if (&node->list == l->next_inactive_rotation) in __bpf_lru_node_move() 126 list_move(&node->list, &l->lists[tgt_type]); in __bpf_lru_node_move() 151 first_node = list_first_entry(active, struct bpf_lru_node, list); in __bpf_lru_list_rotate_active() 152 list_for_each_entry_safe_reverse(node, tmp_node, active, list) { in __bpf_lru_list_rotate_active() 193 node = list_entry(cur, struct bpf_lru_node, list); in __bpf_lru_list_rotate_inactive() 222 list_for_each_entry_safe_reverse(node, tmp_node, inactive, list) { in __bpf_lru_list_shrink_inactive() 283 list) { in __bpf_lru_list_shrink() [all …]
|
D | bpf_local_storage.c | 117 &local_storage->list); in bpf_selem_unlink_storage_nolock() 174 hlist_add_head_rcu(&selem->snode, &local_storage->list); in bpf_selem_link_storage_nolock() 203 hlist_add_head_rcu(&selem->map_node, &b->list); in bpf_selem_link_map() 231 hlist_for_each_entry_rcu(selem, &local_storage->list, snode) in bpf_local_storage_lookup() 290 INIT_HLIST_HEAD(&storage->list); in bpf_local_storage_alloc() 356 if (!local_storage || hlist_empty(&local_storage->list)) { in bpf_local_storage_update() 396 if (unlikely(hlist_empty(&local_storage->list))) { in bpf_local_storage_update() 515 rcu_dereference_raw(hlist_first_rcu(&b->list)), in bpf_local_storage_map_free() 592 INIT_HLIST_HEAD(&smap->buckets[i].list); in bpf_local_storage_map_alloc()
|
D | cpumap.c | 189 list_for_each_entry_safe(skb, tmp, listp, list) { in cpu_map_bpf_prog_run_skb() 283 struct list_head *list) in cpu_map_bpf_prog_run() argument 297 if (unlikely(!list_empty(list))) in cpu_map_bpf_prog_run() 298 cpu_map_bpf_prog_run_skb(rcpu, list, stats); in cpu_map_bpf_prog_run() 324 LIST_HEAD(list); in cpu_map_kthread_run() 355 list_add_tail(&skb->list, &list); in cpu_map_kthread_run() 370 nframes = cpu_map_bpf_prog_run(rcpu, frames, xdp_n, &stats, &list); in cpu_map_kthread_run() 392 list_add_tail(&skb->list, &list); in cpu_map_kthread_run() 394 netif_receive_skb_list(&list); in cpu_map_kthread_run()
|
/kernel/trace/ |
D | trace_dynevent.h | 41 struct list_head list; member 62 struct list_head list; member 74 INIT_LIST_HEAD(&ev->list); in dyn_event_init() 88 list_add_tail(&ev->list, &dyn_event_list); in dyn_event_add() 95 list_del_init(&ev->list); in dyn_event_remove() 112 list_for_each_entry(pos, &dyn_event_list, list) 120 list_for_each_entry_safe(pos, n, &dyn_event_list, list)
|
D | trace_events_trigger.c | 67 list_for_each_entry_rcu(data, &file->triggers, list) { in event_triggers_call() 105 list_for_each_entry_rcu(data, &file->triggers, list) { in event_triggers_post_call() 131 list_for_each_entry_rcu(data, &file->triggers, list, in check_user_trigger() 170 list_for_each_entry_reverse(p, &trigger_commands, list) in trigger_show() 177 data = list_entry(v, struct event_trigger_data, list); in trigger_show() 212 list_for_each_entry(p, &trigger_commands, list) { in event_trigger_regex_open() 247 list_for_each_entry(p, &trigger_commands, list) { in trigger_process_regex() 349 list_for_each_entry(p, &trigger_commands, list) { in register_event_command() 355 list_add(&cmd->list, &trigger_commands); in register_event_command() 372 list_for_each_entry_safe(p, n, &trigger_commands, list) { in unregister_event_command() [all …]
|
D | ring_buffer.c | 295 struct list_head list; /* list of buffer pages */ member 687 static struct list_head *rb_list_head(struct list_head *list); 709 next = rb_list_head(page->list.next); in verify_event() 710 page = list_entry(next, struct buffer_page, list); in verify_event() 1211 static struct list_head *rb_list_head(struct list_head *list) in rb_list_head() argument 1213 unsigned long val = (unsigned long)list; in rb_list_head() 1227 rb_is_head_page(struct buffer_page *page, struct list_head *list) in rb_is_head_page() argument 1231 val = (unsigned long)list->next; in rb_is_head_page() 1233 if ((val & ~RB_FLAG_MASK) != (unsigned long)&page->list) in rb_is_head_page() 1248 struct list_head *list = page->list.prev; in rb_is_reader_page() local [all …]
|
D | trace_printk.c | 34 struct list_head list; member 45 list_for_each_entry(pos, &trace_bprintk_fmt_list, list) { in lookup_format() 76 list_add_tail(&tb_fmt->list, &trace_bprintk_fmt_list); in hold_module_trace_bprintk_format() 140 list_for_each_entry(p, &trace_bprintk_fmt_list, list) { in find_next_mod_format() 154 if (mod_fmt->list.next == &trace_bprintk_fmt_list) in find_next_mod_format() 157 mod_fmt = container_of(mod_fmt->list.next, typeof(*mod_fmt), list); in find_next_mod_format()
|
D | trace_event_perf.c | 99 struct hlist_head __percpu *list; in perf_trace_event_reg() local 107 list = alloc_percpu(struct hlist_head); in perf_trace_event_reg() 108 if (!list) in perf_trace_event_reg() 112 INIT_HLIST_HEAD(per_cpu_ptr(list, cpu)); in perf_trace_event_reg() 114 tp_event->perf_events = list; in perf_trace_event_reg() 222 list_for_each_entry(tp_event, &ftrace_events, list) { in perf_trace_init() 370 struct hlist_head *list; in perf_trace_add() local 376 list = this_cpu_ptr(pcpu_list); in perf_trace_add() 377 hlist_add_head_rcu(&p_event->hlist_entry, list); in perf_trace_add()
|
D | trace_probe.h | 249 struct list_head list; member 258 struct list_head list; member 311 return list_first_entry_or_null(&tpe->probes, struct trace_probe, list); in trace_probe_primary_from_call() 321 struct list_head *list = trace_probe_probe_list(tp); in trace_probe_has_sibling() local 323 return !list_empty(list) && !list_is_singular(list); in trace_probe_has_sibling() 354 list_for_each_entry(pos, &(tp)->event->files, list) 356 list_for_each_entry_rcu(pos, &(tp)->event->files, list)
|
/kernel/power/ |
D | qos.c | 61 if (plist_head_empty(&c->list)) in pm_qos_get_value() 66 return plist_first(&c->list)->prio; in pm_qos_get_value() 69 return plist_last(&c->list)->prio; in pm_qos_get_value() 115 plist_del(node, &c->list); in pm_qos_update_target() 122 plist_del(node, &c->list); in pm_qos_update_target() 126 plist_add(node, &c->list); in pm_qos_update_target() 160 list_for_each_entry(req, &pqf->list, node) in pm_qos_flags_remove_req() 184 prev_value = list_empty(&pqf->list) ? 0 : pqf->effective_flags; in pm_qos_update_flags() 196 list_add_tail(&req->node, &pqf->list); in pm_qos_update_flags() 204 curr_value = list_empty(&pqf->list) ? 0 : pqf->effective_flags; in pm_qos_update_flags() [all …]
|
D | snapshot.c | 268 static inline void free_list_of_pages(struct linked_page *list, in free_list_of_pages() argument 271 while (list) { in free_list_of_pages() 272 struct linked_page *lp = list->next; in free_list_of_pages() 274 free_image_page(list, clear_page_nosave); in free_list_of_pages() 275 list = lp; in free_list_of_pages() 382 struct list_head list; member 391 struct list_head list; /* Link Zones together */ member 437 struct list_head *list) in alloc_rtree_node() argument 449 list_add_tail(&node->list, list); in alloc_rtree_node() 574 list_for_each_entry(node, &zone->nodes, list) in free_zone_bm_rtree() [all …]
|
/kernel/rcu/ |
D | tiny.c | 109 struct rcu_head *next, *list; in rcu_process_callbacks() local 119 list = rcu_ctrlblk.rcucblist; in rcu_process_callbacks() 128 while (list) { in rcu_process_callbacks() 129 next = list->next; in rcu_process_callbacks() 131 debug_rcu_head_unqueue(list); in rcu_process_callbacks() 133 rcu_reclaim_tiny(list); in rcu_process_callbacks() 135 list = next; in rcu_process_callbacks()
|
/kernel/time/ |
D | clockevents.c | 349 struct clock_event_device, list); in clockevents_notify_released() 350 list_move(&dev->list, &clockevent_devices); in clockevents_notify_released() 362 list_for_each_entry(dev, &clockevent_devices, list) { in clockevents_replace() 378 list_del_init(&ced->list); in clockevents_replace() 390 list_del_init(&ced->list); in __clockevents_try_unbind() 463 list_add(&dev->list, &clockevent_devices); in clockevents_register_device() 578 list_move(&old->list, &clockevents_released); in clockevents_exchange_device() 594 list_for_each_entry_reverse(dev, &clockevent_devices, list) in clockevents_suspend() 606 list_for_each_entry(dev, &clockevent_devices, list) in clockevents_resume() 644 list_for_each_entry_safe(dev, tmp, &clockevents_released, list) in tick_cleanup_dead_cpu() [all …]
|
D | clocksource.c | 171 if (list_empty(&cs->list)) { in __clocksource_unstable() 197 if (!list_empty(&cs->list) && list_empty(&cs->wd_list)) in clocksource_mark_unstable() 615 list_for_each_entry(cs, &clocksource_list, list) { in clocksource_select_watchdog() 758 list_for_each_entry(cs, &clocksource_list, list) { in clocksource_suspend_select() 859 list_for_each_entry_reverse(cs, &clocksource_list, list) in clocksource_suspend() 871 list_for_each_entry(cs, &clocksource_list, list) in clocksource_resume() 975 list_for_each_entry(cs, &clocksource_list, list) { in clocksource_find_best() 999 list_for_each_entry(cs, &clocksource_list, list) { in __clocksource_select() 1084 list_for_each_entry(tmp, &clocksource_list, list) { in clocksource_enqueue() 1088 entry = &tmp->list; in clocksource_enqueue() [all …]
|
/kernel/locking/ |
D | mutex-debug.c | 32 INIT_LIST_HEAD(&waiter->list); in debug_mutex_lock_common() 41 DEBUG_LOCKS_WARN_ON(list_empty(&waiter->list)); in debug_mutex_wake_waiter() 46 DEBUG_LOCKS_WARN_ON(!list_empty(&waiter->list)); in debug_mutex_free_waiter() 62 DEBUG_LOCKS_WARN_ON(list_empty(&waiter->list)); in debug_mutex_remove_waiter() 67 INIT_LIST_HEAD(&waiter->list); in debug_mutex_remove_waiter()
|
D | ww_mutex.h | 13 w = list_first_entry(&lock->wait_list, struct mutex_waiter, list); in __ww_waiter_first() 14 if (list_entry_is_head(w, &lock->wait_list, list)) in __ww_waiter_first() 23 w = list_next_entry(w, list); in __ww_waiter_next() 24 if (list_entry_is_head(w, &lock->wait_list, list)) in __ww_waiter_next() 33 w = list_prev_entry(w, list); in __ww_waiter_prev() 34 if (list_entry_is_head(w, &lock->wait_list, list)) in __ww_waiter_prev() 45 w = list_last_entry(&lock->wait_list, struct mutex_waiter, list); in __ww_waiter_last() 46 if (list_entry_is_head(w, &lock->wait_list, list)) in __ww_waiter_last() 57 p = &pos->list; in __ww_waiter_add()
|
D | semaphore.c | 198 struct list_head list; member 213 list_add_tail(&waiter.list, &sem->wait_list); in __down_common() 231 list_del(&waiter.list); in __down_common() 235 list_del(&waiter.list); in __down_common() 262 struct semaphore_waiter, list); in __up() 263 list_del(&waiter->list); in __up()
|
/kernel/dma/ |
D | debug.c | 67 struct list_head list; member 87 struct list_head list; member 299 list_for_each_entry(entry, &bucket->list, list) { in __hash_bucket_find() 381 list_add_tail(&entry->list, &bucket->list); in hash_bucket_add() 389 list_del(&entry->list); in hash_bucket_del() 414 list_for_each_entry(entry, &bucket->list, list) { in debug_dma_dump_mappings() 585 list_add_tail(&entry[i].list, &free_entries); in dma_debug_create_entries() 597 entry = list_entry(free_entries.next, struct dma_debug_entry, list); in __dma_entry_alloc() 598 list_del(&entry->list); in __dma_entry_alloc() 674 list_add(&entry->list, &free_entries); in dma_entry_free() [all …]
|