Home
last modified time | relevance | path

Searched refs:list (Results 1 – 25 of 106) sorted by relevance

12345

/kernel/
Daudit_tree.c19 struct list_head list; member
34 struct list_head list; member
103 INIT_LIST_HEAD(&tree->list); in alloc_tree()
202 INIT_LIST_HEAD(&chunk->owners[i].list); in alloc_chunk()
228 struct list_head *list; in insert_hash() local
237 list = chunk_hash(chunk->key); in insert_hash()
238 list_add_rcu(&chunk->hash, list); in insert_hash()
245 struct list_head *list = chunk_hash(key); in audit_tree_lookup() local
248 list_for_each_entry_rcu(p, list, hash) { in audit_tree_lookup()
313 list_replace_init(&old->owners[j].list, &new->owners[i].list); in replace_chunk()
[all …]
Dkprobes.c90 struct list_head list; member
153 list_for_each_entry_rcu(kip, &c->pages, list) { in __get_insn_slot()
187 INIT_LIST_HEAD(&kip->list); in __get_insn_slot()
193 list_add_rcu(&kip->list, &c->pages); in __get_insn_slot()
216 if (!list_is_singular(&kip->list)) { in collect_one_slot()
224 list_del_rcu(&kip->list); in collect_one_slot()
241 list_for_each_entry_safe(kip, next, &c->pages, list) { in collect_garbage_slots()
264 list_for_each_entry_rcu(kip, &c->pages, list) { in __free_insn_slot()
302 list_for_each_entry_rcu(kip, &c->pages, list) { in __is_insn_slot_addr()
321 list_for_each_entry_rcu(kip, &c->pages, list) { in kprobe_cache_get_kallsym()
[all …]
Dauditfilter.c168 int __init audit_register_class(int class, unsigned *list) in audit_register_class() argument
173 while (*list != ~0U) { in audit_register_class()
174 unsigned n = *list++; in audit_register_class()
903 struct list_head *list; in audit_find_rule() local
908 *p = list = &audit_inode_hash[h]; in audit_find_rule()
912 list = &audit_inode_hash[h]; in audit_find_rule()
913 list_for_each_entry(e, list, list) in audit_find_rule()
921 *p = list = &audit_filter_list[entry->rule.listnr]; in audit_find_rule()
924 list_for_each_entry(e, list, list) in audit_find_rule()
943 struct list_head *list; in audit_add_rule() local
[all …]
Dtaskstats.c49 struct list_head list; member
56 struct list_head list; member
122 list_for_each_entry(s, &listeners->list, list) { in send_cpu_listeners()
126 if (!list_is_last(&s->list, &listeners->list)) { in send_cpu_listeners()
148 list_for_each_entry_safe(s, tmp, &listeners->list, list) { in send_cpu_listeners()
150 list_del(&s->list); in send_cpu_listeners()
325 list_for_each_entry(s2, &listeners->list, list) { in add_del_listener()
329 list_add(&s->list, &listeners->list); in add_del_listener()
343 list_for_each_entry_safe(s, tmp, &listeners->list, list) { in add_del_listener()
345 list_del(&s->list); in add_del_listener()
[all …]
Dirq_work.c80 struct llist_head *list; in __irq_work_queue_local() local
93 list = this_cpu_ptr(&lazy_list); in __irq_work_queue_local()
95 list = this_cpu_ptr(&raised_list); in __irq_work_queue_local()
97 if (!llist_add(&work->node.llist, list)) in __irq_work_queue_local()
226 static void irq_work_run_list(struct llist_head *list) in irq_work_run_list() argument
238 if (llist_empty(list)) in irq_work_run_list()
241 llnode = llist_del_all(list); in irq_work_run_list()
Dfail_function.c27 struct list_head list; member
70 INIT_LIST_HEAD(&attr->list); in fei_attr_new()
87 list_for_each_entry(attr, &fei_attr_list, list) { in fei_attr_lookup()
99 list_for_each_entry(attr, &fei_attr_list, list) { in fei_attr_is_valid()
201 struct fei_attr *attr = list_entry(v, struct fei_attr, list); in fei_seq_show()
223 list_del(&attr->list); in fei_attr_remove()
231 list_for_each_entry_safe(attr, n, &fei_attr_list, list) { in fei_attr_remove_all()
299 list_add_tail(&attr->list, &fei_attr_list); in fei_write()
Dpadata.c250 if (list_empty(&reorder->list)) { in padata_find_next()
255 padata = list_entry(reorder->list.next, struct padata_priv, list); in padata_find_next()
267 list_del_init(&padata->list); in padata_find_next()
312 list_add_tail(&padata->list, &squeue->serial.list); in padata_reorder()
331 if (!list_empty(&reorder->list) && padata_find_next(pd, false)) in padata_reorder()
357 list_replace_init(&squeue->serial.list, &local_list); in padata_serial_worker()
366 struct padata_priv, list); in padata_serial_worker()
368 list_del_init(&padata->list); in padata_serial_worker()
397 list_for_each_prev(pos, &reorder->list) { in padata_do_serial()
398 cur = list_entry(pos, struct padata_priv, list); in padata_do_serial()
[all …]
/kernel/bpf/
Dbpf_lru_list.c77 if (&node->list == l->next_inactive_rotation) in __bpf_lru_node_move_to_free()
83 list_move(&node->list, free_list); in __bpf_lru_node_move_to_free()
98 list_move(&node->list, &l->lists[tgt_type]); in __bpf_lru_node_move_in()
123 if (&node->list == l->next_inactive_rotation) in __bpf_lru_node_move()
126 list_move(&node->list, &l->lists[tgt_type]); in __bpf_lru_node_move()
151 first_node = list_first_entry(active, struct bpf_lru_node, list); in __bpf_lru_list_rotate_active()
152 list_for_each_entry_safe_reverse(node, tmp_node, active, list) { in __bpf_lru_list_rotate_active()
193 node = list_entry(cur, struct bpf_lru_node, list); in __bpf_lru_list_rotate_inactive()
222 list_for_each_entry_safe_reverse(node, tmp_node, inactive, list) { in __bpf_lru_list_shrink_inactive()
283 list) { in __bpf_lru_list_shrink()
[all …]
/kernel/trace/
Dtrace_dynevent.h41 struct list_head list; member
62 struct list_head list; member
74 INIT_LIST_HEAD(&ev->list); in dyn_event_init()
88 list_add_tail(&ev->list, &dyn_event_list); in dyn_event_add()
95 list_del_init(&ev->list); in dyn_event_remove()
112 list_for_each_entry(pos, &dyn_event_list, list)
120 list_for_each_entry_safe(pos, n, &dyn_event_list, list)
Dtrace_events_trigger.c67 list_for_each_entry_rcu(data, &file->triggers, list) { in event_triggers_call()
119 list_for_each_entry_rcu(data, &file->triggers, list) { in event_triggers_post_call()
145 list_for_each_entry_rcu(data, &file->triggers, list, in check_user_trigger()
184 list_for_each_entry_reverse(p, &trigger_commands, list) in trigger_show()
191 data = list_entry(v, struct event_trigger_data, list); in trigger_show()
226 list_for_each_entry(p, &trigger_commands, list) { in event_trigger_regex_open()
261 list_for_each_entry(p, &trigger_commands, list) { in trigger_process_regex()
363 list_for_each_entry(p, &trigger_commands, list) { in register_event_command()
369 list_add(&cmd->list, &trigger_commands); in register_event_command()
386 list_for_each_entry_safe(p, n, &trigger_commands, list) { in unregister_event_command()
[all …]
Dring_buffer.c303 struct list_head list; /* list of buffer pages */ member
701 static struct list_head *rb_list_head(struct list_head *list);
723 next = rb_list_head(page->list.next); in verify_event()
724 page = list_entry(next, struct buffer_page, list); in verify_event()
1245 static struct list_head *rb_list_head(struct list_head *list) in rb_list_head() argument
1247 unsigned long val = (unsigned long)list; in rb_list_head()
1261 rb_is_head_page(struct buffer_page *page, struct list_head *list) in rb_is_head_page() argument
1265 val = (unsigned long)list->next; in rb_is_head_page()
1267 if ((val & ~RB_FLAG_MASK) != (unsigned long)&page->list) in rb_is_head_page()
1282 struct list_head *list = page->list.prev; in rb_is_reader_page() local
[all …]
Dtrace_printk.c34 struct list_head list; member
45 list_for_each_entry(pos, &trace_bprintk_fmt_list, list) { in lookup_format()
76 list_add_tail(&tb_fmt->list, &trace_bprintk_fmt_list); in hold_module_trace_bprintk_format()
140 list_for_each_entry(p, &trace_bprintk_fmt_list, list) { in find_next_mod_format()
154 if (mod_fmt->list.next == &trace_bprintk_fmt_list) in find_next_mod_format()
157 mod_fmt = container_of(mod_fmt->list.next, typeof(*mod_fmt), list); in find_next_mod_format()
Dtrace_event_perf.c99 struct hlist_head __percpu *list; in perf_trace_event_reg() local
107 list = alloc_percpu(struct hlist_head); in perf_trace_event_reg()
108 if (!list) in perf_trace_event_reg()
112 INIT_HLIST_HEAD(per_cpu_ptr(list, cpu)); in perf_trace_event_reg()
114 tp_event->perf_events = list; in perf_trace_event_reg()
222 list_for_each_entry(tp_event, &ftrace_events, list) { in perf_trace_init()
370 struct hlist_head *list; in perf_trace_add() local
376 list = this_cpu_ptr(pcpu_list); in perf_trace_add()
377 hlist_add_head_rcu(&p_event->hlist_entry, list); in perf_trace_add()
Dtrace_probe.h248 struct list_head list; member
257 struct list_head list; member
310 return list_first_entry_or_null(&tpe->probes, struct trace_probe, list); in trace_probe_primary_from_call()
320 struct list_head *list = trace_probe_probe_list(tp); in trace_probe_has_sibling() local
322 return !list_empty(list) && !list_is_singular(list); in trace_probe_has_sibling()
353 list_for_each_entry(pos, &(tp)->event->files, list)
355 list_for_each_entry_rcu(pos, &(tp)->event->files, list)
/kernel/power/
Dqos.c61 if (plist_head_empty(&c->list)) in pm_qos_get_value()
66 return plist_first(&c->list)->prio; in pm_qos_get_value()
69 return plist_last(&c->list)->prio; in pm_qos_get_value()
115 plist_del(node, &c->list); in pm_qos_update_target()
122 plist_del(node, &c->list); in pm_qos_update_target()
126 plist_add(node, &c->list); in pm_qos_update_target()
160 list_for_each_entry(req, &pqf->list, node) in pm_qos_flags_remove_req()
184 prev_value = list_empty(&pqf->list) ? 0 : pqf->effective_flags; in pm_qos_update_flags()
196 list_add_tail(&req->node, &pqf->list); in pm_qos_update_flags()
204 curr_value = list_empty(&pqf->list) ? 0 : pqf->effective_flags; in pm_qos_update_flags()
[all …]
Dsnapshot.c268 static inline void free_list_of_pages(struct linked_page *list, in free_list_of_pages() argument
271 while (list) { in free_list_of_pages()
272 struct linked_page *lp = list->next; in free_list_of_pages()
274 free_image_page(list, clear_page_nosave); in free_list_of_pages()
275 list = lp; in free_list_of_pages()
382 struct list_head list; member
391 struct list_head list; /* Link Zones together */ member
442 struct list_head *list) in alloc_rtree_node() argument
454 list_add_tail(&node->list, list); in alloc_rtree_node()
579 list_for_each_entry(node, &zone->nodes, list) in free_zone_bm_rtree()
[all …]
/kernel/time/
Dclockevents.c349 struct clock_event_device, list); in clockevents_notify_released()
350 list_move(&dev->list, &clockevent_devices); in clockevents_notify_released()
362 list_for_each_entry(dev, &clockevent_devices, list) { in clockevents_replace()
378 list_del_init(&ced->list); in clockevents_replace()
390 list_del_init(&ced->list); in __clockevents_try_unbind()
463 list_add(&dev->list, &clockevent_devices); in clockevents_register_device()
578 list_move(&old->list, &clockevents_released); in clockevents_exchange_device()
594 list_for_each_entry_reverse(dev, &clockevent_devices, list) in clockevents_suspend()
606 list_for_each_entry(dev, &clockevent_devices, list) in clockevents_resume()
644 list_for_each_entry_safe(dev, tmp, &clockevents_released, list) in tick_cleanup_dead_cpu()
[all …]
Dclocksource.c177 if (list_empty(&cs->list)) { in __clocksource_unstable()
203 if (!list_empty(&cs->list) && list_empty(&cs->wd_list)) in clocksource_mark_unstable()
621 list_for_each_entry(cs, &clocksource_list, list) { in clocksource_select_watchdog()
764 list_for_each_entry(cs, &clocksource_list, list) { in clocksource_suspend_select()
865 list_for_each_entry_reverse(cs, &clocksource_list, list) in clocksource_suspend()
877 list_for_each_entry(cs, &clocksource_list, list) in clocksource_resume()
981 list_for_each_entry(cs, &clocksource_list, list) { in clocksource_find_best()
1005 list_for_each_entry(cs, &clocksource_list, list) { in __clocksource_select()
1090 list_for_each_entry(tmp, &clocksource_list, list) { in clocksource_enqueue()
1094 entry = &tmp->list; in clocksource_enqueue()
[all …]
/kernel/rcu/
Dtiny.c109 struct rcu_head *next, *list; in rcu_process_callbacks() local
119 list = rcu_ctrlblk.rcucblist; in rcu_process_callbacks()
128 while (list) { in rcu_process_callbacks()
129 next = list->next; in rcu_process_callbacks()
131 debug_rcu_head_unqueue(list); in rcu_process_callbacks()
133 rcu_reclaim_tiny(list); in rcu_process_callbacks()
135 list = next; in rcu_process_callbacks()
/kernel/locking/
Dmutex-debug.c32 INIT_LIST_HEAD(&waiter->list); in debug_mutex_lock_common()
41 DEBUG_LOCKS_WARN_ON(list_empty(&waiter->list)); in debug_mutex_wake_waiter()
46 DEBUG_LOCKS_WARN_ON(!list_empty(&waiter->list)); in debug_mutex_free_waiter()
62 DEBUG_LOCKS_WARN_ON(list_empty(&waiter->list)); in debug_mutex_remove_waiter()
67 INIT_LIST_HEAD(&waiter->list); in debug_mutex_remove_waiter()
Dww_mutex.h13 w = list_first_entry(&lock->wait_list, struct mutex_waiter, list); in __ww_waiter_first()
14 if (list_entry_is_head(w, &lock->wait_list, list)) in __ww_waiter_first()
23 w = list_next_entry(w, list); in __ww_waiter_next()
24 if (list_entry_is_head(w, &lock->wait_list, list)) in __ww_waiter_next()
33 w = list_prev_entry(w, list); in __ww_waiter_prev()
34 if (list_entry_is_head(w, &lock->wait_list, list)) in __ww_waiter_prev()
45 w = list_last_entry(&lock->wait_list, struct mutex_waiter, list); in __ww_waiter_last()
46 if (list_entry_is_head(w, &lock->wait_list, list)) in __ww_waiter_last()
57 p = &pos->list; in __ww_waiter_add()
Dsemaphore.c199 struct list_head list; member
214 list_add_tail(&waiter.list, &sem->wait_list); in ___down_common()
232 list_del(&waiter.list); in ___down_common()
236 list_del(&waiter.list); in ___down_common()
275 struct semaphore_waiter, list); in __up()
276 list_del(&waiter->list); in __up()
/kernel/futex/
Dcore.c444 plist_for_each_entry(this, &hb->chain, list) { in futex_top_waiter()
514 if (WARN_ON_SMP(!q->lock_ptr) || WARN_ON(plist_node_empty(&q->list))) in __futex_unqueue()
519 plist_del(&q->list, &hb->chain); in __futex_unqueue()
569 plist_node_init(&q->list, prio); in __futex_queue()
570 trace_android_vh_alter_futex_plist_add(&q->list, &hb->chain, &already_on_hb); in __futex_queue()
572 plist_add(&q->list, &hb->chain); in __futex_queue()
800 if (fetch_robust_entry(&entry, &head->list.next, &pi)) in exit_robust_list()
815 while (entry != &head->list) { in exit_robust_list()
895 if (compat_fetch_robust_entry(&uentry, &entry, &head->list.next, &pi)) in compat_exit_robust_list()
911 while (entry != (struct robust_list __user *) &head->list) { in compat_exit_robust_list()
[all …]
/kernel/trace/rv/
Drv.c418 list_for_each_entry_continue(m_def, &rv_monitors_list, list) { in enabled_monitors_next()
436 m_def = list_entry(&rv_monitors_list, struct rv_monitor_def, list); in enabled_monitors_start()
489 list_for_each_entry(mdef, &rv_monitors_list, list) in disable_all_monitors()
546 list_for_each_entry(mdef, &rv_monitors_list, list) { in enabled_monitors_write()
617 list_for_each_entry(mdef, &rv_monitors_list, list) { in reset_all_monitors()
707 list_for_each_entry(r, &rv_monitors_list, list) { in rv_register_monitor()
729 list_add_tail(&r->list, &rv_monitors_list); in rv_register_monitor()
748 list_for_each_entry_safe(ptr, next, &rv_monitors_list, list) { in rv_unregister_monitor()
751 list_del(&ptr->list); in rv_unregister_monitor()
/kernel/dma/
Ddebug.c67 struct list_head list; member
87 struct list_head list; member
299 list_for_each_entry(entry, &bucket->list, list) { in __hash_bucket_find()
379 list_add_tail(&entry->list, &bucket->list); in hash_bucket_add()
387 list_del(&entry->list); in hash_bucket_del()
412 list_for_each_entry(entry, &bucket->list, list) { in debug_dma_dump_mappings()
583 list_add_tail(&entry[i].list, &free_entries); in dma_debug_create_entries()
595 entry = list_entry(free_entries.next, struct dma_debug_entry, list); in __dma_entry_alloc()
596 list_del(&entry->list); in __dma_entry_alloc()
672 list_add(&entry->list, &free_entries); in dma_entry_free()
[all …]

12345