/mm/ |
D | mmu_notifier.c | 67 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) in __mmu_notifier_release() 78 while (unlikely(!hlist_empty(&mm->mmu_notifier_mm->list))) { in __mmu_notifier_release() 79 mn = hlist_entry(mm->mmu_notifier_mm->list.first, in __mmu_notifier_release() 118 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_clear_flush_young() 135 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_clear_young() 151 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_test_young() 170 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_change_pte() 184 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_invalidate_range_start() 199 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_invalidate_range_end() 224 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_invalidate_range() [all …]
|
D | slab.c | 212 int node, struct list_head *list); 213 static void slabs_destroy(struct kmem_cache *cachep, struct list_head *list); 218 void **list); 221 void **list); 598 LIST_HEAD(list); in cache_free_pfmemalloc() 604 free_block(cachep, &objp, 1, page_node, &list); in cache_free_pfmemalloc() 607 slabs_destroy(cachep, &list); in cache_free_pfmemalloc() 729 struct list_head *list) in __drain_alien_cache() argument 743 free_block(cachep, ac->entry, ac->avail, node, list); in __drain_alien_cache() 763 LIST_HEAD(list); in reap_alien() [all …]
|
D | list_lru.c | 22 list_add(&lru->list, &list_lrus); in list_lru_register() 29 list_del(&lru->list); in list_lru_unregister() 114 list_add_tail(item, &l->list); in list_lru_add() 145 void list_lru_isolate(struct list_lru_one *list, struct list_head *item) in list_lru_isolate() argument 148 list->nr_items--; in list_lru_isolate() 152 void list_lru_isolate_move(struct list_lru_one *list, struct list_head *item, in list_lru_isolate_move() argument 156 list->nr_items--; in list_lru_isolate_move() 205 list_for_each_safe(item, n, &l->list) { in __list_lru_walk_one() 232 list_move_tail(item, &l->list); in __list_lru_walk_one() 285 INIT_LIST_HEAD(&l->list); in init_one_lru() [all …]
|
D | kmemleak-test.c | 36 struct list_head list; member 84 INIT_LIST_HEAD(&elem->list); in kmemleak_test_init() 85 list_add_tail(&elem->list, &test_list); in kmemleak_test_init() 106 list_for_each_entry_safe(elem, tmp, &test_list, list) in kmemleak_test_exit() 107 list_del(&elem->list); in kmemleak_test_exit()
|
D | zpool.c | 25 struct list_head list; member 42 list_add(&driver->list, &drivers_head); in zpool_register_driver() 67 list_del(&driver->list); in zpool_unregister_driver() 80 list_for_each_entry(driver, &drivers_head, list) { in zpool_get_driver() 194 list_add(&zpool->list, &pools_head); in zpool_create_pool() 216 list_del(&zpool->list); in zpool_destroy_pool()
|
D | swapfile.c | 155 list_for_each_entry(se, &si->first_swap_extent.list, list) { in discard_swap() 201 se = list_next_entry(se, list); in discard_swap_cluster() 321 static inline bool cluster_list_empty(struct swap_cluster_list *list) in cluster_list_empty() argument 323 return cluster_is_null(&list->head); in cluster_list_empty() 326 static inline unsigned int cluster_list_first(struct swap_cluster_list *list) in cluster_list_first() argument 328 return cluster_next(&list->head); in cluster_list_first() 331 static void cluster_list_init(struct swap_cluster_list *list) in cluster_list_init() argument 333 cluster_set_null(&list->head); in cluster_list_init() 334 cluster_set_null(&list->tail); in cluster_list_init() 337 static void cluster_list_add_tail(struct swap_cluster_list *list, in cluster_list_add_tail() argument [all …]
|
D | zswap.c | 127 struct list_head list; member 222 list_for_each_entry_rcu(pool, &zswap_pools, list) in zswap_update_total_size() 430 pool = list_first_or_null_rcu(&zswap_pools, typeof(*pool), list); in __zswap_pool_current() 465 list_for_each_entry_rcu(pool, &zswap_pools, list) in zswap_pool_last_get() 484 list_for_each_entry_rcu(pool, &zswap_pools, list) { in zswap_pool_find_get() 547 INIT_LIST_HEAD(&pool->list); in zswap_pool_create() 642 list_del_rcu(&pool->list); in __zswap_pool_empty() 705 list_del_rcu(&pool->list); in __zswap_param_set() 722 list_add_rcu(&pool->list, &zswap_pools); in __zswap_param_set() 729 list_add_tail_rcu(&pool->list, &zswap_pools); in __zswap_param_set() [all …]
|
D | highmem.c | 385 struct list_head list; member 424 list_for_each_entry(pam, &pas->lh, list) { in page_address() 458 list_add_tail(&pam->list, &pas->lh); in set_page_address() 462 list_for_each_entry(pam, &pas->lh, list) { in set_page_address() 464 list_del(&pam->list); in set_page_address()
|
D | slab_common.c | 86 list_for_each_entry(s, &slab_caches, list) { in kmem_cache_sanity_check() 395 list_add(&s->list, &slab_caches); in create_cache() 528 list_for_each_entry_safe(s, s2, &to_destroy, list) { in slab_caches_to_rcu_destroy_workfn() 546 list_del(&s->list); in shutdown_cache() 552 list_add_tail(&s->list, &slab_caches_to_rcu_destroy); in shutdown_cache() 914 list_add(&s->list, &slab_caches); in create_kmalloc_cache() 1146 static void freelist_randomize(struct rnd_state *state, unsigned int *list, in freelist_randomize() argument 1153 list[i] = i; in freelist_randomize() 1159 swap(list[i], list[rand]); in freelist_randomize()
|
D | bootmem.c | 80 list_for_each_entry(ent, &bdata_list, list) { in link_bootmem() 82 list_add_tail(&bdata->list, &ent->list); in link_bootmem() 87 list_add_tail(&bdata->list, &bdata_list); in link_bootmem() 279 list_for_each_entry(bdata, &bdata_list, list) in free_all_bootmem() 363 list_for_each_entry(bdata, &bdata_list, list) { in mark_bootmem() 616 list_for_each_entry(bdata, &bdata_list, list) { in alloc_bootmem_core()
|
D | frontswap.c | 125 plist_for_each_entry(si, &swap_active_head, list) { in frontswap_register_ops() 147 plist_for_each_entry(si, &swap_active_head, list) { in frontswap_register_ops() 378 plist_for_each_entry(si, &swap_active_head, list) in __frontswap_curr_pages() 393 plist_for_each_entry(si, &swap_active_head, list) { in __frontswap_unuse_pages()
|
D | vmalloc.c | 42 struct llist_head list; member 54 llist_for_each_safe(llnode, t, llist_del_all(&p->list)) in free_work() 389 list_add_rcu(&va->list, &prev->list); in __insert_vmap_area() 391 list_add_rcu(&va->list, &vmap_area_list); in __insert_vmap_area() 494 if (list_is_last(&first->list, &vmap_area_list)) in alloc_vmap_area() 497 first = list_next_entry(first, list); in alloc_vmap_area() 578 list_del_rcu(&va->list); in __free_vmap_area() 1272 init_llist_head(&p->list); in vmalloc_init() 1564 if (llist_add((struct llist_node *)addr, &p->list)) in __vfree_deferred() 2124 list_for_each_entry(va, &vmap_area_list, list) { in vread() [all …]
|
D | zsmalloc.c | 289 struct list_head list; /* fullness list */ member 764 struct zspage, list); in insert_zspage() 771 list_add(&zspage->list, &head->list); in insert_zspage() 775 list_add(&zspage->list, &class->fullness_list[fullness]); in insert_zspage() 789 list_del_init(&zspage->list); in remove_zspage() 1019 VM_BUG_ON(list_empty(&zspage->list)); in free_zspage() 1156 struct zspage, list); in find_get_zspage() 1805 struct zspage, list); in isolate_zspage() 1989 if (list_empty(&zspage->list) && !is_zspage_isolated(zspage)) { in zs_page_isolate() 1998 if (!list_empty(&zspage->list) && !is_zspage_isolated(zspage)) { in zs_page_isolate() [all …]
|
D | hmm.c | 142 list_for_each_entry(range, &hmm->ranges, list) { in hmm_invalidate_range() 157 list_for_each_entry(mirror, &hmm->mirrors, list) in hmm_invalidate_range() 214 list_add(&mirror->list, &mirror->hmm->mirrors); in hmm_mirror_register() 233 list_del(&mirror->list); in hmm_mirror_unregister() 532 list_add_rcu(&range->list, &hmm->ranges); in hmm_vma_get_pfns() 608 list_del_rcu(&range->list); in hmm_vma_range_done() 698 list_add_rcu(&range->list, &hmm->ranges); in hmm_vma_fault()
|
D | percpu.c | 506 list_move(&chunk->list, &pcpu_slot[nslot]); in pcpu_chunk_relocate() 508 list_move_tail(&chunk->list, &pcpu_slot[nslot]); in pcpu_chunk_relocate() 1105 INIT_LIST_HEAD(&chunk->list); in pcpu_alloc_first_chunk() 1171 INIT_LIST_HEAD(&chunk->list); in pcpu_alloc_chunk() 1405 list_for_each_entry(chunk, &pcpu_slot[slot], list) { in pcpu_alloc() 1594 list_for_each_entry_safe(chunk, next, free_head, list) { in pcpu_balance_workfn() 1598 if (chunk == list_first_entry(free_head, struct pcpu_chunk, list)) in pcpu_balance_workfn() 1601 list_move(&chunk->list, &to_free); in pcpu_balance_workfn() 1606 list_for_each_entry_safe(chunk, next, &to_free, list) { in pcpu_balance_workfn() 1647 list_for_each_entry(chunk, &pcpu_slot[slot], list) { in pcpu_balance_workfn() [all …]
|
D | ksm.c | 146 struct list_head list; member 646 list_del(&stable_node->list); in remove_node_from_stable_tree() 920 list_for_each_entry_safe(stable_node, next, &migrate_nodes, list) { in remove_all_stable_nodes() 1667 list_del(&page_node->list); in stable_tree_search() 1693 list_del(&page_node->list); in stable_tree_search() 1711 list_del(&page_node->list); in stable_tree_search() 1723 list_add(&stable_node_dup->list, stable_node_dup->head); in stable_tree_search() 1756 list_del(&page_node->list); in stable_tree_search() 2018 list_add(&stable_node->list, stable_node->head); in cmp_and_merge_page() 2214 &migrate_nodes, list) { in scan_get_next_rmap_item() [all …]
|
D | percpu-stats.c | 41 list_for_each_entry(chunk, &pcpu_slot[slot], list) in find_max_nr_alloc() 206 list_for_each_entry(chunk, &pcpu_slot[slot], list) { in percpu_stats_show()
|
D | swap.c | 845 struct lruvec *lruvec, struct list_head *list) in lru_add_page_tail() argument 855 if (!list) in lru_add_page_tail() 860 else if (list) { in lru_add_page_tail() 863 list_add_tail(&page_tail->lru, list); in lru_add_page_tail()
|
D | page_alloc.c | 1132 struct list_head *list; in free_pcppages_bulk() local 1145 list = &pcp->lists[migratetype]; in free_pcppages_bulk() 1146 } while (list_empty(list)); in free_pcppages_bulk() 1155 page = list_last_entry(list, struct page, lru); in free_pcppages_bulk() 1171 } while (--count && --batch_free && !list_empty(list)); in free_pcppages_bulk() 2337 unsigned long count, struct list_head *list, in rmqueue_bulk() argument 2361 list_add(&page->lru, list); in rmqueue_bulk() 2363 list_add_tail(&page->lru, list); in rmqueue_bulk() 2364 list = &page->lru; in rmqueue_bulk() 2663 void free_hot_cold_page_list(struct list_head *list, bool cold) in free_hot_cold_page_list() argument [all …]
|
D | huge_memory.c | 2313 struct lruvec *lruvec, struct list_head *list) in __split_huge_page_tail() argument 2365 lru_add_page_tail(head, page_tail, lruvec, list); in __split_huge_page_tail() 2368 static void __split_huge_page(struct page *page, struct list_head *list, in __split_huge_page() argument 2382 __split_huge_page_tail(head, i, lruvec, list); in __split_huge_page() 2546 int split_huge_page_to_list(struct page *page, struct list_head *list) in split_huge_page_to_list() argument 2650 __split_huge_page(page, list, end, flags); in split_huge_page_to_list() 2728 LIST_HEAD(list), *pos, *next; in deferred_split_scan() 2738 list_move(page_deferred_list(page), &list); in deferred_split_scan() 2749 list_for_each_safe(pos, next, &list) { in deferred_split_scan() 2762 list_splice_tail(&list, &pgdata->split_queue); in deferred_split_scan()
|
D | slab.h | 28 struct list_head list; /* List of all slab caches on the system */ member 301 #define root_caches_node list
|
D | percpu-internal.h | 30 struct list_head list; /* linked to pcpu_slot lists */ member
|
D | memcontrol.c | 134 struct list_head list; member 153 struct list_head list; member 3406 list_for_each_entry(ev, &memcg->oom_notify, list) in mem_cgroup_oom_notify_cb() 3624 list_add(&event->list, &memcg->oom_notify); in mem_cgroup_oom_register_event() 3641 list_for_each_entry_safe(ev, tmp, &memcg->oom_notify, list) { in mem_cgroup_oom_unregister_event() 3643 list_del(&ev->list); in mem_cgroup_oom_unregister_event() 3828 if (!list_empty(&event->list)) { in memcg_event_wake() 3829 list_del_init(&event->list); in memcg_event_wake() 3891 INIT_LIST_HEAD(&event->list); in memcg_write_event_control() 3969 list_add(&event->list, &memcg->event_list); in memcg_write_event_control() [all …]
|
D | slob.c | 113 static void set_slob_page_free(struct page *sp, struct list_head *list) in set_slob_page_free() argument 115 list_add(&sp->lru, list); in set_slob_page_free()
|
D | vmscan.c | 290 list_add_tail(&shrinker->list, &shrinker_list); in register_shrinker() 304 list_del(&shrinker->list); in unregister_shrinker() 486 list_for_each_entry(shrinker, &shrinker_list, list) { in shrink_slab() 1949 struct list_head *list, in move_active_pages_to_lru() argument 1958 while (!list_empty(list)) { in move_active_pages_to_lru() 1959 page = lru_to_page(list); in move_active_pages_to_lru()
|