/mm/ |
D | mmu_notifier.c | 66 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) in __mmu_notifier_release() 77 while (unlikely(!hlist_empty(&mm->mmu_notifier_mm->list))) { in __mmu_notifier_release() 78 mn = hlist_entry(mm->mmu_notifier_mm->list.first, in __mmu_notifier_release() 117 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_clear_flush_young() 134 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_clear_young() 150 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_test_young() 169 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_change_pte() 183 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_invalidate_page() 197 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_invalidate_range_start() 212 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_invalidate_range_end() [all …]
|
D | list_lru.c | 22 list_add(&lru->list, &list_lrus); in list_lru_register() 29 list_del(&lru->list); in list_lru_unregister() 114 list_add_tail(item, &l->list); in list_lru_add() 145 void list_lru_isolate(struct list_lru_one *list, struct list_head *item) in list_lru_isolate() argument 148 list->nr_items--; in list_lru_isolate() 152 void list_lru_isolate_move(struct list_lru_one *list, struct list_head *item, in list_lru_isolate_move() argument 156 list->nr_items--; in list_lru_isolate_move() 205 list_for_each_safe(item, n, &l->list) { in __list_lru_walk_one() 232 list_move_tail(item, &l->list); in __list_lru_walk_one() 285 INIT_LIST_HEAD(&l->list); in init_one_lru() [all …]
|
D | slab_common.c | 80 list_for_each_entry(s, &slab_caches, list) { in kmem_cache_sanity_check() 134 INIT_LIST_HEAD(&s->memcg_params.list); in slab_init_memcg_params() 201 list_for_each_entry(s, &slab_caches, list) { in memcg_update_all_caches() 267 list_for_each_entry_reverse(s, &slab_caches, list) { in find_mergeable() 350 list_add(&s->list, &slab_caches); in create_cache() 465 list_move(&s->list, release); in shutdown_cache() 476 list_for_each_entry_safe(s, s2, release, list) { in release_caches() 549 list_add(&s->memcg_params.list, &root_cache->memcg_params.list); in memcg_create_kmem_cache() 578 list_for_each_entry(s, &slab_caches, list) { in memcg_deactivate_kmem_caches() 605 list_del(&s->memcg_params.list); in __shutdown_memcg_cache() [all …]
|
D | kmemleak-test.c | 36 struct list_head list; member 84 INIT_LIST_HEAD(&elem->list); in kmemleak_test_init() 85 list_add_tail(&elem->list, &test_list); in kmemleak_test_init() 106 list_for_each_entry_safe(elem, tmp, &test_list, list) in kmemleak_test_exit() 107 list_del(&elem->list); in kmemleak_test_exit()
|
D | slab.c | 248 int node, struct list_head *list); 249 static void slabs_destroy(struct kmem_cache *cachep, struct list_head *list); 909 struct list_head *list) in __drain_alien_cache() argument 923 free_block(cachep, ac->entry, ac->avail, node, list); in __drain_alien_cache() 943 LIST_HEAD(list); in reap_alien() 945 __drain_alien_cache(cachep, ac, node, &list); in reap_alien() 947 slabs_destroy(cachep, &list); in reap_alien() 964 LIST_HEAD(list); in drain_alien_cache() 968 __drain_alien_cache(cachep, ac, i, &list); in drain_alien_cache() 970 slabs_destroy(cachep, &list); in drain_alien_cache() [all …]
|
D | zpool.c | 25 struct list_head list; member 42 list_add(&driver->list, &drivers_head); in zpool_register_driver() 67 list_del(&driver->list); in zpool_unregister_driver() 80 list_for_each_entry(driver, &drivers_head, list) { in zpool_get_driver() 194 list_add(&zpool->list, &pools_head); in zpool_create_pool() 216 list_del(&zpool->list); in zpool_destroy_pool()
|
D | vmalloc.c | 41 struct llist_head list; member 51 struct llist_node *llnode = llist_del_all(&p->list); in free_work() 340 list_add_rcu(&va->list, &prev->list); in __insert_vmap_area() 342 list_add_rcu(&va->list, &vmap_area_list); in __insert_vmap_area() 441 if (list_is_last(&first->list, &vmap_area_list)) in alloc_vmap_area() 444 first = list_entry(first->list.next, in alloc_vmap_area() 445 struct vmap_area, list); in alloc_vmap_area() 504 list_del_rcu(&va->list); in __free_vmap_area() 630 list_for_each_entry_rcu(va, &vmap_area_list, list) { in __purge_vmap_area_lazy() 1216 init_llist_head(&p->list); in vmalloc_init() [all …]
|
D | highmem.c | 388 struct list_head list; member 427 list_for_each_entry(pam, &pas->lh, list) { in page_address() 461 list_add_tail(&pam->list, &pas->lh); in set_page_address() 465 list_for_each_entry(pam, &pas->lh, list) { in set_page_address() 467 list_del(&pam->list); in set_page_address()
|
D | zswap.c | 125 struct list_head list; member 217 list_for_each_entry_rcu(pool, &zswap_pools, list) in zswap_update_total_size() 511 pool = list_first_or_null_rcu(&zswap_pools, typeof(*pool), list); in __zswap_pool_current() 545 list_for_each_entry_rcu(pool, &zswap_pools, list) in zswap_pool_last_get() 562 list_for_each_entry_rcu(pool, &zswap_pools, list) { in zswap_pool_find_get() 613 INIT_LIST_HEAD(&pool->list); in zswap_pool_create() 693 list_del_rcu(&pool->list); in __zswap_pool_empty() 756 list_del_rcu(&pool->list); in __zswap_param_set() 773 list_add_rcu(&pool->list, &zswap_pools); in __zswap_param_set() 779 list_add_tail_rcu(&pool->list, &zswap_pools); in __zswap_param_set() [all …]
|
D | swapfile.c | 143 list_for_each_entry(se, &si->first_swap_extent.list, list) { in discard_swap() 191 lh = se->list.next; in discard_swap_cluster() 192 se = list_entry(lh, struct swap_extent, list); in discard_swap_cluster() 1642 lh = se->list.next; in map_swap_entry() 1643 se = list_entry(lh, struct swap_extent, list); in map_swap_entry() 1664 while (!list_empty(&sis->first_swap_extent.list)) { in destroy_swap_extents() 1667 se = list_entry(sis->first_swap_extent.list.next, in destroy_swap_extents() 1668 struct swap_extent, list); in destroy_swap_extents() 1669 list_del(&se->list); in destroy_swap_extents() 1704 lh = sis->first_swap_extent.list.prev; /* Highest extent */ in add_swap_extent() [all …]
|
D | bootmem.c | 82 list_for_each_entry(ent, &bdata_list, list) { in link_bootmem() 84 list_add_tail(&bdata->list, &ent->list); in link_bootmem() 89 list_add_tail(&bdata->list, &bdata_list); in link_bootmem() 281 list_for_each_entry(bdata, &bdata_list, list) in free_all_bootmem() 365 list_for_each_entry(bdata, &bdata_list, list) { in mark_bootmem() 618 list_for_each_entry(bdata, &bdata_list, list) { in alloc_bootmem_core()
|
D | percpu.c | 106 struct list_head list; /* linked to pcpu_slot lists */ member 375 list_move(&chunk->list, &pcpu_slot[nslot]); in pcpu_chunk_relocate() 377 list_move_tail(&chunk->list, &pcpu_slot[nslot]); in pcpu_chunk_relocate() 742 INIT_LIST_HEAD(&chunk->list); in pcpu_alloc_chunk() 934 list_for_each_entry(chunk, &pcpu_slot[slot], list) { in pcpu_alloc() 1126 list_for_each_entry_safe(chunk, next, free_head, list) { in pcpu_balance_workfn() 1130 if (chunk == list_first_entry(free_head, struct pcpu_chunk, list)) in pcpu_balance_workfn() 1134 list_move(&chunk->list, &to_free); in pcpu_balance_workfn() 1139 list_for_each_entry_safe(chunk, next, &to_free, list) { in pcpu_balance_workfn() 1198 list_for_each_entry(chunk, &pcpu_slot[slot], list) { in pcpu_balance_workfn() [all …]
|
D | frontswap.c | 123 plist_for_each_entry(si, &swap_active_head, list) { in frontswap_register_ops() 143 plist_for_each_entry(si, &swap_active_head, list) { in frontswap_register_ops() 383 plist_for_each_entry(si, &swap_active_head, list) in __frontswap_curr_pages() 398 plist_for_each_entry(si, &swap_active_head, list) { in __frontswap_unuse_pages()
|
D | slab.h | 27 struct list_head list; /* List of all slab caches on the system */ member 181 list_for_each_entry(iter, &(root)->memcg_params.list, \ 182 memcg_params.list)
|
D | ksm.c | 139 struct list_head list; member 512 list_del(&stable_node->list); in remove_node_from_stable_tree() 762 stable_node = list_entry(this, struct stable_node, list); in remove_all_stable_nodes() 1231 list_del(&page_node->list); in stable_tree_search() 1240 list_del(&page_node->list); in stable_tree_search() 1249 list_add(&stable_node->list, stable_node->head); in stable_tree_search() 1452 list_add(&stable_node->list, stable_node->head); in cmp_and_merge_page() 1622 struct stable_node, list); in scan_get_next_rmap_item() 2068 stable_node = list_entry(this, struct stable_node, list); in ksm_check_stable_tree()
|
D | swap.c | 995 struct lruvec *lruvec, struct list_head *list) in lru_add_page_tail() argument 1005 if (!list) in lru_add_page_tail() 1010 else if (list) { in lru_add_page_tail() 1013 list_add_tail(&page_tail->lru, list); in lru_add_page_tail()
|
D | page_alloc.c | 864 struct list_head *list; in free_pcppages_bulk() local 877 list = &pcp->lists[migratetype]; in free_pcppages_bulk() 878 } while (list_empty(list)); in free_pcppages_bulk() 887 page = list_entry(list->prev, struct page, lru); in free_pcppages_bulk() 900 } while (--count && --batch_free && !list_empty(list)); in free_pcppages_bulk() 1898 unsigned long count, struct list_head *list, in rmqueue_bulk() argument 1919 list_add(&page->lru, list); in rmqueue_bulk() 1921 list_add_tail(&page->lru, list); in rmqueue_bulk() 1922 list = &page->lru; in rmqueue_bulk() 2158 void free_hot_cold_page_list(struct list_head *list, bool cold) in free_hot_cold_page_list() argument [all …]
|
D | swap_state.c | 162 int add_to_swap(struct page *page, struct list_head *list) in add_to_swap() argument 175 if (unlikely(split_huge_page_to_list(page, list))) { in add_to_swap()
|
D | memcontrol.c | 141 struct list_head list; member 160 struct list_head list; member 3368 list_for_each_entry(ev, &memcg->oom_notify, list) in mem_cgroup_oom_notify_cb() 3586 list_add(&event->list, &memcg->oom_notify); in mem_cgroup_oom_register_event() 3603 list_for_each_entry_safe(ev, tmp, &memcg->oom_notify, list) { in mem_cgroup_oom_unregister_event() 3605 list_del(&ev->list); in mem_cgroup_oom_unregister_event() 3874 if (!list_empty(&event->list)) { in memcg_event_wake() 3875 list_del_init(&event->list); in memcg_event_wake() 3937 INIT_LIST_HEAD(&event->list); in memcg_write_event_control() 4015 list_add(&event->list, &memcg->event_list); in memcg_write_event_control() [all …]
|
D | slob.c | 112 static void set_slob_page_free(struct page *sp, struct list_head *list) in set_slob_page_free() argument 114 list_add(&sp->lru, list); in set_slob_page_free()
|
D | vmscan.c | 246 list_add_tail(&shrinker->list, &shrinker_list); in register_shrinker() 260 list_del(&shrinker->list); in unregister_shrinker() 442 list_for_each_entry(shrinker, &shrinker_list, list) { in shrink_slab() 1742 struct list_head *list, in move_active_pages_to_lru() argument 1751 while (!list_empty(list)) { in move_active_pages_to_lru() 1752 page = lru_to_page(list); in move_active_pages_to_lru()
|
D | huge_memory.c | 1746 struct list_head *list) in __split_huge_page_refcount() argument 1833 lru_add_page_tail(page, page_tail, lruvec, list); in __split_huge_page_refcount() 1943 struct list_head *list) in __split_huge_page() argument 1975 __split_huge_page_refcount(page, list); in __split_huge_page() 1998 int split_huge_page_to_list(struct page *page, struct list_head *list) in split_huge_page_to_list() argument 2023 __split_huge_page(page, anon_vma, list); in split_huge_page_to_list()
|
D | compaction.c | 58 static void map_pages(struct list_head *list) in map_pages() argument 62 list_for_each_entry(page, list, lru) { in map_pages()
|
D | slub.c | 3865 list_for_each_entry(s, &slab_caches, list) in slab_mem_going_offline_callback() 3889 list_for_each_entry(s, &slab_caches, list) { in slab_mem_offline_callback() 3928 list_for_each_entry(s, &slab_caches, list) { in slab_mem_going_online_callback() 4015 list_add(&s->list, &slab_caches); in bootstrap() 4140 list_for_each_entry(s, &slab_caches, list) { in slab_cpuup_callback() 5646 list_for_each_entry(s, &slab_caches, list) { in slab_sysfs_init()
|
D | hugetlb.c | 2050 list_add(&m->list, &huge_boot_pages); in alloc_bootmem_huge_page() 2069 list_for_each_entry(m, &huge_boot_pages, list) { in gather_bootmem_prealloc() 4623 bool isolate_huge_page(struct page *page, struct list_head *list) in isolate_huge_page() argument 4634 list_move_tail(&page->lru, list); in isolate_huge_page()
|