Home
last modified time | relevance | path

Searched refs:list (Results 1 – 24 of 24) sorted by relevance

/mm/
Dmmu_notifier.c66 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) in __mmu_notifier_release()
77 while (unlikely(!hlist_empty(&mm->mmu_notifier_mm->list))) { in __mmu_notifier_release()
78 mn = hlist_entry(mm->mmu_notifier_mm->list.first, in __mmu_notifier_release()
117 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_clear_flush_young()
133 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_test_young()
152 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_change_pte()
166 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_invalidate_page()
180 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_invalidate_range_start()
195 hlist_for_each_entry_rcu(mn, &mm->mmu_notifier_mm->list, hlist) { in __mmu_notifier_invalidate_range_end()
230 INIT_HLIST_HEAD(&mmu_notifier_mm->list); in do_mmu_notifier_register()
[all …]
Dzpool.c27 struct list_head list; member
44 list_add(&driver->list, &drivers_head); in zpool_register_driver()
69 list_del(&driver->list); in zpool_unregister_driver()
89 list_for_each_entry(zpool, &pools_head, list) { in zpool_evict()
108 list_for_each_entry(driver, &drivers_head, list) { in zpool_get_driver()
186 list_add(&zpool->list, &pools_head); in zpool_create_pool()
208 list_del(&zpool->list); in zpool_destroy_pool()
Dkmemleak-test.c36 struct list_head list; member
84 INIT_LIST_HEAD(&elem->list); in kmemleak_test_init()
85 list_add_tail(&elem->list, &test_list); in kmemleak_test_init()
106 list_for_each_entry_safe(elem, tmp, &test_list, list) in kmemleak_test_exit()
107 list_del(&elem->list); in kmemleak_test_exit()
Dslab.c248 int node, struct list_head *list);
249 static void slabs_destroy(struct kmem_cache *cachep, struct list_head *list);
916 struct list_head *list) in __drain_alien_cache() argument
930 free_block(cachep, ac->entry, ac->avail, node, list); in __drain_alien_cache()
950 LIST_HEAD(list); in reap_alien()
952 __drain_alien_cache(cachep, ac, node, &list); in reap_alien()
954 slabs_destroy(cachep, &list); in reap_alien()
971 LIST_HEAD(list); in drain_alien_cache()
975 __drain_alien_cache(cachep, ac, i, &list); in drain_alien_cache()
977 slabs_destroy(cachep, &list); in drain_alien_cache()
[all …]
Dvmalloc.c38 struct llist_head list; member
48 struct llist_node *llnode = llist_del_all(&p->list); in free_work()
333 list_add_rcu(&va->list, &prev->list); in __insert_vmap_area()
335 list_add_rcu(&va->list, &vmap_area_list); in __insert_vmap_area()
434 if (list_is_last(&first->list, &vmap_area_list)) in alloc_vmap_area()
437 first = list_entry(first->list.next, in alloc_vmap_area()
438 struct vmap_area, list); in alloc_vmap_area()
494 list_del_rcu(&va->list); in __free_vmap_area()
620 list_for_each_entry_rcu(va, &vmap_area_list, list) { in __purge_vmap_area_lazy()
1191 init_llist_head(&p->list); in vmalloc_init()
[all …]
Dlist_lru.c21 list_add_tail(item, &nlru->list); in list_lru_add()
77 list_for_each_safe(item, n, &nlru->list) { in list_lru_walk_node()
106 list_move_tail(item, &nlru->list); in list_lru_walk_node()
141 INIT_LIST_HEAD(&lru->node[i].list); in list_lru_init_key()
Dhighmem.c388 struct list_head list; member
427 list_for_each_entry(pam, &pas->lh, list) { in page_address()
461 list_add_tail(&pam->list, &pas->lh); in set_page_address()
465 list_for_each_entry(pam, &pas->lh, list) { in set_page_address()
467 list_del(&pam->list); in set_page_address()
Dswapfile.c143 list_for_each_entry(se, &si->first_swap_extent.list, list) { in discard_swap()
191 lh = se->list.next; in discard_swap_cluster()
192 se = list_entry(lh, struct swap_extent, list); in discard_swap_cluster()
1642 lh = se->list.next; in map_swap_entry()
1643 se = list_entry(lh, struct swap_extent, list); in map_swap_entry()
1664 while (!list_empty(&sis->first_swap_extent.list)) { in destroy_swap_extents()
1667 se = list_entry(sis->first_swap_extent.list.next, in destroy_swap_extents()
1668 struct swap_extent, list); in destroy_swap_extents()
1669 list_del(&se->list); in destroy_swap_extents()
1704 lh = sis->first_swap_extent.list.prev; /* Highest extent */ in add_swap_extent()
[all …]
Dbootmem.c82 list_for_each_entry(ent, &bdata_list, list) { in link_bootmem()
84 list_add_tail(&bdata->list, &ent->list); in link_bootmem()
89 list_add_tail(&bdata->list, &bdata_list); in link_bootmem()
280 list_for_each_entry(bdata, &bdata_list, list) in free_all_bootmem()
358 list_for_each_entry(bdata, &bdata_list, list) { in mark_bootmem()
611 list_for_each_entry(bdata, &bdata_list, list) { in alloc_bootmem_core()
Dpercpu.c105 struct list_head list; /* linked to pcpu_slot lists */ member
374 list_move(&chunk->list, &pcpu_slot[nslot]); in pcpu_chunk_relocate()
376 list_move_tail(&chunk->list, &pcpu_slot[nslot]); in pcpu_chunk_relocate()
741 INIT_LIST_HEAD(&chunk->list); in pcpu_alloc_chunk()
933 list_for_each_entry(chunk, &pcpu_slot[slot], list) { in pcpu_alloc()
1125 list_for_each_entry_safe(chunk, next, free_head, list) { in pcpu_balance_workfn()
1129 if (chunk == list_first_entry(free_head, struct pcpu_chunk, list)) in pcpu_balance_workfn()
1133 list_move(&chunk->list, &to_free); in pcpu_balance_workfn()
1138 list_for_each_entry_safe(chunk, next, &to_free, list) { in pcpu_balance_workfn()
1197 list_for_each_entry(chunk, &pcpu_slot[slot], list) { in pcpu_balance_workfn()
[all …]
Dslab_common.c81 list_for_each_entry(s, &slab_caches, list) { in kmem_cache_sanity_check()
174 list_for_each_entry(s, &slab_caches, list) { in memcg_update_all_caches()
243 list_for_each_entry(s, &slab_caches, list) { in find_mergeable()
326 list_add(&s->list, &slab_caches); in do_kmem_cache_create()
522 list_del(&s->list); in kmem_cache_destroy()
597 list_add(&s->list, &slab_caches); in create_kmalloc_cache()
908 struct kmem_cache *s = list_entry(p, struct kmem_cache, list); in s_show()
Dksm.c139 struct list_head list; member
511 list_del(&stable_node->list); in remove_node_from_stable_tree()
760 stable_node = list_entry(this, struct stable_node, list); in remove_all_stable_nodes()
1223 list_del(&page_node->list); in stable_tree_search()
1232 list_del(&page_node->list); in stable_tree_search()
1241 list_add(&stable_node->list, stable_node->head); in stable_tree_search()
1432 list_add(&stable_node->list, stable_node->head); in cmp_and_merge_page()
1574 struct stable_node, list); in scan_get_next_rmap_item()
2018 stable_node = list_entry(this, struct stable_node, list); in ksm_check_stable_tree()
Dmemcontrol.c227 struct list_head list; member
246 struct list_head list; member
2799 list_for_each_entry(params, &memcg->memcg_slab_caches, list) in mem_cgroup_slabinfo_read()
2957 list_add(&cachep->memcg_params->list, &memcg->memcg_slab_caches); in memcg_register_cache()
2987 list_del(&cachep->memcg_params->list); in memcg_unregister_cache()
3055 list_for_each_entry_safe(params, tmp, &memcg->memcg_slab_caches, list) { in memcg_unregister_all_caches()
3811 struct list_head *list; in mem_cgroup_force_empty_list() local
3817 list = &lruvec->lists[lru]; in mem_cgroup_force_empty_list()
3825 if (list_empty(list)) { in mem_cgroup_force_empty_list()
3829 page = list_entry(list->prev, struct page, lru); in mem_cgroup_force_empty_list()
[all …]
Dswap.c984 struct lruvec *lruvec, struct list_head *list) in lru_add_page_tail() argument
994 if (!list) in lru_add_page_tail()
999 else if (list) { in lru_add_page_tail()
1002 list_add_tail(&page_tail->lru, list); in lru_add_page_tail()
Dpage_alloc.c716 struct list_head *list; in free_pcppages_bulk() local
729 list = &pcp->lists[migratetype]; in free_pcppages_bulk()
730 } while (list_empty(list)); in free_pcppages_bulk()
739 page = list_entry(list->prev, struct page, lru); in free_pcppages_bulk()
749 } while (--to_free && --batch_free && !list_empty(list)); in free_pcppages_bulk()
1249 unsigned long count, struct list_head *list, in rmqueue_bulk() argument
1270 list_add(&page->lru, list); in rmqueue_bulk()
1272 list_add_tail(&page->lru, list); in rmqueue_bulk()
1273 list = &page->lru; in rmqueue_bulk()
1478 void free_hot_cold_page_list(struct list_head *list, bool cold) in free_hot_cold_page_list() argument
[all …]
Dfrontswap.c336 plist_for_each_entry(si, &swap_active_head, list) in __frontswap_curr_pages()
351 plist_for_each_entry(si, &swap_active_head, list) { in __frontswap_unuse_pages()
Dswap_state.c168 int add_to_swap(struct page *page, struct list_head *list) in add_to_swap() argument
181 if (unlikely(split_huge_page_to_list(page, list))) { in add_to_swap()
Dvmscan.c201 list_for_each_entry(shrinker, &shrinker_list, list) { in debug_shrinker_show()
246 list_add_tail(&shrinker->list, &shrinker_list); in register_shrinker()
267 list_del(&shrinker->list); in unregister_shrinker()
426 list_for_each_entry(shrinker, &shrinker_list, list) { in shrink_slab()
1674 struct list_head *list, in move_active_pages_to_lru() argument
1683 while (!list_empty(list)) { in move_active_pages_to_lru()
1684 page = lru_to_page(list); in move_active_pages_to_lru()
Dslob.c112 static void set_slob_page_free(struct page *sp, struct list_head *list) in set_slob_page_free() argument
114 list_add(&sp->lru, list); in set_slob_page_free()
Dhuge_memory.c1667 struct list_head *list) in __split_huge_page_refcount() argument
1755 lru_add_page_tail(page, page_tail, lruvec, list); in __split_huge_page_refcount()
1865 struct list_head *list) in __split_huge_page() argument
1897 __split_huge_page_refcount(page, list); in __split_huge_page()
1920 int split_huge_page_to_list(struct page *page, struct list_head *list) in split_huge_page_to_list() argument
1945 __split_huge_page(page, anon_vma, list); in split_huge_page_to_list()
Dslab.h27 struct list_head list; /* List of all slab caches on the system */ member
Dcompaction.c55 static void map_pages(struct list_head *list) in map_pages() argument
59 list_for_each_entry(page, list, lru) { in map_pages()
Dslub.c3495 list_for_each_entry(s, &slab_caches, list) in slab_mem_going_offline_callback()
3519 list_for_each_entry(s, &slab_caches, list) { in slab_mem_offline_callback()
3558 list_for_each_entry(s, &slab_caches, list) { in slab_mem_going_online_callback()
3644 list_add(&s->list, &slab_caches); in bootstrap()
3774 list_for_each_entry(s, &slab_caches, list) { in slab_cpuup_callback()
5280 list_for_each_entry(s, &slab_caches, list) { in slab_sysfs_init()
Dhugetlb.c1488 list_add(&m->list, &huge_boot_pages); in alloc_bootmem_huge_page()
1507 list_for_each_entry(m, &huge_boot_pages, list) { in gather_bootmem_prealloc()
3843 bool isolate_huge_page(struct page *page, struct list_head *list) in isolate_huge_page() argument
3854 list_move_tail(&page->lru, list); in isolate_huge_page()