/mm/ |
D | slab.c | 213 int node, struct list_head *list); 214 static void slabs_destroy(struct kmem_cache *cachep, struct list_head *list); 219 void **list); 222 void **list); 557 LIST_HEAD(list); in cache_free_pfmemalloc() 563 free_block(cachep, &objp, 1, page_node, &list); in cache_free_pfmemalloc() 566 slabs_destroy(cachep, &list); in cache_free_pfmemalloc() 697 struct list_head *list) in __drain_alien_cache() argument 711 free_block(cachep, ac->entry, ac->avail, node, list); in __drain_alien_cache() 731 LIST_HEAD(list); in reap_alien() [all …]
|
D | mmu_notifier.c | 39 struct hlist_head list; member 310 hlist_for_each_entry_rcu(subscription, &subscriptions->list, hlist, in mn_hlist_release() 322 while (unlikely(!hlist_empty(&subscriptions->list))) { in mn_hlist_release() 323 subscription = hlist_entry(subscriptions->list.first, in mn_hlist_release() 356 if (!hlist_empty(&subscriptions->list)) in __mmu_notifier_release() 374 &mm->notifier_subscriptions->list, hlist, in __mmu_notifier_clear_flush_young() 394 &mm->notifier_subscriptions->list, hlist, in __mmu_notifier_clear_young() 413 &mm->notifier_subscriptions->list, hlist, in __mmu_notifier_test_young() 435 &mm->notifier_subscriptions->list, hlist, in __mmu_notifier_change_pte() 484 hlist_for_each_entry_rcu(subscription, &subscriptions->list, hlist, in mn_hlist_invalidate_range_start() [all …]
|
D | list_lru.c | 24 list_add(&lru->list, &list_lrus); in list_lru_register() 31 list_del(&lru->list); in list_lru_unregister() 125 list_add_tail(item, &l->list); in list_lru_add() 159 void list_lru_isolate(struct list_lru_one *list, struct list_head *item) in list_lru_isolate() argument 162 list->nr_items--; in list_lru_isolate() 166 void list_lru_isolate_move(struct list_lru_one *list, struct list_head *item, in list_lru_isolate_move() argument 170 list->nr_items--; in list_lru_isolate_move() 211 list_for_each_safe(item, n, &l->list) { in __list_lru_walk_one() 239 list_move_tail(item, &l->list); in __list_lru_walk_one() 317 INIT_LIST_HEAD(&l->list); in init_one_lru() [all …]
|
D | page_reporting.c | 120 struct list_head *list = &area->free_list[mt]; in page_reporting_cycle() local 130 if (list_empty(list)) in page_reporting_cycle() 152 list_for_each_entry_safe(page, next, list, lru) { in page_reporting_cycle() 191 if (!list_is_first(&page->lru, list)) in page_reporting_cycle() 192 list_rotate_to_front(&page->lru, list); in page_reporting_cycle() 216 next = list_first_entry(list, struct page, lru); in page_reporting_cycle() 224 if (!list_entry_is_head(next, list, lru) && !list_is_first(&next->lru, list)) in page_reporting_cycle() 225 list_rotate_to_front(&next->lru, list); in page_reporting_cycle()
|
D | zpool.c | 28 struct list_head list; member 45 list_add(&driver->list, &drivers_head); in zpool_register_driver() 70 list_del(&driver->list); in zpool_unregister_driver() 83 list_for_each_entry(driver, &drivers_head, list) { in zpool_get_driver() 199 list_add(&zpool->list, &pools_head); in zpool_create_pool() 221 list_del(&zpool->list); in zpool_destroy_pool()
|
D | slab_common.c | 214 list_for_each_entry_reverse(s, &slab_caches, list) { in find_mergeable() 271 list_add(&s->list, &slab_caches); in create_cache() 451 list_for_each_entry_safe(s, s2, &to_destroy, list) { in slab_caches_to_rcu_destroy_workfn() 470 list_del(&s->list); in shutdown_cache() 476 list_add_tail(&s->list, &slab_caches_to_rcu_destroy); in shutdown_cache() 692 list_add(&s->list, &slab_caches); in create_kmalloc_cache() 992 static void freelist_randomize(struct rnd_state *state, unsigned int *list, in freelist_randomize() argument 999 list[i] = i; in freelist_randomize() 1005 swap(list[i], list[rand]); in freelist_randomize() 1104 struct kmem_cache *s = list_entry(p, struct kmem_cache, list); in slab_show() [all …]
|
D | zswap.c | 144 struct list_head list; member 251 list_for_each_entry_rcu(pool, &zswap_pools, list) in zswap_update_total_size() 507 pool = list_first_or_null_rcu(&zswap_pools, typeof(*pool), list); in __zswap_pool_current() 542 list_for_each_entry_rcu(pool, &zswap_pools, list) in zswap_pool_last_get() 561 list_for_each_entry_rcu(pool, &zswap_pools, list) { in zswap_pool_find_get() 635 INIT_LIST_HEAD(&pool->list); in zswap_pool_create() 735 list_del_rcu(&pool->list); in __zswap_pool_empty() 798 list_del_rcu(&pool->list); in __zswap_param_set() 815 list_add_rcu(&pool->list, &zswap_pools); in __zswap_param_set() 822 list_add_tail_rcu(&pool->list, &zswap_pools); in __zswap_param_set() [all …]
|
D | swapfile.c | 395 static inline bool cluster_list_empty(struct swap_cluster_list *list) in cluster_list_empty() argument 397 return cluster_is_null(&list->head); in cluster_list_empty() 400 static inline unsigned int cluster_list_first(struct swap_cluster_list *list) in cluster_list_first() argument 402 return cluster_next(&list->head); in cluster_list_first() 405 static void cluster_list_init(struct swap_cluster_list *list) in cluster_list_init() argument 407 cluster_set_null(&list->head); in cluster_list_init() 408 cluster_set_null(&list->tail); in cluster_list_init() 411 static void cluster_list_add_tail(struct swap_cluster_list *list, in cluster_list_add_tail() argument 415 if (cluster_list_empty(list)) { in cluster_list_add_tail() 416 cluster_set_next_flag(&list->head, idx, 0); in cluster_list_add_tail() [all …]
|
D | vmalloc.c | 83 struct llist_head list; member 95 llist_for_each_safe(llnode, t, llist_del_all(&p->list)) in free_work() 914 struct list_head *list; in get_va_next_sibling() local 925 list = &rb_entry(parent, struct vmap_area, rb_node)->list; in get_va_next_sibling() 926 return (&parent->rb_right == link ? list->next : list); in get_va_next_sibling() 938 head = &rb_entry(parent, struct vmap_area, rb_node)->list; in link_va() 965 list_add(&va->list, head); in link_va() 980 list_del(&va->list); in unlink_va() 991 list_for_each_entry(va, &free_vmap_area_list, list) { in augment_tree_propagate_check() 1117 sibling = list_entry(next, struct vmap_area, list); in merge_or_add_vmap_area() [all …]
|
D | frontswap.c | 128 plist_for_each_entry(si, &swap_active_head, list) { in frontswap_register_ops() 150 plist_for_each_entry(si, &swap_active_head, list) { in frontswap_register_ops() 381 plist_for_each_entry(si, &swap_active_head, list) in __frontswap_curr_pages() 396 plist_for_each_entry(si, &swap_active_head, list) { in __frontswap_unuse_pages()
|
D | vmscan.c | 682 list_add_tail(&shrinker->list, &shrinker_list); in register_shrinker_prepared() 707 list_del(&shrinker->list); in unregister_shrinker() 946 list_for_each_entry(shrinker, &shrinker_list, list) { in shrink_slab() 2157 struct list_head *list) in move_pages_to_lru() argument 2163 while (!list_empty(list)) { in move_pages_to_lru() 2164 page = lru_to_page(list); in move_pages_to_lru() 2215 list_splice(&pages_to_free, list); in move_pages_to_lru() 3029 VM_WARN_ON_ONCE(!list_empty(&mm->lru_gen.list)); in lru_gen_add_mm() 3044 lruvec->mm_state.tail = &mm->lru_gen.list; in lru_gen_add_mm() 3047 list_add_tail(&mm->lru_gen.list, &mm_list->fifo); in lru_gen_add_mm() [all …]
|
D | sparse-vmemmap.c | 223 static void free_vmemmap_page_list(struct list_head *list) in free_vmemmap_page_list() argument 227 list_for_each_entry_safe(page, next, list, lru) { in free_vmemmap_page_list() 331 gfp_t gfp_mask, struct list_head *list) in alloc_vmemmap_page_list() argument 341 list_add_tail(&page->lru, list); in alloc_vmemmap_page_list() 346 list_for_each_entry_safe(page, next, list, lru) in alloc_vmemmap_page_list()
|
D | highmem.c | 700 struct list_head list; member 739 list_for_each_entry(pam, &pas->lh, list) { in page_address() 772 list_add_tail(&pam->list, &pas->lh); in set_page_address() 776 list_for_each_entry(pam, &pas->lh, list) { in set_page_address() 778 list_del(&pam->list); in set_page_address()
|
D | zsmalloc.c | 289 struct list_head list; /* fullness list */ member 724 struct zspage, list); in insert_zspage() 730 list_add(&zspage->list, &head->list); in insert_zspage() 732 list_add(&zspage->list, &class->fullness_list[fullness]); in insert_zspage() 746 list_del_init(&zspage->list); in remove_zspage() 964 VM_BUG_ON(list_empty(&zspage->list)); in free_zspage() 1100 struct zspage, list); in find_get_zspage() 1707 struct zspage, list); in isolate_zspage() 1929 if (list_empty(&zspage->list) && !is_zspage_isolated(zspage)) { in zs_page_isolate() 1938 if (!list_empty(&zspage->list) && !is_zspage_isolated(zspage)) { in zs_page_isolate() [all …]
|
D | percpu.c | 535 list_move(&chunk->list, &pcpu_chunk_lists[slot]); in __pcpu_chunk_move() 537 list_move_tail(&chunk->list, &pcpu_chunk_lists[slot]); in __pcpu_chunk_move() 579 list_move(&chunk->list, &pcpu_chunk_lists[pcpu_to_depopulate_slot]); in pcpu_isolate_chunk() 1370 INIT_LIST_HEAD(&chunk->list); in pcpu_alloc_first_chunk() 1450 INIT_LIST_HEAD(&chunk->list); in pcpu_alloc_chunk() 1804 list) { in pcpu_alloc() 1998 list_for_each_entry_safe(chunk, next, free_head, list) { in pcpu_balance_free() 2002 if (chunk == list_first_entry(free_head, struct pcpu_chunk, list)) in pcpu_balance_free() 2006 list_move(&chunk->list, &to_free); in pcpu_balance_free() 2013 list_for_each_entry_safe(chunk, next, &to_free, list) { in pcpu_balance_free() [all …]
|
D | ksm.c | 162 struct list_head list; member 658 list_del(&stable_node->list); in remove_node_from_stable_tree() 951 list_for_each_entry_safe(stable_node, next, &migrate_nodes, list) { in remove_all_stable_nodes() 1696 list_del(&page_node->list); in stable_tree_search() 1722 list_del(&page_node->list); in stable_tree_search() 1740 list_del(&page_node->list); in stable_tree_search() 1752 list_add(&stable_node_dup->list, stable_node_dup->head); in stable_tree_search() 1783 list_del(&page_node->list); in stable_tree_search() 2046 list_add(&stable_node->list, stable_node->head); in cmp_and_merge_page() 2253 &migrate_nodes, list) { in scan_get_next_rmap_item() [all …]
|
D | percpu-stats.c | 40 list_for_each_entry(chunk, &pcpu_chunk_lists[slot], list) in find_max_nr_alloc() 206 list_for_each_entry(chunk, &pcpu_chunk_lists[slot], list) { in percpu_stats_show()
|
D | huge_memory.c | 2322 struct lruvec *lruvec, struct list_head *list) in lru_add_page_tail() argument 2329 if (list) { in lru_add_page_tail() 2333 list_add_tail(&tail->lru, list); in lru_add_page_tail() 2343 struct lruvec *lruvec, struct list_head *list) in __split_huge_page_tail() argument 2405 lru_add_page_tail(head, page_tail, lruvec, list); in __split_huge_page_tail() 2408 static void __split_huge_page(struct page *page, struct list_head *list, in __split_huge_page() argument 2435 __split_huge_page_tail(head, i, lruvec, list); in __split_huge_page() 2615 int split_huge_page_to_list(struct page *page, struct list_head *list) in split_huge_page_to_list() argument 2723 __split_huge_page(page, list, end); in split_huge_page_to_list() 2820 LIST_HEAD(list), *pos, *next; in deferred_split_scan() [all …]
|
D | page_alloc.c | 1568 struct list_head *list; in free_pcppages_bulk() local 1581 list = &pcp->lists[pindex]; in free_pcppages_bulk() 1582 } while (list_empty(list)); in free_pcppages_bulk() 1591 page = list_last_entry(list, struct page, pcp_list); in free_pcppages_bulk() 1619 } while (count > 0 && --batch_free && !list_empty(list)); in free_pcppages_bulk() 3201 unsigned long count, struct list_head *list, in rmqueue_bulk() argument 3237 list_add_tail(&page->pcp_list, list); in rmqueue_bulk() 3264 struct list_head *list = &pcp->lists[order_to_pindex(migratetype, order)]; in get_populated_pcp_list() local 3266 if (list_empty(list)) { in get_populated_pcp_list() 3270 trace_android_vh_rmqueue_bulk_bypass(order, pcp, migratetype, list); in get_populated_pcp_list() [all …]
|
D | memcontrol.c | 131 struct list_head list; member 150 struct list_head list; member 305 list_del(&objcg->list); in obj_cgroup_release() 327 INIT_LIST_HEAD(&objcg->list); in obj_cgroup_alloc() 341 list_add(&objcg->list, &memcg->objcg_list); in memcg_reparent_objcgs() 343 list_for_each_entry(iter, &memcg->objcg_list, list) in memcg_reparent_objcgs() 4231 list_for_each_entry(ev, &memcg->oom_notify, list) in mem_cgroup_oom_notify_cb() 4447 list_add(&event->list, &memcg->oom_notify); in mem_cgroup_oom_register_event() 4464 list_for_each_entry_safe(ev, tmp, &memcg->oom_notify, list) { in mem_cgroup_oom_unregister_event() 4466 list_del(&ev->list); in mem_cgroup_oom_unregister_event() [all …]
|
D | compaction.c | 89 static void split_map_pages(struct list_head *list) in split_map_pages() argument 95 list_for_each_entry_safe(page, next, list, lru) { in split_map_pages() 111 list_splice(&tmp_list, list); in split_map_pages() 765 struct list_head *list) in isolate_and_split_free_page() argument 779 list_add(&page->lru, list); in isolate_and_split_free_page() 781 split_map_pages(list); in isolate_and_split_free_page()
|
D | percpu-internal.h | 40 struct list_head list; /* linked to pcpu_slot lists */ member
|
/mm/kfence/ |
D | core.c | 343 meta = list_entry(kfence_freelist.next, struct kfence_metadata, list); in kfence_guarded_alloc() 344 list_del_init(&meta->list); in kfence_guarded_alloc() 364 list_add_tail(&meta->list, &kfence_freelist); in kfence_guarded_alloc() 489 KFENCE_WARN_ON(!list_empty(&meta->list)); in kfence_guarded_free() 490 list_add_tail(&meta->list, &kfence_freelist); in kfence_guarded_free() 565 INIT_LIST_HEAD(&meta->list); in kfence_init_pool() 569 list_add_tail(&meta->list, &kfence_freelist); in kfence_init_pool()
|
D | kfence.h | 47 struct list_head list; /* Freelist node; access under kfence_freelist_lock. */ member
|
/mm/damon/ |
D | core.c | 44 INIT_LIST_HEAD(®ion->list); in damon_new_region() 54 list_add_tail(&r->list, &t->regions_list); in damon_add_region() 60 list_del(&r->list); in damon_del_region() 95 INIT_LIST_HEAD(&scheme->list); in damon_new_scheme() 123 list_add_tail(&s->list, &ctx->schemes); in damon_add_scheme() 128 list_del(&s->list); in damon_del_scheme() 164 list_add_tail(&t->list, &ctx->adaptive_targets); in damon_add_target() 174 list_del(&t->list); in damon_del_target()
|