Home
last modified time | relevance | path

Searched refs:list (Results 1 – 25 of 30) sorted by relevance

12

/mm/
Dslab.c213 int node, struct list_head *list);
214 static void slabs_destroy(struct kmem_cache *cachep, struct list_head *list);
219 void **list);
222 void **list);
557 LIST_HEAD(list); in cache_free_pfmemalloc()
563 free_block(cachep, &objp, 1, page_node, &list); in cache_free_pfmemalloc()
566 slabs_destroy(cachep, &list); in cache_free_pfmemalloc()
697 struct list_head *list) in __drain_alien_cache() argument
711 free_block(cachep, ac->entry, ac->avail, node, list); in __drain_alien_cache()
731 LIST_HEAD(list); in reap_alien()
[all …]
Dlist_lru.c24 list_add(&lru->list, &list_lrus); in list_lru_register()
31 list_del(&lru->list); in list_lru_unregister()
125 list_add_tail(item, &l->list); in list_lru_add()
159 void list_lru_isolate(struct list_lru_one *list, struct list_head *item) in list_lru_isolate() argument
162 list->nr_items--; in list_lru_isolate()
166 void list_lru_isolate_move(struct list_lru_one *list, struct list_head *item, in list_lru_isolate_move() argument
170 list->nr_items--; in list_lru_isolate_move()
211 list_for_each_safe(item, n, &l->list) { in __list_lru_walk_one()
239 list_move_tail(item, &l->list); in __list_lru_walk_one()
317 INIT_LIST_HEAD(&l->list); in init_one_lru()
[all …]
Dpage_reporting.c114 struct list_head *list = &area->free_list[mt]; in page_reporting_cycle() local
124 if (list_empty(list)) in page_reporting_cycle()
146 list_for_each_entry_safe(page, next, list, lru) { in page_reporting_cycle()
181 if (!list_is_first(&page->lru, list)) in page_reporting_cycle()
182 list_rotate_to_front(&page->lru, list); in page_reporting_cycle()
206 next = list_first_entry(list, struct page, lru); in page_reporting_cycle()
214 if (&next->lru != list && !list_is_first(&next->lru, list)) in page_reporting_cycle()
215 list_rotate_to_front(&next->lru, list); in page_reporting_cycle()
Dmmu_notifier.c45 struct hlist_head list; member
316 hlist_for_each_entry_rcu(subscription, &subscriptions->list, hlist, in mn_hlist_release()
328 while (unlikely(!hlist_empty(&subscriptions->list))) { in mn_hlist_release()
329 subscription = hlist_entry(subscriptions->list.first, in mn_hlist_release()
362 if (!hlist_empty(&subscriptions->list)) in __mmu_notifier_release()
380 &mm->notifier_subscriptions->list, hlist, in __mmu_notifier_clear_flush_young()
400 &mm->notifier_subscriptions->list, hlist, in __mmu_notifier_clear_young()
419 &mm->notifier_subscriptions->list, hlist, in __mmu_notifier_test_young()
441 &mm->notifier_subscriptions->list, hlist, in __mmu_notifier_change_pte()
490 hlist_for_each_entry_rcu(subscription, &subscriptions->list, hlist, in mn_hlist_invalidate_range_start()
[all …]
Dzpool.c27 struct list_head list; member
44 list_add(&driver->list, &drivers_head); in zpool_register_driver()
69 list_del(&driver->list); in zpool_unregister_driver()
82 list_for_each_entry(driver, &drivers_head, list) { in zpool_get_driver()
197 list_add(&zpool->list, &pools_head); in zpool_create_pool()
219 list_del(&zpool->list); in zpool_destroy_pool()
Dslab_common.c205 list_for_each_entry_reverse(s, &slab_caches, list) { in find_mergeable()
262 list_add(&s->list, &slab_caches); in create_cache()
448 list_for_each_entry_safe(s, s2, &to_destroy, list) { in slab_caches_to_rcu_destroy_workfn()
467 list_del(&s->list); in shutdown_cache()
473 list_add_tail(&s->list, &slab_caches_to_rcu_destroy); in shutdown_cache()
597 list_add(&s->list, &slab_caches); in create_kmalloc_cache()
882 static void freelist_randomize(struct rnd_state *state, unsigned int *list, in freelist_randomize() argument
889 list[i] = i; in freelist_randomize()
895 swap(list[i], list[rand]); in freelist_randomize()
992 struct kmem_cache *s = list_entry(p, struct kmem_cache, list); in slab_show()
[all …]
Dhighmem.c384 struct list_head list; member
423 list_for_each_entry(pam, &pas->lh, list) { in page_address()
457 list_add_tail(&pam->list, &pas->lh); in set_page_address()
461 list_for_each_entry(pam, &pas->lh, list) { in set_page_address()
463 list_del(&pam->list); in set_page_address()
Dzswap.c134 struct list_head list; member
241 list_for_each_entry_rcu(pool, &zswap_pools, list) in zswap_update_total_size()
453 pool = list_first_or_null_rcu(&zswap_pools, typeof(*pool), list); in __zswap_pool_current()
488 list_for_each_entry_rcu(pool, &zswap_pools, list) in zswap_pool_last_get()
507 list_for_each_entry_rcu(pool, &zswap_pools, list) { in zswap_pool_find_get()
580 INIT_LIST_HEAD(&pool->list); in zswap_pool_create()
679 list_del_rcu(&pool->list); in __zswap_pool_empty()
742 list_del_rcu(&pool->list); in __zswap_param_set()
759 list_add_rcu(&pool->list, &zswap_pools); in __zswap_param_set()
766 list_add_tail_rcu(&pool->list, &zswap_pools); in __zswap_param_set()
[all …]
Dvmalloc.c55 struct llist_head list; member
67 llist_for_each_safe(llnode, t, llist_del_all(&p->list)) in free_work()
575 struct list_head *list; in get_va_next_sibling() local
586 list = &rb_entry(parent, struct vmap_area, rb_node)->list; in get_va_next_sibling()
587 return (&parent->rb_right == link ? list->next : list); in get_va_next_sibling()
599 head = &rb_entry(parent, struct vmap_area, rb_node)->list; in link_va()
626 list_add(&va->list, head); in link_va()
641 list_del(&va->list); in unlink_va()
652 list_for_each_entry(va, &free_vmap_area_list, list) { in augment_tree_propagate_check()
778 sibling = list_entry(next, struct vmap_area, list); in merge_or_add_vmap_area()
[all …]
Dswapfile.c398 static inline bool cluster_list_empty(struct swap_cluster_list *list) in cluster_list_empty() argument
400 return cluster_is_null(&list->head); in cluster_list_empty()
403 static inline unsigned int cluster_list_first(struct swap_cluster_list *list) in cluster_list_first() argument
405 return cluster_next(&list->head); in cluster_list_first()
408 static void cluster_list_init(struct swap_cluster_list *list) in cluster_list_init() argument
410 cluster_set_null(&list->head); in cluster_list_init()
411 cluster_set_null(&list->tail); in cluster_list_init()
414 static void cluster_list_add_tail(struct swap_cluster_list *list, in cluster_list_add_tail() argument
418 if (cluster_list_empty(list)) { in cluster_list_add_tail()
419 cluster_set_next_flag(&list->head, idx, 0); in cluster_list_add_tail()
[all …]
Dfrontswap.c124 plist_for_each_entry(si, &swap_active_head, list) { in frontswap_register_ops()
146 plist_for_each_entry(si, &swap_active_head, list) { in frontswap_register_ops()
377 plist_for_each_entry(si, &swap_active_head, list) in __frontswap_curr_pages()
392 plist_for_each_entry(si, &swap_active_head, list) { in __frontswap_unuse_pages()
Dzsmalloc.c289 struct list_head list; /* fullness list */ member
724 struct zspage, list); in insert_zspage()
731 list_add(&zspage->list, &head->list); in insert_zspage()
735 list_add(&zspage->list, &class->fullness_list[fullness]); in insert_zspage()
749 list_del_init(&zspage->list); in remove_zspage()
967 VM_BUG_ON(list_empty(&zspage->list)); in free_zspage()
1104 struct zspage, list); in find_get_zspage()
1712 struct zspage, list); in isolate_zspage()
1934 if (list_empty(&zspage->list) && !is_zspage_isolated(zspage)) { in zs_page_isolate()
1943 if (!list_empty(&zspage->list) && !is_zspage_isolated(zspage)) { in zs_page_isolate()
[all …]
Dksm.c162 struct list_head list; member
664 list_del(&stable_node->list); in remove_node_from_stable_tree()
958 list_for_each_entry_safe(stable_node, next, &migrate_nodes, list) { in remove_all_stable_nodes()
1703 list_del(&page_node->list); in stable_tree_search()
1729 list_del(&page_node->list); in stable_tree_search()
1747 list_del(&page_node->list); in stable_tree_search()
1759 list_add(&stable_node_dup->list, stable_node_dup->head); in stable_tree_search()
1792 list_del(&page_node->list); in stable_tree_search()
2055 list_add(&stable_node->list, stable_node->head); in cmp_and_merge_page()
2262 &migrate_nodes, list) { in scan_get_next_rmap_item()
[all …]
Dpercpu-stats.c43 list) in find_max_nr_alloc()
221 list) { in percpu_stats_show()
Dpercpu.c513 list_move(&chunk->list, &pcpu_slot[slot]); in __pcpu_chunk_move()
515 list_move_tail(&chunk->list, &pcpu_slot[slot]); in __pcpu_chunk_move()
1325 INIT_LIST_HEAD(&chunk->list); in pcpu_alloc_first_chunk()
1405 INIT_LIST_HEAD(&chunk->list); in pcpu_alloc_chunk()
1761 list_for_each_entry_safe(chunk, next, &pcpu_slot[slot], list) { in pcpu_alloc()
1959 list_for_each_entry_safe(chunk, next, free_head, list) { in __pcpu_balance_workfn()
1963 if (chunk == list_first_entry(free_head, struct pcpu_chunk, list)) in __pcpu_balance_workfn()
1966 list_move(&chunk->list, &to_free); in __pcpu_balance_workfn()
1971 list_for_each_entry_safe(chunk, next, &to_free, list) { in __pcpu_balance_workfn()
2013 list_for_each_entry(chunk, &pcpu_slot[slot], list) { in __pcpu_balance_workfn()
[all …]
Dpage_alloc.c1451 struct list_head *list; in free_pcppages_bulk() local
1464 list = &pcp->lists[migratetype]; in free_pcppages_bulk()
1465 } while (list_empty(list)); in free_pcppages_bulk()
1472 page = list_last_entry(list, struct page, lru); in free_pcppages_bulk()
1493 } while (--count && --batch_free && !list_empty(list)); in free_pcppages_bulk()
3000 unsigned long count, struct list_head *list, in rmqueue_bulk() argument
3031 list_add_tail(&page->lru, list); in rmqueue_bulk()
3058 struct list_head *list = &pcp->lists[migratetype]; in get_populated_pcp_list() local
3060 if (list_empty(list)) { in get_populated_pcp_list()
3061 trace_android_vh_rmqueue_bulk_bypass(order, pcp, migratetype, list); in get_populated_pcp_list()
[all …]
Dswap.c1110 struct lruvec *lruvec, struct list_head *list) in lru_add_page_tail() argument
1117 if (!list) in lru_add_page_tail()
1122 else if (list) { in lru_add_page_tail()
1125 list_add_tail(&page_tail->lru, list); in lru_add_page_tail()
Dhuge_memory.c2394 struct lruvec *lruvec, struct list_head *list) in __split_huge_page_tail() argument
2455 lru_add_page_tail(head, page_tail, lruvec, list); in __split_huge_page_tail()
2458 static void __split_huge_page(struct page *page, struct list_head *list, in __split_huge_page() argument
2483 __split_huge_page_tail(head, i, lruvec, list); in __split_huge_page()
2662 int split_huge_page_to_list(struct page *page, struct list_head *list) in split_huge_page_to_list() argument
2760 __split_huge_page(page, list, end, flags); in split_huge_page_to_list()
2857 LIST_HEAD(list), *pos, *next; in deferred_split_scan()
2872 list_move(page_deferred_list(page), &list); in deferred_split_scan()
2883 list_for_each_safe(pos, next, &list) { in deferred_split_scan()
2896 list_splice_tail(&list, &ds_queue->split_queue); in deferred_split_scan()
Dcompaction.c92 static void split_map_pages(struct list_head *list) in split_map_pages() argument
98 list_for_each_entry_safe(page, next, list, lru) { in split_map_pages()
114 list_splice(&tmp_list, list); in split_map_pages()
773 struct list_head *list) in isolate_and_split_free_page() argument
787 list_add(&page->lru, list); in isolate_and_split_free_page()
789 split_map_pages(list); in isolate_and_split_free_page()
Dmemcontrol.c126 struct list_head list; member
145 struct list_head list; member
293 list_del(&objcg->list); in obj_cgroup_release()
316 INIT_LIST_HEAD(&objcg->list); in obj_cgroup_alloc()
332 list_add(&objcg->list, &parent->objcg_list); in memcg_reparent_objcgs()
335 list_for_each_entry(iter, &memcg->objcg_list, list) { in memcg_reparent_objcgs()
4310 list_for_each_entry(ev, &memcg->oom_notify, list) in mem_cgroup_oom_notify_cb()
4526 list_add(&event->list, &memcg->oom_notify); in mem_cgroup_oom_register_event()
4543 list_for_each_entry_safe(ev, tmp, &memcg->oom_notify, list) { in mem_cgroup_oom_unregister_event()
4545 list_del(&ev->list); in mem_cgroup_oom_unregister_event()
[all …]
Dpercpu-internal.h59 struct list_head list; /* linked to pcpu_slot lists */ member
Dslob.c113 static void set_slob_page_free(struct page *sp, struct list_head *list) in set_slob_page_free() argument
115 list_add(&sp->slab_list, list); in set_slob_page_free()
/mm/kfence/
Dcore.c267 meta = list_entry(kfence_freelist.next, struct kfence_metadata, list); in kfence_guarded_alloc()
268 list_del_init(&meta->list); in kfence_guarded_alloc()
286 list_add_tail(&meta->list, &kfence_freelist); in kfence_guarded_alloc()
405 KFENCE_WARN_ON(!list_empty(&meta->list)); in kfence_guarded_free()
406 list_add_tail(&meta->list, &kfence_freelist); in kfence_guarded_free()
474 INIT_LIST_HEAD(&meta->list); in kfence_init_pool()
478 list_add_tail(&meta->list, &kfence_freelist); in kfence_init_pool()
Dkfence.h45 struct list_head list; /* Freelist node; access under kfence_freelist_lock. */ member
/mm/damon/
Dcore.c44 INIT_LIST_HEAD(&region->list); in damon_new_region()
54 list_add_tail(&r->list, &t->regions_list); in damon_add_region()
60 list_del(&r->list); in damon_del_region()
95 INIT_LIST_HEAD(&scheme->list); in damon_new_scheme()
123 list_add_tail(&s->list, &ctx->schemes); in damon_add_scheme()
128 list_del(&s->list); in damon_del_scheme()
158 INIT_LIST_HEAD(&t->list); in damon_new_target()
165 list_add_tail(&t->list, &ctx->adaptive_targets); in damon_add_target()
175 list_del(&t->list); in damon_del_target()

12