/mm/ |
D | list_lru.c | 21 static void list_lru_register(struct list_lru *lru) in list_lru_register() argument 24 list_add(&lru->list, &list_lrus); in list_lru_register() 28 static void list_lru_unregister(struct list_lru *lru) in list_lru_unregister() argument 31 list_del(&lru->list); in list_lru_unregister() 35 static int lru_shrinker_id(struct list_lru *lru) in lru_shrinker_id() argument 37 return lru->shrinker_id; in lru_shrinker_id() 40 static inline bool list_lru_memcg_aware(struct list_lru *lru) in list_lru_memcg_aware() argument 42 return lru->memcg_aware; in list_lru_memcg_aware() 56 return memcg_lrus->lru[idx]; in list_lru_from_memcg_idx() 57 return &nlru->lru; in list_lru_from_memcg_idx() [all …]
|
D | zbud.c | 111 struct list_head lru; member 128 struct list_head lru; member 159 INIT_LIST_HEAD(&zhdr->lru); in init_zbud_page() 232 INIT_LIST_HEAD(&pool->lru); in zbud_create_pool() 323 if (!list_empty(&zhdr->lru)) in zbud_alloc() 324 list_del(&zhdr->lru); in zbud_alloc() 325 list_add(&zhdr->lru, &pool->lru); in zbud_alloc() 368 list_del(&zhdr->lru); in zbud_free() 422 if (!pool->ops || !pool->ops->evict || list_empty(&pool->lru) || in zbud_reclaim_page() 428 zhdr = list_last_entry(&pool->lru, struct zbud_header, lru); in zbud_reclaim_page() [all …]
|
D | balloon_compaction.c | 48 list_for_each_entry_safe(page, tmp, pages, lru) { in balloon_page_list_enqueue() 49 list_del(&page->lru); in balloon_page_list_enqueue() 84 list_for_each_entry_safe(page, tmp, &b_dev_info->pages, lru) { in balloon_page_list_dequeue() 104 list_add(&page->lru, pages); in balloon_page_list_dequeue() 200 return list_first_entry(&pages, struct page, lru); in balloon_page_dequeue() 213 list_del(&page->lru); in balloon_page_isolate() 226 list_add(&page->lru, &b_dev_info->pages); in balloon_page_putback()
|
D | vmscan.c | 181 if ((_page)->lru.prev != _base) { \ 184 prev = lru_to_page(&(_page->lru)); \ 619 static unsigned long lruvec_lru_size(struct lruvec *lruvec, enum lru_list lru, in lruvec_lru_size() argument 632 size += mem_cgroup_get_zone_lru_size(lruvec, lru, zid); in lruvec_lru_size() 634 size += zone_page_state(zone, NR_ZONE_LRU_BASE + lru); in lruvec_lru_size() 1434 list_del(&page->lru); in shrink_page_list() 1558 list_add_tail(&page->lru, page_list); in shrink_page_list() 1583 list_add(&page->lru, &demote_pages); in shrink_page_list() 1812 list_add(&page->lru, &free_pages); in shrink_page_list() 1839 list_add(&page->lru, &ret_pages); in shrink_page_list() [all …]
|
D | page_reporting.c | 152 list_for_each_entry_safe(page, next, list, lru) { in page_reporting_cycle() 191 if (!list_is_first(&page->lru, list)) in page_reporting_cycle() 192 list_rotate_to_front(&page->lru, list); in page_reporting_cycle() 216 next = list_first_entry(list, struct page, lru); in page_reporting_cycle() 224 if (!list_entry_is_head(next, list, lru) && !list_is_first(&next->lru, list)) in page_reporting_cycle() 225 list_rotate_to_front(&next->lru, list); in page_reporting_cycle()
|
D | z3fold.c | 162 struct list_head lru; member 391 INIT_LIST_HEAD(&page->lru); in init_z3fold_page() 513 if (!list_empty(&page->lru)) in __release_z3fold_page() 514 list_del_init(&page->lru); in __release_z3fold_page() 999 INIT_LIST_HEAD(&pool->lru); in z3fold_create_pool() 1174 if (!list_empty(&page->lru)) in z3fold_alloc() 1175 list_del(&page->lru); in z3fold_alloc() 1177 list_add(&page->lru, &pool->lru); in z3fold_alloc() 1216 list_del(&page->lru); in z3fold_free() 1331 if (list_empty(&pool->lru)) { in z3fold_reclaim_page() [all …]
|
D | pgtable-generic.c | 168 INIT_LIST_HEAD(&pgtable->lru); in pgtable_trans_huge_deposit() 170 list_add(&pgtable->lru, &pmd_huge_pte(mm, pmdp)->lru); in pgtable_trans_huge_deposit() 185 pmd_huge_pte(mm, pmdp) = list_first_entry_or_null(&pgtable->lru, in pgtable_trans_huge_withdraw() 186 struct page, lru); in pgtable_trans_huge_withdraw() 188 list_del(&pgtable->lru); in pgtable_trans_huge_withdraw()
|
D | sparse-vmemmap.c | 227 list_for_each_entry_safe(page, next, list, lru) { in free_vmemmap_page_list() 228 list_del(&page->lru); in free_vmemmap_page_list() 244 list_add_tail(&page->lru, walk->vmemmap_pages); in vmemmap_remap_pte() 257 page = list_first_entry(walk->vmemmap_pages, struct page, lru); in vmemmap_restore_pte() 258 list_del(&page->lru); in vmemmap_restore_pte() 341 list_add_tail(&page->lru, list); in alloc_vmemmap_page_list() 346 list_for_each_entry_safe(page, next, list, lru) in alloc_vmemmap_page_list()
|
D | mmzone.c | 78 enum lru_list lru; in lruvec_init() local 83 for_each_lru(lru) in lruvec_init() 84 INIT_LIST_HEAD(&lruvec->lists[lru]); in lruvec_init()
|
D | compaction.c | 78 list_for_each_entry_safe(page, next, freelist, lru) { in release_freepages() 80 list_del(&page->lru); in release_freepages() 95 list_for_each_entry_safe(page, next, list, lru) { in split_map_pages() 96 list_del(&page->lru); in split_map_pages() 106 list_add(&page->lru, &tmp_list); in split_map_pages() 633 list_add_tail(&page->lru, freelist); in isolate_freepages_block() 779 list_add(&page->lru, list); in isolate_and_split_free_page() 1132 list_add(&page->lru, &cc->migratepages); in isolate_migratepages_block() 1366 if (!list_is_last(freelist, &freepage->lru)) { in move_freelist_head() 1367 list_cut_before(&sublist, freelist, &freepage->lru); in move_freelist_head() [all …]
|
D | memory-failure.c | 1093 #define lru (1UL << PG_lru) macro 1123 { lru|dirty, lru|dirty, MF_MSG_DIRTY_LRU, me_pagecache_dirty }, 1124 { lru|dirty, lru, MF_MSG_CLEAN_LRU, me_pagecache_clean }, 1136 #undef lru 2106 bool lru = PageLRU(page); in isolate_page() local 2111 if (lru) in isolate_page() 2117 list_add(&page->lru, pagelist); in isolate_page() 2120 if (isolated && lru) in isolate_page()
|
D | khugepaged.c | 584 list_for_each_entry_safe(page, tmp, compound_pagelist, lru) { in release_pte_pages() 585 list_del(&page->lru); in release_pte_pages() 654 list_for_each_entry(p, compound_pagelist, lru) { in __collapse_huge_page_isolate() 714 list_add_tail(&page->lru, compound_pagelist); in __collapse_huge_page_isolate() 792 list_for_each_entry_safe(src_page, tmp, compound_pagelist, lru) { in __collapse_huge_page_copy() 793 list_del(&src_page->lru); in __collapse_huge_page_copy() 1898 list_add_tail(&page->lru, &pagelist); in collapse_file() 1948 list_for_each_entry_safe(page, tmp, &pagelist, lru) { in collapse_file() 1955 list_del(&page->lru); in collapse_file() 1995 struct page, lru); in collapse_file() [all …]
|
D | shuffle.c | 136 list_swap(&page_i->lru, &page_j->lru); in __shuffle_zone()
|
D | migrate.c | 144 list_for_each_entry_safe(page, page2, l, lru) { in putback_movable_pages() 149 list_del(&page->lru); in putback_movable_pages() 1226 list_del(&page->lru); in unmap_and_move() 1251 list_add_tail(&page->lru, ret); in unmap_and_move() 1300 list_move_tail(&hpage->lru, ret); in unmap_and_move_huge_page() 1394 list_move_tail(&hpage->lru, ret); in unmap_and_move_huge_page() 1418 list_safe_reset_next(page, *page2, lru); in try_split_thp() 1475 list_for_each_entry_safe(page, page2, from, lru) { in migrate_pages() 1740 list_add_tail(&head->lru, pagelist); in add_page_for_migration() 2203 list_add(&page->lru, &migratepages); in migrate_misplaced_page() [all …]
|
D | swapfile.c | 1550 page = list_next_entry(page, lru); in swp_swapcount() 3632 INIT_LIST_HEAD(&head->lru); in add_swap_count_continuation() 3637 list_for_each_entry(list_page, &head->lru, lru) { in add_swap_count_continuation() 3659 list_add_tail(&page->lru, &head->lru); in add_swap_count_continuation() 3698 page = list_next_entry(head, lru); in swap_count_continued() 3710 page = list_next_entry(page, lru); in swap_count_continued() 3716 page = list_next_entry(page, lru); in swap_count_continued() 3726 while ((page = list_prev_entry(page, lru)) != head) { in swap_count_continued() 3740 page = list_next_entry(page, lru); in swap_count_continued() 3749 while ((page = list_prev_entry(page, lru)) != head) { in swap_count_continued() [all …]
|
D | readahead.c | 73 list_del(&victim->lru); in read_cache_pages_invalidate_pages() 98 list_del(&page->lru); in read_cache_pages() 231 list_add(&page->lru, &page_pool); in page_cache_ra_unbounded()
|
D | hugetlb.c | 1079 list_move(&page->lru, &h->hugepage_freelists[nid]); in enqueue_huge_page() 1091 list_for_each_entry(page, &h->hugepage_freelists[nid], lru) { in dequeue_huge_page_node_exact() 1098 list_move(&page->lru, &h->hugepage_activelist); in dequeue_huge_page_node_exact() 1376 list_del(&page->lru); in remove_hugetlb_page() 1424 INIT_LIST_HEAD(&page->lru); in add_hugetlb_page() 1562 list_for_each_entry_safe(page, t_page, list, lru) { in update_and_free_pages_bulk() 1656 INIT_LIST_HEAD(&page->lru); in __prep_new_huge_page() 1933 struct page, lru); in remove_pool_huge_page() 2253 list_add(&page->lru, &surplus_list); in gather_surplus_pages() 2288 list_for_each_entry_safe(page, tmp, &surplus_list, lru) { in gather_surplus_pages() [all …]
|
D | page_alloc.c | 1604 list_add_tail(&page->lru, &head); in free_pcppages_bulk() 1631 list_for_each_entry_safe(page, tmp, &head, lru) { in free_pcppages_bulk() 1676 INIT_LIST_HEAD(&page->lru); in __init_single_page() 1728 INIT_LIST_HEAD(&page->lru); in reserve_bootmem_region() 3646 list_for_each_entry_safe(page, next, list, lru) { in free_unref_page_list() 3649 list_del(&page->lru); in free_unref_page_list() 3662 list_del(&page->lru); in free_unref_page_list() 3670 list_for_each_entry_safe(page, next, list, lru) { in free_unref_page_list() 5622 list_add(&page->lru, page_list); in __alloc_pages_bulk() 5644 list_add(&page->lru, page_list); in __alloc_pages_bulk() [all …]
|
D | workingset.c | 625 struct list_lru_one *lru, in shadow_lru_isolate() argument 654 list_lru_isolate(lru, item); in shadow_lru_isolate()
|
D | bootmem_info.c | 37 INIT_LIST_HEAD(&page->lru); in put_page_bootmem()
|
D | memcontrol.c | 1303 void mem_cgroup_update_lru_size(struct lruvec *lruvec, enum lru_list lru, in mem_cgroup_update_lru_size() argument 1314 lru_size = &mz->lru_zone_size[zid][lru]; in mem_cgroup_update_lru_size() 1322 __func__, lruvec, lru, nr_pages, size)) { in mem_cgroup_update_lru_size() 3930 enum lru_list lru; in mem_cgroup_node_nr_lru_pages() local 3934 for_each_lru(lru) { in mem_cgroup_node_nr_lru_pages() 3935 if (!(BIT(lru) & lru_mask)) in mem_cgroup_node_nr_lru_pages() 3938 nr += lruvec_page_state(lruvec, NR_LRU_BASE + lru); in mem_cgroup_node_nr_lru_pages() 3940 nr += lruvec_page_state_local(lruvec, NR_LRU_BASE + lru); in mem_cgroup_node_nr_lru_pages() 3950 enum lru_list lru; in mem_cgroup_nr_lru_pages() local 3952 for_each_lru(lru) { in mem_cgroup_nr_lru_pages() [all …]
|
D | swap.c | 149 list_del(&victim->lru); in put_pages_list() 1024 list_add(&page->lru, &pages_to_free); in release_pages()
|
D | hugetlb_cgroup.c | 208 list_for_each_entry(page, &h->hugepage_activelist, lru) in hugetlb_cgroup_css_offline() 796 list_move(&newhpage->lru, &h->hugepage_activelist); in hugetlb_cgroup_migrate()
|
D | memory_hotplug.c | 1787 list_add_tail(&page->lru, &source); in do_migrate_range() 1811 mtc.nid = page_to_nid(list_first_entry(&source, struct page, lru)); in do_migrate_range() 1824 list_for_each_entry(page, &source, lru) { in do_migrate_range()
|
/mm/damon/ |
D | paddr.c | 241 list_add(&page->lru, &page_list); in damon_pa_apply_scheme()
|