/mm/ |
D | swap.c | 64 struct lruvec *lruvec; in __page_cache_release() local 68 lruvec = mem_cgroup_page_lruvec(page, pgdat); in __page_cache_release() 71 del_page_from_lru_list(page, lruvec, page_off_lru(page)); in __page_cache_release() 191 void (*move_fn)(struct page *page, struct lruvec *lruvec, void *arg), in pagevec_lru_move_fn() argument 196 struct lruvec *lruvec; in pagevec_lru_move_fn() local 210 lruvec = mem_cgroup_page_lruvec(page, pgdat); in pagevec_lru_move_fn() 211 (*move_fn)(page, lruvec, arg); in pagevec_lru_move_fn() 219 static void pagevec_move_tail_fn(struct page *page, struct lruvec *lruvec, in pagevec_move_tail_fn() argument 225 del_page_from_lru_list(page, lruvec, page_lru(page)); in pagevec_move_tail_fn() 227 add_page_to_lru_list_tail(page, lruvec, page_lru(page)); in pagevec_move_tail_fn() [all …]
|
D | workingset.c | 229 struct lruvec *lruvec; in workingset_eviction() local 236 lruvec = mem_cgroup_lruvec(pgdat, memcg); in workingset_eviction() 237 eviction = atomic_long_inc_return(&lruvec->inactive_age); in workingset_eviction() 256 struct lruvec *lruvec; in workingset_refault() local 283 lruvec = mem_cgroup_lruvec(pgdat, memcg); in workingset_refault() 284 refault = atomic_long_read(&lruvec->inactive_age); in workingset_refault() 285 active_file = lruvec_lru_size(lruvec, LRU_ACTIVE_FILE, MAX_NR_ZONES); in workingset_refault() 305 inc_lruvec_state(lruvec, WORKINGSET_REFAULT); in workingset_refault() 316 atomic_long_inc(&lruvec->inactive_age); in workingset_refault() 317 inc_lruvec_state(lruvec, WORKINGSET_ACTIVATE); in workingset_refault() [all …]
|
D | vmscan.c | 352 unsigned long lruvec_lru_size(struct lruvec *lruvec, enum lru_list lru, int zone_idx) in lruvec_lru_size() argument 359 lru_size += mem_cgroup_get_zone_lru_size(lruvec, lru, zid); in lruvec_lru_size() 361 lru_size = node_page_state(lruvec_pgdat(lruvec), NR_LRU_BASE + lru); in lruvec_lru_size() 364 struct zone *zone = &lruvec_pgdat(lruvec)->node_zones[zid]; in lruvec_lru_size() 371 size = mem_cgroup_get_zone_lru_size(lruvec, lru, zid); in lruvec_lru_size() 373 size = zone_page_state(&lruvec_pgdat(lruvec)->node_zones[zid], in lruvec_lru_size() 1646 static __always_inline void update_lru_sizes(struct lruvec *lruvec, in update_lru_sizes() argument 1655 __update_lru_size(lruvec, lru, zid, -nr_zone_taken[zid]); in update_lru_sizes() 1657 mem_cgroup_update_lru_size(lruvec, lru, zid, -nr_zone_taken[zid]); in update_lru_sizes() 1684 struct lruvec *lruvec, struct list_head *dst, in isolate_lru_pages() argument [all …]
|
D | mmzone.c | 89 void lruvec_init(struct lruvec *lruvec) in lruvec_init() argument 93 memset(lruvec, 0, sizeof(struct lruvec)); in lruvec_init() 96 INIT_LIST_HEAD(&lruvec->lists[lru]); in lruvec_init()
|
D | slab.h | 352 struct lruvec *lruvec; in memcg_charge_slab() local 372 lruvec = mem_cgroup_lruvec(page_pgdat(page), memcg); in memcg_charge_slab() 373 mod_lruvec_state(lruvec, cache_vmstat_idx(s), 1 << order); in memcg_charge_slab() 391 struct lruvec *lruvec; in memcg_uncharge_slab() local 396 lruvec = mem_cgroup_lruvec(page_pgdat(page), memcg); in memcg_uncharge_slab() 397 mod_lruvec_state(lruvec, cache_vmstat_idx(s), -(1 << order)); in memcg_uncharge_slab()
|
D | memcontrol.c | 733 void __mod_lruvec_state(struct lruvec *lruvec, enum node_stat_item idx, in __mod_lruvec_state() argument 736 pg_data_t *pgdat = lruvec_pgdat(lruvec); in __mod_lruvec_state() 747 pn = container_of(lruvec, struct mem_cgroup_per_node, lruvec); in __mod_lruvec_state() 772 struct lruvec *lruvec; in __mod_lruvec_slab_state() local 781 lruvec = mem_cgroup_lruvec(pgdat, memcg); in __mod_lruvec_slab_state() 782 __mod_lruvec_state(lruvec, idx, val); in __mod_lruvec_slab_state() 1234 struct lruvec *mem_cgroup_page_lruvec(struct page *page, struct pglist_data *pgdat) in mem_cgroup_page_lruvec() 1238 struct lruvec *lruvec; in mem_cgroup_page_lruvec() local 1241 lruvec = &pgdat->lruvec; in mem_cgroup_page_lruvec() 1254 lruvec = &mz->lruvec; in mem_cgroup_page_lruvec() [all …]
|
D | mlock.c | 112 struct lruvec *lruvec; in __munlock_isolate_lru_page() local 114 lruvec = mem_cgroup_page_lruvec(page, page_pgdat(page)); in __munlock_isolate_lru_page() 118 del_page_from_lru_list(page, lruvec, page_lru(page)); in __munlock_isolate_lru_page()
|
D | huge_memory.c | 2446 struct lruvec *lruvec, struct list_head *list) in __split_huge_page_tail() argument 2504 lru_add_page_tail(head, page_tail, lruvec, list); in __split_huge_page_tail() 2512 struct lruvec *lruvec; in __split_huge_page() local 2517 lruvec = mem_cgroup_page_lruvec(head, pgdat); in __split_huge_page() 2531 __split_huge_page_tail(head, i, lruvec, list); in __split_huge_page()
|
D | compaction.c | 787 struct lruvec *lruvec; in isolate_migratepages_block() local 978 lruvec = mem_cgroup_page_lruvec(page, pgdat); in isolate_migratepages_block() 987 del_page_from_lru_list(page, lruvec, page_lru(page)); in isolate_migratepages_block()
|
D | page-writeback.c | 2718 struct lruvec *lruvec; in test_clear_page_writeback() local 2722 lruvec = mem_cgroup_page_lruvec(page, page_pgdat(page)); in test_clear_page_writeback() 2756 dec_lruvec_state(lruvec, NR_WRITEBACK); in test_clear_page_writeback()
|