• Home
  • Raw
  • Download

Lines Matching refs:lruvec

342 unsigned long lruvec_lru_size(struct lruvec *lruvec, enum lru_list lru, int zone_idx)  in lruvec_lru_size()  argument
348 struct zone *zone = &lruvec_pgdat(lruvec)->node_zones[zid]; in lruvec_lru_size()
354 size += mem_cgroup_get_zone_lru_size(lruvec, lru, zid); in lruvec_lru_size()
1683 static __always_inline void update_lru_sizes(struct lruvec *lruvec, in update_lru_sizes() argument
1692 update_lru_size(lruvec, lru, zid, -nr_zone_taken[zid]); in update_lru_sizes()
1717 struct lruvec *lruvec, struct list_head *dst, in isolate_lru_pages() argument
1721 struct list_head *src = &lruvec->lists[lru]; in isolate_lru_pages()
1800 update_lru_sizes(lruvec, lru, nr_zone_taken); in isolate_lru_pages()
1839 struct lruvec *lruvec; in isolate_lru_page() local
1842 lruvec = mem_cgroup_page_lruvec(page, pgdat); in isolate_lru_page()
1847 del_page_from_lru_list(page, lruvec, lru); in isolate_lru_page()
1912 static unsigned noinline_for_stack move_pages_to_lru(struct lruvec *lruvec, in move_pages_to_lru() argument
1915 struct pglist_data *pgdat = lruvec_pgdat(lruvec); in move_pages_to_lru()
1931 lruvec = mem_cgroup_page_lruvec(page, pgdat); in move_pages_to_lru()
1937 update_lru_size(lruvec, lru, page_zonenum(page), nr_pages); in move_pages_to_lru()
1938 list_move(&page->lru, &lruvec->lists[lru]); in move_pages_to_lru()
1944 del_page_from_lru_list(page, lruvec, lru); in move_pages_to_lru()
1955 workingset_age_nonresident(lruvec, nr_pages); in move_pages_to_lru()
1985 shrink_inactive_list(unsigned long nr_to_scan, struct lruvec *lruvec, in shrink_inactive_list() argument
1995 struct pglist_data *pgdat = lruvec_pgdat(lruvec); in shrink_inactive_list()
2015 nr_taken = isolate_lru_pages(nr_to_scan, lruvec, &page_list, in shrink_inactive_list()
2022 __count_memcg_events(lruvec_memcg(lruvec), item, nr_scanned); in shrink_inactive_list()
2035 move_pages_to_lru(lruvec, &page_list); in shrink_inactive_list()
2038 lru_note_cost(lruvec, file, stat.nr_pageout); in shrink_inactive_list()
2042 __count_memcg_events(lruvec_memcg(lruvec), item, nr_reclaimed); in shrink_inactive_list()
2078 struct lruvec *lruvec, in shrink_active_list() argument
2092 struct pglist_data *pgdat = lruvec_pgdat(lruvec); in shrink_active_list()
2100 nr_taken = isolate_lru_pages(nr_to_scan, lruvec, &l_hold, in shrink_active_list()
2107 __count_memcg_events(lruvec_memcg(lruvec), PGREFILL, nr_scanned); in shrink_active_list()
2171 nr_activate = move_pages_to_lru(lruvec, &l_active); in shrink_active_list()
2172 nr_deactivate = move_pages_to_lru(lruvec, &l_inactive); in shrink_active_list()
2177 __count_memcg_events(lruvec_memcg(lruvec), PGDEACTIVATE, nr_deactivate); in shrink_active_list()
2244 struct lruvec *lruvec, struct scan_control *sc) in shrink_list() argument
2248 shrink_active_list(nr_to_scan, lruvec, sc, lru); in shrink_list()
2254 return shrink_inactive_list(nr_to_scan, lruvec, sc, lru); in shrink_list()
2285 static bool inactive_is_low(struct lruvec *lruvec, enum lru_list inactive_lru) in inactive_is_low() argument
2293 inactive = lruvec_page_state(lruvec, NR_LRU_BASE + inactive_lru); in inactive_is_low()
2294 active = lruvec_page_state(lruvec, NR_LRU_BASE + active_lru); in inactive_is_low()
2328 static void get_scan_count(struct lruvec *lruvec, struct scan_control *sc, in get_scan_count() argument
2331 struct mem_cgroup *memcg = lruvec_memcg(lruvec); in get_scan_count()
2429 lruvec_size = lruvec_lru_size(lruvec, lru, sc->reclaim_idx); in get_scan_count()
2531 static void shrink_lruvec(struct lruvec *lruvec, struct scan_control *sc) in shrink_lruvec() argument
2542 get_scan_count(lruvec, sc, nr); in shrink_lruvec()
2573 lruvec, sc); in shrink_lruvec()
2638 if (total_swap_pages && inactive_is_low(lruvec, LRU_INACTIVE_ANON)) in shrink_lruvec()
2639 shrink_active_list(SWAP_CLUSTER_MAX, lruvec, in shrink_lruvec()
2721 struct lruvec *lruvec = mem_cgroup_lruvec(memcg, pgdat); in shrink_node_memcgs() local
2763 shrink_lruvec(lruvec, sc); in shrink_node_memcgs()
2780 struct lruvec *target_lruvec; in shrink_node()
3095 struct lruvec *target_lruvec; in snapshot_refaults()
3165 struct lruvec *lruvec; in do_try_to_free_pages() local
3167 lruvec = mem_cgroup_lruvec(sc->target_mem_cgroup, in do_try_to_free_pages()
3169 clear_bit(LRUVEC_CONGESTED, &lruvec->flags); in do_try_to_free_pages()
3394 struct lruvec *lruvec = mem_cgroup_lruvec(memcg, pgdat); in mem_cgroup_shrink_node() local
3419 shrink_lruvec(lruvec, &sc); in mem_cgroup_shrink_node()
3472 struct lruvec *lruvec; in age_active_anon() local
3477 lruvec = mem_cgroup_lruvec(NULL, pgdat); in age_active_anon()
3478 if (!inactive_is_low(lruvec, LRU_INACTIVE_ANON)) in age_active_anon()
3483 lruvec = mem_cgroup_lruvec(memcg, pgdat); in age_active_anon()
3484 shrink_active_list(SWAP_CLUSTER_MAX, lruvec, in age_active_anon()
3553 struct lruvec *lruvec = mem_cgroup_lruvec(NULL, pgdat); in clear_pgdat_congested() local
3555 clear_bit(LRUVEC_CONGESTED, &lruvec->flags); in clear_pgdat_congested()
4431 struct lruvec *lruvec; in check_move_unevictable_pages() local
4454 lruvec = mem_cgroup_page_lruvec(page, pgdat); in check_move_unevictable_pages()
4464 del_page_from_lru_list(page, lruvec, LRU_UNEVICTABLE); in check_move_unevictable_pages()
4465 add_page_to_lru_list(page, lruvec, lru); in check_move_unevictable_pages()