/mm/ |
D | vmstat.c | 246 struct pglist_data *pgdat; in refresh_zone_stat_thresholds() local 252 for_each_online_pgdat(pgdat) { in refresh_zone_stat_thresholds() 254 per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold = 0; in refresh_zone_stat_thresholds() 259 struct pglist_data *pgdat = zone->zone_pgdat; in refresh_zone_stat_thresholds() local 271 pgdat_threshold = per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold; in refresh_zone_stat_thresholds() 272 per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold in refresh_zone_stat_thresholds() 289 void set_pgdat_percpu_threshold(pg_data_t *pgdat, in set_pgdat_percpu_threshold() argument 297 for (i = 0; i < pgdat->nr_zones; i++) { in set_pgdat_percpu_threshold() 298 zone = &pgdat->node_zones[i]; in set_pgdat_percpu_threshold() 347 void __mod_node_page_state(struct pglist_data *pgdat, enum node_stat_item item, in __mod_node_page_state() argument [all …]
|
D | vmscan.c | 1380 struct pglist_data *pgdat) in demote_page_list() argument 1382 int target_nid = next_demotion_node(pgdat->node_id); in demote_page_list() 1409 struct pglist_data *pgdat, in shrink_page_list() argument 1424 do_demote_pass = can_demote(pgdat->node_id, sc); in shrink_page_list() 1534 test_bit(PGDAT_WRITEBACK, &pgdat->flags)) { in shrink_page_list() 1687 !test_bit(PGDAT_DIRTY, &pgdat->flags))) { in shrink_page_list() 1866 nr_reclaimed += demote_page_list(&demote_pages, pgdat); in shrink_page_list() 2142 static int too_many_isolated(struct pglist_data *pgdat, int file, in too_many_isolated() argument 2154 inactive = node_page_state(pgdat, NR_INACTIVE_FILE); in too_many_isolated() 2155 isolated = node_page_state(pgdat, NR_ISOLATED_FILE); in too_many_isolated() [all …]
|
D | compaction.c | 388 void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() argument 393 struct zone *zone = &pgdat->node_zones[zoneid]; in reset_isolation_suitable() 792 static bool too_many_isolated(pg_data_t *pgdat) in too_many_isolated() argument 796 inactive = node_page_state(pgdat, NR_INACTIVE_FILE) + in too_many_isolated() 797 node_page_state(pgdat, NR_INACTIVE_ANON); in too_many_isolated() 798 active = node_page_state(pgdat, NR_ACTIVE_FILE) + in too_many_isolated() 799 node_page_state(pgdat, NR_ACTIVE_ANON); in too_many_isolated() 800 isolated = node_page_state(pgdat, NR_ISOLATED_FILE) + in too_many_isolated() 801 node_page_state(pgdat, NR_ISOLATED_ANON); in too_many_isolated() 828 pg_data_t *pgdat = cc->zone->zone_pgdat; in isolate_migratepages_block() local [all …]
|
D | memory_hotplug.c | 433 static void update_pgdat_span(struct pglist_data *pgdat) in update_pgdat_span() argument 438 for (zone = pgdat->node_zones; in update_pgdat_span() 439 zone < pgdat->node_zones + MAX_NR_ZONES; zone++) { in update_pgdat_span() 457 pgdat->node_start_pfn = node_start_pfn; in update_pgdat_span() 458 pgdat->node_spanned_pages = node_end_pfn - node_start_pfn; in update_pgdat_span() 466 struct pglist_data *pgdat = zone->zone_pgdat; in remove_pfn_range_from_zone() local 491 update_pgdat_span(pgdat); in remove_pfn_range_from_zone() 664 static void __meminit resize_pgdat_range(struct pglist_data *pgdat, unsigned long start_pfn, in resize_pgdat_range() argument 667 unsigned long old_end_pfn = pgdat_end_pfn(pgdat); in resize_pgdat_range() 669 if (!pgdat->node_spanned_pages || start_pfn < pgdat->node_start_pfn) in resize_pgdat_range() [all …]
|
D | workingset.c | 187 static void *pack_shadow(int memcgid, pg_data_t *pgdat, unsigned long eviction, in pack_shadow() argument 192 eviction = (eviction << NODES_SHIFT) | pgdat->node_id; in pack_shadow() 198 static void unpack_shadow(void *shadow, int *memcgidp, pg_data_t **pgdat, in unpack_shadow() argument 213 *pgdat = NODE_DATA(nid); in unpack_shadow() 232 struct pglist_data *pgdat = page_pgdat(page); in lru_gen_eviction() local 236 lruvec = mem_cgroup_lruvec(memcg, pgdat); in lru_gen_eviction() 244 return pack_shadow(mem_cgroup_id(memcg), pgdat, token, refs); in lru_gen_eviction() 257 struct pglist_data *pgdat; in lru_gen_refault() local 261 unpack_shadow(shadow, &memcg_id, &pgdat, &token, &workingset); in lru_gen_refault() 263 if (pgdat != page_pgdat(page)) in lru_gen_refault() [all …]
|
D | page_alloc.c | 1709 pg_data_t *pgdat; in init_reserved_page() local 1716 pgdat = NODE_DATA(nid); in init_reserved_page() 1719 struct zone *zone = &pgdat->node_zones[zid]; in init_reserved_page() 2172 pg_data_t *pgdat = data; in deferred_init_memmap() local 2173 const struct cpumask *cpumask = cpumask_of_node(pgdat->node_id); in deferred_init_memmap() 2185 pgdat_resize_lock(pgdat, &flags); in deferred_init_memmap() 2186 first_init_pfn = pgdat->first_deferred_pfn; in deferred_init_memmap() 2188 pgdat_resize_unlock(pgdat, &flags); in deferred_init_memmap() 2194 BUG_ON(pgdat->first_deferred_pfn < pgdat->node_start_pfn); in deferred_init_memmap() 2195 BUG_ON(pgdat->first_deferred_pfn > pgdat_end_pfn(pgdat)); in deferred_init_memmap() [all …]
|
D | mmzone.c | 19 struct pglist_data *next_online_pgdat(struct pglist_data *pgdat) in next_online_pgdat() argument 21 int nid = next_online_node(pgdat->node_id); in next_online_pgdat() 33 pg_data_t *pgdat = zone->zone_pgdat; in next_zone() local 35 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone() 38 pgdat = next_online_pgdat(pgdat); in next_zone() 39 if (pgdat) in next_zone() 40 zone = pgdat->node_zones; in next_zone()
|
D | shuffle.h | 11 extern void __shuffle_free_memory(pg_data_t *pgdat); 13 static inline void __meminit shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory() argument 17 __shuffle_free_memory(pgdat); in shuffle_free_memory() 40 static inline void shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory() argument
|
D | bootmem_info.c | 103 void __init register_page_bootmem_info_node(struct pglist_data *pgdat) in register_page_bootmem_info_node() argument 106 int node = pgdat->node_id; in register_page_bootmem_info_node() 110 page = virt_to_page(pgdat); in register_page_bootmem_info_node() 115 pfn = pgdat->node_start_pfn; in register_page_bootmem_info_node() 116 end_pfn = pgdat_end_pfn(pgdat); in register_page_bootmem_info_node()
|
D | page_owner.c | 284 pg_data_t *pgdat, struct zone *zone) in pagetypeinfo_showmixedcount_print() argument 360 seq_printf(m, "Node %d, zone %8s ", pgdat->node_id, zone->name); in pagetypeinfo_showmixedcount_print() 587 static void init_pages_in_zone(pg_data_t *pgdat, struct zone *zone) in init_pages_in_zone() argument 653 pgdat->node_id, zone->name, count); in init_pages_in_zone() 656 static void init_zones_in_node(pg_data_t *pgdat) in init_zones_in_node() argument 659 struct zone *node_zones = pgdat->node_zones; in init_zones_in_node() 665 init_pages_in_zone(pgdat, zone); in init_zones_in_node() 671 pg_data_t *pgdat; in init_early_allocated_pages() local 673 for_each_online_pgdat(pgdat) in init_early_allocated_pages() 674 init_zones_in_node(pgdat); in init_early_allocated_pages()
|
D | sparse.c | 321 static inline phys_addr_t pgdat_to_phys(struct pglist_data *pgdat) in pgdat_to_phys() argument 324 VM_BUG_ON(pgdat != &contig_page_data); in pgdat_to_phys() 327 return __pa(pgdat); in pgdat_to_phys() 333 sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat, in sparse_early_usemaps_alloc_pgdat_section() argument 349 goal = pgdat_to_phys(pgdat) & (PAGE_SECTION_MASK << PAGE_SHIFT); in sparse_early_usemaps_alloc_pgdat_section() 367 struct pglist_data *pgdat = NODE_DATA(nid); in check_usemap_section_nr() local 377 pgdat_snr = pfn_to_section_nr(pgdat_to_phys(pgdat) >> PAGE_SHIFT); in check_usemap_section_nr() 405 sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat, in sparse_early_usemaps_alloc_pgdat_section() argument 408 return memblock_alloc_node(size, SMP_CACHE_BYTES, pgdat->node_id); in sparse_early_usemaps_alloc_pgdat_section()
|
D | shuffle.c | 154 void __meminit __shuffle_free_memory(pg_data_t *pgdat) in __shuffle_free_memory() argument 158 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in __shuffle_free_memory()
|
D | mm_init.c | 31 pg_data_t *pgdat = NODE_DATA(nid); in mminit_verify_zonelist() local 43 zonelist = &pgdat->node_zonelists[listid]; in mminit_verify_zonelist() 44 zone = &pgdat->node_zones[zoneid]; in mminit_verify_zonelist()
|
D | page-writeback.c | 271 static unsigned long node_dirtyable_memory(struct pglist_data *pgdat) in node_dirtyable_memory() argument 277 struct zone *zone = pgdat->node_zones + z; in node_dirtyable_memory() 290 nr_pages -= min(nr_pages, pgdat->totalreserve_pages); in node_dirtyable_memory() 292 nr_pages += node_page_state(pgdat, NR_INACTIVE_FILE); in node_dirtyable_memory() 293 nr_pages += node_page_state(pgdat, NR_ACTIVE_FILE); in node_dirtyable_memory() 470 static unsigned long node_dirty_limit(struct pglist_data *pgdat) in node_dirty_limit() argument 472 unsigned long node_memory = node_dirtyable_memory(pgdat); in node_dirty_limit() 495 bool node_dirty_ok(struct pglist_data *pgdat) in node_dirty_ok() argument 497 unsigned long limit = node_dirty_limit(pgdat); in node_dirty_ok() 500 nr_pages += node_page_state(pgdat, NR_FILE_DIRTY); in node_dirty_ok() [all …]
|
D | memcontrol.c | 780 pg_data_t *pgdat = page_pgdat(page); in __mod_lruvec_page_state() local 788 __mod_node_page_state(pgdat, idx, val); in __mod_lruvec_page_state() 792 lruvec = mem_cgroup_lruvec(memcg, pgdat); in __mod_lruvec_page_state() 800 pg_data_t *pgdat = page_pgdat(virt_to_page(p)); in __mod_lruvec_kmem_state() local 814 __mod_node_page_state(pgdat, idx, val); in __mod_lruvec_kmem_state() 816 lruvec = mem_cgroup_lruvec(memcg, pgdat); in __mod_lruvec_kmem_state() 827 struct pglist_data *pgdat, in mod_objcg_mlstate() argument 835 lruvec = mem_cgroup_lruvec(memcg, pgdat); in mod_objcg_mlstate() 1055 mz = root->nodeinfo[reclaim->pgdat->node_id]; in mem_cgroup_iter() 1284 struct lruvec *page_to_lruvec(struct page *page, pg_data_t *pgdat) in page_to_lruvec() argument [all …]
|
D | page_ext.c | 177 void __meminit pgdat_page_ext_init(struct pglist_data *pgdat) in pgdat_page_ext_init() argument 179 pgdat->node_page_ext = NULL; in pgdat_page_ext_init() 517 void __meminit pgdat_page_ext_init(struct pglist_data *pgdat) in pgdat_page_ext_init() argument
|
D | memblock.c | 2086 void reset_node_managed_pages(pg_data_t *pgdat) in reset_node_managed_pages() argument 2090 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in reset_node_managed_pages() 2096 struct pglist_data *pgdat; in reset_all_zones_managed_pages() local 2101 for_each_online_pgdat(pgdat) in reset_all_zones_managed_pages() 2102 reset_node_managed_pages(pgdat); in reset_all_zones_managed_pages()
|
D | migrate.c | 2081 static bool migrate_balanced_pgdat(struct pglist_data *pgdat, in migrate_balanced_pgdat() argument 2086 for (z = pgdat->nr_zones - 1; z >= 0; z--) { in migrate_balanced_pgdat() 2087 struct zone *zone = pgdat->node_zones + z; in migrate_balanced_pgdat() 2135 static int numamigrate_isolate_page(pg_data_t *pgdat, struct page *page) in numamigrate_isolate_page() argument 2147 if (!migrate_balanced_pgdat(pgdat, nr_pages)) in numamigrate_isolate_page() 2174 pg_data_t *pgdat = NODE_DATA(node); in migrate_misplaced_page() local 2209 isolated = numamigrate_isolate_page(pgdat, page); in migrate_misplaced_page()
|
D | internal.h | 554 static inline int node_reclaim(struct pglist_data *pgdat, gfp_t mask, in node_reclaim() argument
|
D | huge_memory.c | 505 struct pglist_data *pgdat = NODE_DATA(page_to_nid(page)); in get_deferred_split_queue() local 510 return &pgdat->deferred_split_queue; in get_deferred_split_queue() 515 struct pglist_data *pgdat = NODE_DATA(page_to_nid(page)); in get_deferred_split_queue() local 517 return &pgdat->deferred_split_queue; in get_deferred_split_queue()
|
D | slab.h | 281 void mod_objcg_state(struct obj_cgroup *objcg, struct pglist_data *pgdat,
|