Home
last modified time | relevance | path

Searched refs:node_zones (Results 1 – 21 of 21) sorted by relevance

/kernel/linux/linux-5.10/mm/
Dmmzone.c34 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone()
39 zone = pgdat->node_zones; in next_zone()
Dshuffle.c158 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in __shuffle_free_memory()
Dmemory_hotplug.c447 for (zone = pgdat->node_zones; in update_pgdat_span()
448 zone < pgdat->node_zones + MAX_NR_ZONES; zone++) { in update_pgdat_span()
736 struct zone *zone = &pgdat->node_zones[zid]; in default_kernel_zone_for_pfn()
742 return &pgdat->node_zones[ZONE_NORMAL]; in default_kernel_zone_for_pfn()
750 struct zone *movable_zone = &NODE_DATA(nid)->node_zones[ZONE_MOVABLE]; in default_zone_for_pfn()
776 return &NODE_DATA(nid)->node_zones[ZONE_MOVABLE]; in zone_for_pfn_range()
880 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in reset_node_present_pages()
1401 present_pages += pgdat->node_zones[zt].present_pages; in node_states_check_changes_offline()
1414 present_pages += pgdat->node_zones[ZONE_HIGHMEM].present_pages; in node_states_check_changes_offline()
1429 present_pages += pgdat->node_zones[ZONE_MOVABLE].present_pages; in node_states_check_changes_offline()
Dpage_owner.c633 struct zone *node_zones = pgdat->node_zones; in init_zones_in_node() local
635 for (zone = node_zones; zone - node_zones < MAX_NR_ZONES; ++zone) { in init_zones_in_node()
Dvmstat.c300 zone = &pgdat->node_zones[i]; in set_pgdat_percpu_threshold()
976 struct zone *zones = NODE_DATA(node)->node_zones; in sum_zone_node_page_state()
993 struct zone *zones = NODE_DATA(node)->node_zones; in sum_zone_numa_state()
1422 struct zone *node_zones = pgdat->node_zones; in walk_zones_in_node() local
1425 for (zone = node_zones; zone - node_zones < MAX_NR_ZONES; ++zone) { in walk_zones_in_node()
1634 struct zone *compare = &pgdat->node_zones[zid]; in is_zone_first_populated()
Dmemcg_reclaim.c70 struct zone *zone = &pgdat->node_zones[z]; in get_scan_count_hyperhold()
437 struct zone *zone = &pgdat->node_zones[z]; in shrink_node_hyperhold()
Dvmscan.c234 struct zone *zone = &lruvec_pgdat(lruvec)->node_zones[zid]; in lruvec_lru_size()
246 struct zone *zone = &lruvec_pgdat(lruvec)->node_zones[zid]; in lruvec_lru_size()
2571 struct zone *zone = &pgdat->node_zones[z]; in should_continue_reclaim()
2739 struct zone *zone = &pgdat->node_zones[z]; in shrink_node()
3113 zone = &pgdat->node_zones[i]; in allow_direct_reclaim()
3406 zone = pgdat->node_zones + i; in pgdat_watermark_boosted()
3432 zone = pgdat->node_zones + i; in pgdat_balanced()
3517 zone = pgdat->node_zones + z; in kswapd_shrink_node()
3589 zone = pgdat->node_zones + i; in balance_pgdat()
3620 zone = pgdat->node_zones + i; in balance_pgdat()
[all …]
Dmm_init.c48 zone = &pgdat->node_zones[zoneid]; in mminit_verify_zonelist()
Dmemremap.c278 zone = &NODE_DATA(nid)->node_zones[ZONE_DEVICE]; in pagemap_range()
292 memmap_init_zone_device(&NODE_DATA(nid)->node_zones[ZONE_DEVICE], in pagemap_range()
Dcompaction.c391 struct zone *zone = &pgdat->node_zones[zoneid]; in reset_isolation_suitable()
1918 zone = &pgdat->node_zones[zoneid]; in fragmentation_score_node()
2556 zone = &pgdat->node_zones[zoneid]; in proactive_compact_node()
2586 zone = &pgdat->node_zones[zoneid]; in compact_node()
2675 zone = &pgdat->node_zones[zoneid]; in kcompactd_node_suitable()
2711 zone = &pgdat->node_zones[zoneid]; in kcompactd_do_work()
Dpage_alloc.c1478 struct zone *zone = &pgdat->node_zones[zid]; in init_reserved_page()
1952 zone = pgdat->node_zones + zid; in deferred_init_memmap()
5495 managed_pages += zone_managed_pages(&pgdat->node_zones[zone_type]); in si_meminfo_node()
5501 struct zone *zone = &pgdat->node_zones[zone_type]; in si_meminfo_node()
5790 zone = pgdat->node_zones + zone_type; in build_zonerefs_node()
6416 struct zone *zone = node->node_zones + j; in memmap_init()
6875 struct zone *zone = pgdat->node_zones + i; in calculate_node_totalpages()
7078 zone_init_internals(&pgdat->node_zones[z], z, nid, 0); in free_area_init_core_hotplug()
7100 struct zone *zone = pgdat->node_zones + j; in free_area_init_core()
7567 struct zone *zone = &pgdat->node_zones[zone_type]; in check_for_memory()
[all …]
Dmemblock.c1966 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in reset_node_managed_pages()
Dpage-writeback.c278 struct zone *zone = pgdat->node_zones + z; in node_dirtyable_memory()
314 z = &NODE_DATA(node)->node_zones[i]; in highmem_dirtyable_memory()
Dmigrate.c2010 struct zone *zone = pgdat->node_zones + z; in migrate_balanced_pgdat()
/kernel/linux/linux-5.10/lib/
Dshow_mem.c23 struct zone *zone = &pgdat->node_zones[zoneid]; in show_mem()
/kernel/linux/linux-5.10/include/linux/
Dmmzone.h743 struct zone node_zones[MAX_NR_ZONES]; member
929 #define zone_idx(zone) ((zone) - (zone)->zone_pgdat->node_zones)
1068 for (zone = (first_online_pgdat())->node_zones; \
1073 for (zone = (first_online_pgdat())->node_zones; \
Dmm_inline.h35 __mod_zone_page_state(&pgdat->node_zones[zid], in __update_lru_size()
Dmm.h1489 return &NODE_DATA(page_to_nid(page))->node_zones[page_zonenum(page)]; in page_zone()
/kernel/linux/linux-5.10/kernel/
Dcrash_core.c484 VMCOREINFO_OFFSET(pglist_data, node_zones); in crash_save_vmcoreinfo_init()
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdkfd/
Dkfd_crat.c900 mem_in_bytes += zone_managed_pages(&pgdat->node_zones[zone_type]); in kfd_fill_mem_info_for_cpu()
/kernel/linux/linux-5.10/Documentation/admin-guide/kdump/
Dvmcoreinfo.rst144 (pglist_data, node_zones|nr_zones|node_mem_map|node_start_pfn|node_spanned_pages|node_id)