/mm/ |
D | shuffle.c | 105 void __meminit __shuffle_zone(struct zone *z) in __shuffle_zone() argument 108 unsigned long start_pfn = z->zone_start_pfn; in __shuffle_zone() 109 unsigned long end_pfn = zone_end_pfn(z); in __shuffle_zone() 113 spin_lock_irqsave(&z->lock, flags); in __shuffle_zone() 137 j = z->zone_start_pfn + in __shuffle_zone() 138 ALIGN_DOWN(get_random_long() % z->spanned_pages, in __shuffle_zone() 166 spin_unlock_irqrestore(&z->lock, flags); in __shuffle_zone() 168 spin_lock_irqsave(&z->lock, flags); in __shuffle_zone() 171 spin_unlock_irqrestore(&z->lock, flags); in __shuffle_zone() 180 struct zone *z; in __shuffle_free_memory() local [all …]
|
D | mmzone.c | 56 struct zoneref *__next_zones_zonelist(struct zoneref *z, in __next_zones_zonelist() argument 65 while (zonelist_zone_idx(z) > highest_zoneidx) in __next_zones_zonelist() 66 z++; in __next_zones_zonelist() 68 while (zonelist_zone_idx(z) > highest_zoneidx || in __next_zones_zonelist() 69 (z->zone && !zref_in_nodemask(z, nodes))) in __next_zones_zonelist() 70 z++; in __next_zones_zonelist() 72 return z; in __next_zones_zonelist()
|
D | shuffle.h | 32 extern void __shuffle_zone(struct zone *z); 33 static inline void shuffle_zone(struct zone *z) in shuffle_zone() argument 37 __shuffle_zone(z); in shuffle_zone() 51 static inline void shuffle_zone(struct zone *z) in shuffle_zone() argument
|
D | page_alloc.c | 2561 struct zoneref *z; in unreserve_highatomic_pageblock() local 2567 for_each_zone_zonelist_nodemask(zone, z, zonelist, ac->high_zoneidx, in unreserve_highatomic_pageblock() 2928 struct zone *z; in drain_all_pages() local 2936 for_each_populated_zone(z) { in drain_all_pages() 2937 pcp = per_cpu_ptr(z->pageset, cpu); in drain_all_pages() 3193 static inline void zone_statistics(struct zone *preferred_zone, struct zone *z) in zone_statistics() argument 3202 if (zone_to_nid(z) != numa_node_id()) in zone_statistics() 3205 if (zone_to_nid(z) == zone_to_nid(preferred_zone)) in zone_statistics() 3206 __inc_numa_state(z, NUMA_HIT); in zone_statistics() 3208 __inc_numa_state(z, NUMA_MISS); in zone_statistics() [all …]
|
D | vmscan.c | 2359 int z; in get_scan_count() local 2366 for (z = 0; z < MAX_NR_ZONES; z++) { in get_scan_count() 2367 struct zone *zone = &pgdat->node_zones[z]; in get_scan_count() 2700 int z; in should_continue_reclaim() local 2720 for (z = 0; z <= sc->reclaim_idx; z++) { in should_continue_reclaim() 2721 struct zone *zone = &pgdat->node_zones[z]; in should_continue_reclaim() 2952 struct zoneref *z; in shrink_zones() local 2970 for_each_zone_zonelist_nodemask(zone, z, zonelist, in shrink_zones() 3071 struct zoneref *z; in do_try_to_free_pages() local 3100 for_each_zone_zonelist_nodemask(zone, z, zonelist, sc->reclaim_idx, in do_try_to_free_pages() [all …]
|
D | page-writeback.c | 280 int z; in node_dirtyable_memory() local 282 for (z = 0; z < MAX_NR_ZONES; z++) { in node_dirtyable_memory() 283 struct zone *zone = pgdat->node_zones + z; in node_dirtyable_memory() 313 struct zone *z; in highmem_dirtyable_memory() local 319 z = &NODE_DATA(node)->node_zones[i]; in highmem_dirtyable_memory() 320 if (!populated_zone(z)) in highmem_dirtyable_memory() 323 nr_pages = zone_page_state(z, NR_FREE_PAGES); in highmem_dirtyable_memory() 325 nr_pages -= min(nr_pages, high_wmark_pages(z)); in highmem_dirtyable_memory() 326 nr_pages += zone_page_state(z, NR_ZONE_INACTIVE_FILE); in highmem_dirtyable_memory() 327 nr_pages += zone_page_state(z, NR_ZONE_ACTIVE_FILE); in highmem_dirtyable_memory()
|
D | vmstat.c | 485 long o, n, t, z; in mod_zone_state() local 488 z = 0; /* overflow to zone counters */ in mod_zone_state() 509 z = n + os; in mod_zone_state() 514 if (z) in mod_zone_state() 515 zone_page_state_add(z, zone, item); in mod_zone_state() 542 long o, n, t, z; in mod_node_state() local 545 z = 0; /* overflow to node counters */ in mod_node_state() 566 z = n + os; in mod_node_state() 571 if (z) in mod_node_state() 572 node_page_state_add(z, pgdat, item); in mod_node_state()
|
D | mm_init.c | 36 struct zoneref *z; in mminit_verify_zonelist() local 57 for_each_zone_zonelist(zone, z, zonelist, zoneid) in mminit_verify_zonelist()
|
D | memory_hotplug.c | 378 unsigned long z = zone_end_pfn(zone); /* zone_end_pfn namespace clash */ in shrink_zone_span() local 379 unsigned long zone_end_pfn = z; in shrink_zone_span() 877 struct zone *z; in reset_node_present_pages() local 879 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in reset_node_present_pages() 880 z->present_pages = 0; in reset_node_present_pages()
|
D | memblock.c | 1950 struct zone *z; in reset_node_managed_pages() local 1952 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in reset_node_managed_pages() 1953 atomic_long_set(&z->managed_pages, 0); in reset_node_managed_pages()
|
D | mempolicy.c | 1865 struct zoneref *z; in mempolicy_slab_node() local 1874 z = first_zones_zonelist(zonelist, highest_zoneidx, in mempolicy_slab_node() 1876 return z->zone ? zone_to_nid(z->zone) : node; in mempolicy_slab_node() 2389 struct zoneref *z; in mpol_misplaced() local 2425 z = first_zones_zonelist( in mpol_misplaced() 2429 polnid = zone_to_nid(z->zone); in mpol_misplaced()
|
D | oom_kill.c | 255 struct zoneref *z; in constrained_alloc() local 295 for_each_zone_zonelist_nodemask(zone, z, oc->zonelist, in constrained_alloc()
|
D | migrate.c | 1865 int z; in migrate_balanced_pgdat() local 1867 for (z = pgdat->nr_zones - 1; z >= 0; z--) { in migrate_balanced_pgdat() 1868 struct zone *zone = pgdat->node_zones + z; in migrate_balanced_pgdat()
|
D | compaction.c | 2043 struct zoneref *z; in compaction_zonelist_suitable() local 2049 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx, in compaction_zonelist_suitable() 2344 struct zoneref *z; in try_to_compact_pages() local 2358 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx, in try_to_compact_pages()
|
D | hugetlb.c | 900 struct zoneref *z; in dequeue_huge_page_nodemask() local 907 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nmask) { in dequeue_huge_page_nodemask() 1081 static bool pfn_range_valid_gigantic(struct zone *z, in pfn_range_valid_gigantic() argument 1092 if (page_zone(page) != z) in pfn_range_valid_gigantic() 1123 struct zoneref *z; in alloc_gigantic_page() local 1126 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nodemask) { in alloc_gigantic_page()
|
D | slab.c | 2747 #define cache_free_debugcheck(x,objp,z) (objp) argument 3106 struct zoneref *z; in fallback_alloc() local 3126 for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { in fallback_alloc()
|
D | slub.c | 1889 struct zoneref *z; in get_any_partial() local 1920 for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { in get_any_partial()
|