Home
last modified time | relevance | path

Searched refs:z (Results 1 – 17 of 17) sorted by relevance

/mm/
Dshuffle.c105 void __meminit __shuffle_zone(struct zone *z) in __shuffle_zone() argument
108 unsigned long start_pfn = z->zone_start_pfn; in __shuffle_zone()
109 unsigned long end_pfn = zone_end_pfn(z); in __shuffle_zone()
113 spin_lock_irqsave(&z->lock, flags); in __shuffle_zone()
137 j = z->zone_start_pfn + in __shuffle_zone()
138 ALIGN_DOWN(get_random_long() % z->spanned_pages, in __shuffle_zone()
166 spin_unlock_irqrestore(&z->lock, flags); in __shuffle_zone()
168 spin_lock_irqsave(&z->lock, flags); in __shuffle_zone()
171 spin_unlock_irqrestore(&z->lock, flags); in __shuffle_zone()
180 struct zone *z; in __shuffle_free_memory() local
[all …]
Dmmzone.c56 struct zoneref *__next_zones_zonelist(struct zoneref *z, in __next_zones_zonelist() argument
65 while (zonelist_zone_idx(z) > highest_zoneidx) in __next_zones_zonelist()
66 z++; in __next_zones_zonelist()
68 while (zonelist_zone_idx(z) > highest_zoneidx || in __next_zones_zonelist()
69 (z->zone && !zref_in_nodemask(z, nodes))) in __next_zones_zonelist()
70 z++; in __next_zones_zonelist()
72 return z; in __next_zones_zonelist()
Dshuffle.h32 extern void __shuffle_zone(struct zone *z);
33 static inline void shuffle_zone(struct zone *z) in shuffle_zone() argument
37 __shuffle_zone(z); in shuffle_zone()
51 static inline void shuffle_zone(struct zone *z) in shuffle_zone() argument
Dpage_alloc.c2561 struct zoneref *z; in unreserve_highatomic_pageblock() local
2567 for_each_zone_zonelist_nodemask(zone, z, zonelist, ac->high_zoneidx, in unreserve_highatomic_pageblock()
2928 struct zone *z; in drain_all_pages() local
2936 for_each_populated_zone(z) { in drain_all_pages()
2937 pcp = per_cpu_ptr(z->pageset, cpu); in drain_all_pages()
3193 static inline void zone_statistics(struct zone *preferred_zone, struct zone *z) in zone_statistics() argument
3202 if (zone_to_nid(z) != numa_node_id()) in zone_statistics()
3205 if (zone_to_nid(z) == zone_to_nid(preferred_zone)) in zone_statistics()
3206 __inc_numa_state(z, NUMA_HIT); in zone_statistics()
3208 __inc_numa_state(z, NUMA_MISS); in zone_statistics()
[all …]
Dvmscan.c2359 int z; in get_scan_count() local
2366 for (z = 0; z < MAX_NR_ZONES; z++) { in get_scan_count()
2367 struct zone *zone = &pgdat->node_zones[z]; in get_scan_count()
2700 int z; in should_continue_reclaim() local
2720 for (z = 0; z <= sc->reclaim_idx; z++) { in should_continue_reclaim()
2721 struct zone *zone = &pgdat->node_zones[z]; in should_continue_reclaim()
2952 struct zoneref *z; in shrink_zones() local
2970 for_each_zone_zonelist_nodemask(zone, z, zonelist, in shrink_zones()
3071 struct zoneref *z; in do_try_to_free_pages() local
3100 for_each_zone_zonelist_nodemask(zone, z, zonelist, sc->reclaim_idx, in do_try_to_free_pages()
[all …]
Dpage-writeback.c280 int z; in node_dirtyable_memory() local
282 for (z = 0; z < MAX_NR_ZONES; z++) { in node_dirtyable_memory()
283 struct zone *zone = pgdat->node_zones + z; in node_dirtyable_memory()
313 struct zone *z; in highmem_dirtyable_memory() local
319 z = &NODE_DATA(node)->node_zones[i]; in highmem_dirtyable_memory()
320 if (!populated_zone(z)) in highmem_dirtyable_memory()
323 nr_pages = zone_page_state(z, NR_FREE_PAGES); in highmem_dirtyable_memory()
325 nr_pages -= min(nr_pages, high_wmark_pages(z)); in highmem_dirtyable_memory()
326 nr_pages += zone_page_state(z, NR_ZONE_INACTIVE_FILE); in highmem_dirtyable_memory()
327 nr_pages += zone_page_state(z, NR_ZONE_ACTIVE_FILE); in highmem_dirtyable_memory()
Dvmstat.c485 long o, n, t, z; in mod_zone_state() local
488 z = 0; /* overflow to zone counters */ in mod_zone_state()
509 z = n + os; in mod_zone_state()
514 if (z) in mod_zone_state()
515 zone_page_state_add(z, zone, item); in mod_zone_state()
542 long o, n, t, z; in mod_node_state() local
545 z = 0; /* overflow to node counters */ in mod_node_state()
566 z = n + os; in mod_node_state()
571 if (z) in mod_node_state()
572 node_page_state_add(z, pgdat, item); in mod_node_state()
Dmm_init.c36 struct zoneref *z; in mminit_verify_zonelist() local
57 for_each_zone_zonelist(zone, z, zonelist, zoneid) in mminit_verify_zonelist()
Dmemory_hotplug.c378 unsigned long z = zone_end_pfn(zone); /* zone_end_pfn namespace clash */ in shrink_zone_span() local
379 unsigned long zone_end_pfn = z; in shrink_zone_span()
877 struct zone *z; in reset_node_present_pages() local
879 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in reset_node_present_pages()
880 z->present_pages = 0; in reset_node_present_pages()
Dmemblock.c1950 struct zone *z; in reset_node_managed_pages() local
1952 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in reset_node_managed_pages()
1953 atomic_long_set(&z->managed_pages, 0); in reset_node_managed_pages()
Dmempolicy.c1865 struct zoneref *z; in mempolicy_slab_node() local
1874 z = first_zones_zonelist(zonelist, highest_zoneidx, in mempolicy_slab_node()
1876 return z->zone ? zone_to_nid(z->zone) : node; in mempolicy_slab_node()
2389 struct zoneref *z; in mpol_misplaced() local
2425 z = first_zones_zonelist( in mpol_misplaced()
2429 polnid = zone_to_nid(z->zone); in mpol_misplaced()
Doom_kill.c255 struct zoneref *z; in constrained_alloc() local
295 for_each_zone_zonelist_nodemask(zone, z, oc->zonelist, in constrained_alloc()
Dmigrate.c1865 int z; in migrate_balanced_pgdat() local
1867 for (z = pgdat->nr_zones - 1; z >= 0; z--) { in migrate_balanced_pgdat()
1868 struct zone *zone = pgdat->node_zones + z; in migrate_balanced_pgdat()
Dcompaction.c2043 struct zoneref *z; in compaction_zonelist_suitable() local
2049 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx, in compaction_zonelist_suitable()
2344 struct zoneref *z; in try_to_compact_pages() local
2358 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx, in try_to_compact_pages()
Dhugetlb.c900 struct zoneref *z; in dequeue_huge_page_nodemask() local
907 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nmask) { in dequeue_huge_page_nodemask()
1081 static bool pfn_range_valid_gigantic(struct zone *z, in pfn_range_valid_gigantic() argument
1092 if (page_zone(page) != z) in pfn_range_valid_gigantic()
1123 struct zoneref *z; in alloc_gigantic_page() local
1126 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nodemask) { in alloc_gigantic_page()
Dslab.c2747 #define cache_free_debugcheck(x,objp,z) (objp) argument
3106 struct zoneref *z; in fallback_alloc() local
3126 for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { in fallback_alloc()
Dslub.c1889 struct zoneref *z; in get_any_partial() local
1920 for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { in get_any_partial()