• Home
  • Raw
  • Download

Lines Matching refs:z

2561 	struct zoneref *z;  in unreserve_highatomic_pageblock()  local
2567 for_each_zone_zonelist_nodemask(zone, z, zonelist, ac->high_zoneidx, in unreserve_highatomic_pageblock()
2928 struct zone *z; in drain_all_pages() local
2936 for_each_populated_zone(z) { in drain_all_pages()
2937 pcp = per_cpu_ptr(z->pageset, cpu); in drain_all_pages()
3193 static inline void zone_statistics(struct zone *preferred_zone, struct zone *z) in zone_statistics() argument
3202 if (zone_to_nid(z) != numa_node_id()) in zone_statistics()
3205 if (zone_to_nid(z) == zone_to_nid(preferred_zone)) in zone_statistics()
3206 __inc_numa_state(z, NUMA_HIT); in zone_statistics()
3208 __inc_numa_state(z, NUMA_MISS); in zone_statistics()
3211 __inc_numa_state(z, local_stat); in zone_statistics()
3402 bool __zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark, in __zone_watermark_ok() argument
3422 free_pages -= z->nr_reserved_highatomic; in __zone_watermark_ok()
3440 free_pages -= zone_page_state(z, NR_FREE_CMA_PAGES); in __zone_watermark_ok()
3448 if (free_pages <= min + z->lowmem_reserve[classzone_idx]) in __zone_watermark_ok()
3457 struct free_area *area = &z->free_area[o]; in __zone_watermark_ok()
3481 bool zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark, in zone_watermark_ok() argument
3484 return __zone_watermark_ok(z, order, mark, classzone_idx, alloc_flags, in zone_watermark_ok()
3485 zone_page_state(z, NR_FREE_PAGES)); in zone_watermark_ok()
3488 static inline bool zone_watermark_fast(struct zone *z, unsigned int order, in zone_watermark_fast() argument
3491 long free_pages = zone_page_state(z, NR_FREE_PAGES); in zone_watermark_fast()
3497 cma_pages = zone_page_state(z, NR_FREE_CMA_PAGES); in zone_watermark_fast()
3507 if (!order && (free_pages - cma_pages) > mark + z->lowmem_reserve[classzone_idx]) in zone_watermark_fast()
3510 return __zone_watermark_ok(z, order, mark, classzone_idx, alloc_flags, in zone_watermark_fast()
3514 bool zone_watermark_ok_safe(struct zone *z, unsigned int order, in zone_watermark_ok_safe() argument
3517 long free_pages = zone_page_state(z, NR_FREE_PAGES); in zone_watermark_ok_safe()
3519 if (z->percpu_drift_mark && free_pages < z->percpu_drift_mark) in zone_watermark_ok_safe()
3520 free_pages = zone_page_state_snapshot(z, NR_FREE_PAGES); in zone_watermark_ok_safe()
3522 return __zone_watermark_ok(z, order, mark, classzone_idx, 0, in zone_watermark_ok_safe()
3584 struct zoneref *z; in get_page_from_freelist() local
3595 z = ac->preferred_zoneref; in get_page_from_freelist()
3596 for_next_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx, in get_page_from_freelist()
4036 struct zoneref *z; in should_compact_retry() local
4047 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx, in should_compact_retry()
4171 struct zoneref *z; in wake_all_kswapds() local
4176 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, high_zoneidx, in wake_all_kswapds()
4283 struct zoneref *z; in should_reclaim_retry() local
4311 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx, in should_reclaim_retry()
4327 trace_reclaim_retry_zone(z, order, reclaimable, in should_reclaim_retry()
5049 struct zoneref *z; in nr_free_zone_pages() local
5057 for_each_zone_zonelist(zone, z, zonelist, offset) { in nr_free_zone_pages()
5680 struct zoneref *z; in local_memory_node() local
5682 z = first_zones_zonelist(node_zonelist(node, GFP_KERNEL), in local_memory_node()
5685 return zone_to_nid(z->zone); in local_memory_node()
6746 enum zone_type z; in free_area_init_core_hotplug() local
6750 for (z = 0; z < MAX_NR_ZONES; z++) in free_area_init_core_hotplug()
6751 zone_init_internals(&pgdat->node_zones[z], z, nid, 0); in free_area_init_core_hotplug()