Lines Matching refs:zone
39 static void zero_zone_numa_counters(struct zone *zone) in zero_zone_numa_counters() argument
44 atomic_long_set(&zone->vm_numa_stat[item], 0); in zero_zone_numa_counters()
46 per_cpu_ptr(zone->pageset, cpu)->vm_numa_stat_diff[item] in zero_zone_numa_counters()
54 struct zone *zone; in zero_zones_numa_counters() local
56 for_each_populated_zone(zone) in zero_zones_numa_counters()
57 zero_zone_numa_counters(zone); in zero_zones_numa_counters()
171 int calculate_pressure_threshold(struct zone *zone) in calculate_pressure_threshold() argument
184 watermark_distance = low_wmark_pages(zone) - min_wmark_pages(zone); in calculate_pressure_threshold()
195 int calculate_normal_threshold(struct zone *zone) in calculate_normal_threshold() argument
230 mem = zone->managed_pages >> (27 - PAGE_SHIFT); in calculate_normal_threshold()
248 struct zone *zone; in refresh_zone_stat_thresholds() local
259 for_each_populated_zone(zone) { in refresh_zone_stat_thresholds()
260 struct pglist_data *pgdat = zone->zone_pgdat; in refresh_zone_stat_thresholds()
263 threshold = calculate_normal_threshold(zone); in refresh_zone_stat_thresholds()
268 per_cpu_ptr(zone->pageset, cpu)->stat_threshold in refresh_zone_stat_thresholds()
282 tolerate_drift = low_wmark_pages(zone) - min_wmark_pages(zone); in refresh_zone_stat_thresholds()
285 zone->percpu_drift_mark = high_wmark_pages(zone) + in refresh_zone_stat_thresholds()
291 int (*calculate_pressure)(struct zone *)) in set_pgdat_percpu_threshold() argument
293 struct zone *zone; in set_pgdat_percpu_threshold() local
299 zone = &pgdat->node_zones[i]; in set_pgdat_percpu_threshold()
300 if (!zone->percpu_drift_mark) in set_pgdat_percpu_threshold()
303 threshold = (*calculate_pressure)(zone); in set_pgdat_percpu_threshold()
305 per_cpu_ptr(zone->pageset, cpu)->stat_threshold in set_pgdat_percpu_threshold()
315 void __mod_zone_page_state(struct zone *zone, enum zone_stat_item item, in __mod_zone_page_state() argument
318 struct per_cpu_pageset __percpu *pcp = zone->pageset; in __mod_zone_page_state()
328 zone_page_state_add(x, zone, item); in __mod_zone_page_state()
378 void __inc_zone_state(struct zone *zone, enum zone_stat_item item) in __inc_zone_state() argument
380 struct per_cpu_pageset __percpu *pcp = zone->pageset; in __inc_zone_state()
389 zone_page_state_add(v + overstep, zone, item); in __inc_zone_state()
422 void __dec_zone_state(struct zone *zone, enum zone_stat_item item) in __dec_zone_state() argument
424 struct per_cpu_pageset __percpu *pcp = zone->pageset; in __dec_zone_state()
433 zone_page_state_add(v - overstep, zone, item); in __dec_zone_state()
479 static inline void mod_zone_state(struct zone *zone, in mod_zone_state() argument
482 struct per_cpu_pageset __percpu *pcp = zone->pageset; in mod_zone_state()
514 zone_page_state_add(z, zone, item); in mod_zone_state()
517 void mod_zone_page_state(struct zone *zone, enum zone_stat_item item, in mod_zone_page_state() argument
520 mod_zone_state(zone, item, delta, 0); in mod_zone_page_state()
601 void mod_zone_page_state(struct zone *zone, enum zone_stat_item item, in mod_zone_page_state() argument
607 __mod_zone_page_state(zone, item, delta); in mod_zone_page_state()
615 struct zone *zone; in inc_zone_page_state() local
617 zone = page_zone(page); in inc_zone_page_state()
619 __inc_zone_state(zone, item); in inc_zone_page_state()
747 struct zone *zone; in refresh_cpu_vm_stats() local
756 for_each_populated_zone(zone) { in refresh_cpu_vm_stats()
757 struct per_cpu_pageset __percpu *p = zone->pageset; in refresh_cpu_vm_stats()
765 atomic_long_add(v, &zone->vm_stat[i]); in refresh_cpu_vm_stats()
780 atomic_long_add(v, &zone->vm_numa_stat[i]); in refresh_cpu_vm_stats()
802 if (zone_to_nid(zone) == numa_node_id()) { in refresh_cpu_vm_stats()
811 drain_zone_pages(zone, this_cpu_ptr(&p->pcp)); in refresh_cpu_vm_stats()
849 struct zone *zone; in cpu_vm_stats_fold() local
857 for_each_populated_zone(zone) { in cpu_vm_stats_fold()
860 p = per_cpu_ptr(zone->pageset, cpu); in cpu_vm_stats_fold()
868 atomic_long_add(v, &zone->vm_stat[i]); in cpu_vm_stats_fold()
879 atomic_long_add(v, &zone->vm_numa_stat[i]); in cpu_vm_stats_fold()
912 void drain_zonestat(struct zone *zone, struct per_cpu_pageset *pset) in drain_zonestat() argument
920 atomic_long_add(v, &zone->vm_stat[i]); in drain_zonestat()
930 atomic_long_add(v, &zone->vm_numa_stat[i]); in drain_zonestat()
938 void __inc_numa_state(struct zone *zone, in __inc_numa_state() argument
941 struct per_cpu_pageset __percpu *pcp = zone->pageset; in __inc_numa_state()
948 zone_numa_state_add(v, zone, item); in __inc_numa_state()
961 struct zone *zones = NODE_DATA(node)->node_zones; in sum_zone_node_page_state()
978 struct zone *zones = NODE_DATA(node)->node_zones; in sum_zone_numa_state()
1019 static void fill_contig_page_info(struct zone *zone, in fill_contig_page_info() argument
1033 blocks = zone->free_area[order].nr_free; in fill_contig_page_info()
1077 int fragmentation_index(struct zone *zone, unsigned int order) in fragmentation_index() argument
1081 fill_contig_page_info(zone, order, &info); in fragmentation_index()
1326 void (*print)(struct seq_file *m, pg_data_t *, struct zone *)) in walk_zones_in_node() argument
1328 struct zone *zone; in walk_zones_in_node() local
1329 struct zone *node_zones = pgdat->node_zones; in walk_zones_in_node()
1332 for (zone = node_zones; zone - node_zones < MAX_NR_ZONES; ++zone) { in walk_zones_in_node()
1333 if (assert_populated && !populated_zone(zone)) in walk_zones_in_node()
1337 spin_lock_irqsave(&zone->lock, flags); in walk_zones_in_node()
1338 print(m, pgdat, zone); in walk_zones_in_node()
1340 spin_unlock_irqrestore(&zone->lock, flags); in walk_zones_in_node()
1347 struct zone *zone) in frag_show_print() argument
1351 seq_printf(m, "Node %d, zone %8s ", pgdat->node_id, zone->name); in frag_show_print()
1353 seq_printf(m, "%6lu ", zone->free_area[order].nr_free); in frag_show_print()
1368 pg_data_t *pgdat, struct zone *zone) in pagetypeinfo_showfree_print() argument
1375 zone->name, in pagetypeinfo_showfree_print()
1382 area = &(zone->free_area[order]); in pagetypeinfo_showfree_print()
1410 pg_data_t *pgdat, struct zone *zone) in pagetypeinfo_showblockcount_print() argument
1414 unsigned long start_pfn = zone->zone_start_pfn; in pagetypeinfo_showblockcount_print()
1415 unsigned long end_pfn = zone_end_pfn(zone); in pagetypeinfo_showblockcount_print()
1426 if (!memmap_valid_within(pfn, page, zone)) in pagetypeinfo_showblockcount_print()
1429 if (page_zone(page) != zone) in pagetypeinfo_showblockcount_print()
1439 seq_printf(m, "Node %d, zone %8s ", pgdat->node_id, zone->name); in pagetypeinfo_showblockcount_print()
1523 static bool is_zone_first_populated(pg_data_t *pgdat, struct zone *zone) in is_zone_first_populated() argument
1528 struct zone *compare = &pgdat->node_zones[zid]; in is_zone_first_populated()
1531 return zone == compare; in is_zone_first_populated()
1538 struct zone *zone) in zoneinfo_show_print() argument
1541 seq_printf(m, "Node %d, zone %8s", pgdat->node_id, zone->name); in zoneinfo_show_print()
1542 if (is_zone_first_populated(pgdat, zone)) { in zoneinfo_show_print()
1563 zone_page_state(zone, NR_FREE_PAGES), in zoneinfo_show_print()
1564 min_wmark_pages(zone), in zoneinfo_show_print()
1565 low_wmark_pages(zone), in zoneinfo_show_print()
1566 high_wmark_pages(zone), in zoneinfo_show_print()
1567 zone->spanned_pages, in zoneinfo_show_print()
1568 zone->present_pages, in zoneinfo_show_print()
1569 zone->managed_pages); in zoneinfo_show_print()
1573 zone->lowmem_reserve[0]); in zoneinfo_show_print()
1574 for (i = 1; i < ARRAY_SIZE(zone->lowmem_reserve); i++) in zoneinfo_show_print()
1575 seq_printf(m, ", %ld", zone->lowmem_reserve[i]); in zoneinfo_show_print()
1579 if (!populated_zone(zone)) { in zoneinfo_show_print()
1586 zone_page_state(zone, i)); in zoneinfo_show_print()
1592 zone_numa_state_snapshot(zone, i)); in zoneinfo_show_print()
1599 pageset = per_cpu_ptr(zone->pageset, i); in zoneinfo_show_print()
1618 zone->zone_start_pfn); in zoneinfo_show_print()
1817 struct zone *zone; in need_update() local
1819 for_each_populated_zone(zone) { in need_update()
1820 struct per_cpu_pageset *p = per_cpu_ptr(zone->pageset, cpu); in need_update()
2007 pg_data_t *pgdat, struct zone *zone) in unusable_show_print() argument
2015 zone->name); in unusable_show_print()
2017 fill_contig_page_info(zone, order, &info); in unusable_show_print()
2067 pg_data_t *pgdat, struct zone *zone) in extfrag_show_print() argument
2077 zone->name); in extfrag_show_print()
2079 fill_contig_page_info(zone, order, &info); in extfrag_show_print()