Home
last modified time | relevance | path

Searched refs:pgdat (Results 1 – 25 of 50) sorted by relevance

12

/kernel/linux/linux-5.10/mm/
Dmemcg_reclaim.c31 static void get_scan_count_hyperhold(struct pglist_data *pgdat, in get_scan_count_hyperhold() argument
36 struct lruvec *lruvec = node_lruvec(pgdat); in get_scan_count_hyperhold()
65 pgdatfree = sum_zone_node_page_state(pgdat->node_id, NR_FREE_PAGES); in get_scan_count_hyperhold()
66 pgdatfile = node_page_state(pgdat, NR_ACTIVE_FILE) + in get_scan_count_hyperhold()
67 node_page_state(pgdat, NR_INACTIVE_FILE); in get_scan_count_hyperhold()
70 struct zone *zone = &pgdat->node_zones[z]; in get_scan_count_hyperhold()
186 void shrink_anon_memcg(struct pglist_data *pgdat, in shrink_anon_memcg() argument
190 struct lruvec *lruvec = mem_cgroup_lruvec(memcg, pgdat); in shrink_anon_memcg()
233 static void shrink_anon(struct pglist_data *pgdat, in shrink_anon() argument
242 node_lruvec(pgdat), LRU_ACTIVE_ANON, MAX_NR_ZONES); in shrink_anon()
[all …]
Dvmscan.c1013 struct pglist_data *pgdat, in shrink_page_list() argument
1128 test_bit(PGDAT_WRITEBACK, &pgdat->flags)) { in shrink_page_list()
1269 !test_bit(PGDAT_DIRTY, &pgdat->flags))) { in shrink_page_list()
1714 pg_data_t *pgdat = page_pgdat(page); in isolate_lru_page() local
1717 spin_lock_irq(&pgdat->lru_lock); in isolate_lru_page()
1718 lruvec = mem_cgroup_page_lruvec(page, pgdat); in isolate_lru_page()
1726 spin_unlock_irq(&pgdat->lru_lock); in isolate_lru_page()
1738 static int too_many_isolated(struct pglist_data *pgdat, int file, in too_many_isolated() argument
1750 inactive = node_page_state(pgdat, NR_INACTIVE_FILE); in too_many_isolated()
1751 isolated = node_page_state(pgdat, NR_ISOLATED_FILE); in too_many_isolated()
[all …]
Dvmstat.c248 struct pglist_data *pgdat; in refresh_zone_stat_thresholds() local
254 for_each_online_pgdat(pgdat) { in refresh_zone_stat_thresholds()
256 per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold = 0; in refresh_zone_stat_thresholds()
261 struct pglist_data *pgdat = zone->zone_pgdat; in refresh_zone_stat_thresholds() local
273 pgdat_threshold = per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold; in refresh_zone_stat_thresholds()
274 per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold in refresh_zone_stat_thresholds()
291 void set_pgdat_percpu_threshold(pg_data_t *pgdat, in set_pgdat_percpu_threshold() argument
299 for (i = 0; i < pgdat->nr_zones; i++) { in set_pgdat_percpu_threshold()
300 zone = &pgdat->node_zones[i]; in set_pgdat_percpu_threshold()
336 void __mod_node_page_state(struct pglist_data *pgdat, enum node_stat_item item, in __mod_node_page_state() argument
[all …]
Dzswapd.c436 void wakeup_zswapd(pg_data_t *pgdat) in wakeup_zswapd() argument
440 if (IS_ERR(pgdat->zswapd)) in wakeup_zswapd()
443 if (!wq_has_sleeper(&pgdat->zswapd_wait)) in wakeup_zswapd()
463 atomic_set(&pgdat->zswapd_wait_flag, 1); in wakeup_zswapd()
464 wake_up_interruptible(&pgdat->zswapd_wait); in wakeup_zswapd()
469 pg_data_t *pgdat = NULL; in wake_all_zswapd() local
473 pgdat = NODE_DATA(nid); in wake_all_zswapd()
474 wakeup_zswapd(pgdat); in wake_all_zswapd()
487 struct pglist_data *pgdat = lruvec_pgdat(lruvec); in zswapd_shrink_active_list() local
488 unsigned long *node_anon_cost = &pgdat->__lruvec.anon_cost; in zswapd_shrink_active_list()
[all …]
Dmemory_hotplug.c232 void __init register_page_bootmem_info_node(struct pglist_data *pgdat) in register_page_bootmem_info_node() argument
235 int node = pgdat->node_id; in register_page_bootmem_info_node()
239 page = virt_to_page(pgdat); in register_page_bootmem_info_node()
244 pfn = pgdat->node_start_pfn; in register_page_bootmem_info_node()
245 end_pfn = pgdat_end_pfn(pgdat); in register_page_bootmem_info_node()
442 static void update_pgdat_span(struct pglist_data *pgdat) in update_pgdat_span() argument
447 for (zone = pgdat->node_zones; in update_pgdat_span()
448 zone < pgdat->node_zones + MAX_NR_ZONES; zone++) { in update_pgdat_span()
467 pgdat->node_start_pfn = node_start_pfn; in update_pgdat_span()
468 pgdat->node_spanned_pages = node_end_pfn - node_start_pfn; in update_pgdat_span()
[all …]
Dcompaction.c386 void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() argument
391 struct zone *zone = &pgdat->node_zones[zoneid]; in reset_isolation_suitable()
767 static bool too_many_isolated(pg_data_t *pgdat) in too_many_isolated() argument
771 inactive = node_page_state(pgdat, NR_INACTIVE_FILE) + in too_many_isolated()
772 node_page_state(pgdat, NR_INACTIVE_ANON); in too_many_isolated()
773 active = node_page_state(pgdat, NR_ACTIVE_FILE) + in too_many_isolated()
774 node_page_state(pgdat, NR_ACTIVE_ANON); in too_many_isolated()
775 isolated = node_page_state(pgdat, NR_ISOLATED_FILE) + in too_many_isolated()
776 node_page_state(pgdat, NR_ISOLATED_ANON); in too_many_isolated()
803 pg_data_t *pgdat = cc->zone->zone_pgdat; in isolate_migratepages_block() local
[all …]
Dmmzone.c18 struct pglist_data *next_online_pgdat(struct pglist_data *pgdat) in next_online_pgdat() argument
20 int nid = next_online_node(pgdat->node_id); in next_online_pgdat()
32 pg_data_t *pgdat = zone->zone_pgdat; in next_zone() local
34 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone()
37 pgdat = next_online_pgdat(pgdat); in next_zone()
38 if (pgdat) in next_zone()
39 zone = pgdat->node_zones; in next_zone()
Dworkingset.c185 static void *pack_shadow(int memcgid, pg_data_t *pgdat, unsigned long eviction, in pack_shadow() argument
191 eviction = (eviction << NODES_SHIFT) | pgdat->node_id; in pack_shadow()
197 static void unpack_shadow(void *shadow, int *memcgidp, pg_data_t **pgdat, in unpack_shadow() argument
212 *pgdat = NODE_DATA(nid); in unpack_shadow()
255 struct pglist_data *pgdat = page_pgdat(page); in workingset_eviction() local
265 lruvec = mem_cgroup_lruvec(target_memcg, pgdat); in workingset_eviction()
268 lruvec = node_lruvec(pgdat); in workingset_eviction()
279 return pack_shadow(memcgid, pgdat, eviction, PageWorkingset(page)); in workingset_eviction()
298 struct pglist_data *pgdat; in workingset_refault() local
306 unpack_shadow(shadow, &memcgid, &pgdat, &eviction, &workingset); in workingset_refault()
[all …]
Dpage_alloc.c1468 pg_data_t *pgdat; in init_reserved_page() local
1475 pgdat = NODE_DATA(nid); in init_reserved_page()
1478 struct zone *zone = &pgdat->node_zones[zid]; in init_reserved_page()
1917 pg_data_t *pgdat = data; in deferred_init_memmap() local
1918 const struct cpumask *cpumask = cpumask_of_node(pgdat->node_id); in deferred_init_memmap()
1930 pgdat_resize_lock(pgdat, &flags); in deferred_init_memmap()
1931 first_init_pfn = pgdat->first_deferred_pfn; in deferred_init_memmap()
1933 pgdat_resize_unlock(pgdat, &flags); in deferred_init_memmap()
1939 BUG_ON(pgdat->first_deferred_pfn < pgdat->node_start_pfn); in deferred_init_memmap()
1940 BUG_ON(pgdat->first_deferred_pfn > pgdat_end_pfn(pgdat)); in deferred_init_memmap()
[all …]
Dshuffle.h11 extern void __shuffle_free_memory(pg_data_t *pgdat);
13 static inline void shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory() argument
17 __shuffle_free_memory(pgdat); in shuffle_free_memory()
40 static inline void shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory() argument
Dswap.c82 pg_data_t *pgdat = page_pgdat(page); in __page_cache_release() local
86 spin_lock_irqsave(&pgdat->lru_lock, flags); in __page_cache_release()
87 lruvec = mem_cgroup_page_lruvec(page, pgdat); in __page_cache_release()
91 spin_unlock_irqrestore(&pgdat->lru_lock, flags); in __page_cache_release()
211 struct pglist_data *pgdat = NULL; in pagevec_lru_move_fn() local
219 if (pagepgdat != pgdat) { in pagevec_lru_move_fn()
220 if (pgdat) in pagevec_lru_move_fn()
221 spin_unlock_irqrestore(&pgdat->lru_lock, flags); in pagevec_lru_move_fn()
222 pgdat = pagepgdat; in pagevec_lru_move_fn()
223 spin_lock_irqsave(&pgdat->lru_lock, flags); in pagevec_lru_move_fn()
[all …]
Dpage_owner.c265 pg_data_t *pgdat, struct zone *zone) in pagetypeinfo_showmixedcount_print() argument
341 seq_printf(m, "Node %d, zone %8s ", pgdat->node_id, zone->name); in pagetypeinfo_showmixedcount_print()
558 static void init_pages_in_zone(pg_data_t *pgdat, struct zone *zone) in init_pages_in_zone() argument
627 pgdat->node_id, zone->name, count); in init_pages_in_zone()
630 static void init_zones_in_node(pg_data_t *pgdat) in init_zones_in_node() argument
633 struct zone *node_zones = pgdat->node_zones; in init_zones_in_node()
639 init_pages_in_zone(pgdat, zone); in init_zones_in_node()
645 pg_data_t *pgdat; in init_early_allocated_pages() local
647 for_each_online_pgdat(pgdat) in init_early_allocated_pages()
648 init_zones_in_node(pgdat); in init_early_allocated_pages()
Dpage_idle.c35 pg_data_t *pgdat; in page_idle_get_page() local
41 pgdat = page_pgdat(page); in page_idle_get_page()
42 spin_lock_irq(&pgdat->lru_lock); in page_idle_get_page()
47 spin_unlock_irq(&pgdat->lru_lock); in page_idle_get_page()
Dshuffle.c154 void __meminit __shuffle_free_memory(pg_data_t *pgdat) in __shuffle_free_memory() argument
158 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in __shuffle_free_memory()
/kernel/linux/linux-5.10/include/linux/
Dmemory_hotplug.h177 extern void arch_free_nodedata(pg_data_t *pgdat);
178 extern void arch_refresh_nodedata(int nid, pg_data_t *pgdat);
183 #define arch_free_nodedata(pgdat) generic_free_nodedata(pgdat) argument
200 #define generic_free_nodedata(pgdat) kfree(pgdat) argument
203 static inline void arch_refresh_nodedata(int nid, pg_data_t *pgdat) in arch_refresh_nodedata() argument
205 node_data[nid] = pgdat; in arch_refresh_nodedata()
216 static inline void generic_free_nodedata(pg_data_t *pgdat) in generic_free_nodedata() argument
219 static inline void arch_refresh_nodedata(int nid, pg_data_t *pgdat) in arch_refresh_nodedata() argument
226 extern void __init register_page_bootmem_info_node(struct pglist_data *pgdat);
228 static inline void register_page_bootmem_info_node(struct pglist_data *pgdat) in register_page_bootmem_info_node() argument
[all …]
Dmmzone.h324 struct pglist_data *pgdat; member
851 #define pgdat_page_nr(pgdat, pagenr) ((pgdat)->node_mem_map + (pagenr)) argument
853 #define pgdat_page_nr(pgdat, pagenr) pfn_to_page((pgdat)->node_start_pfn + (pagenr)) argument
860 static inline struct lruvec *node_lruvec(struct pglist_data *pgdat) in node_lruvec() argument
862 return &pgdat->__lruvec; in node_lruvec()
865 static inline unsigned long pgdat_end_pfn(pg_data_t *pgdat) in pgdat_end_pfn() argument
867 return pgdat->node_start_pfn + pgdat->node_spanned_pages; in pgdat_end_pfn()
870 static inline bool pgdat_is_empty(pg_data_t *pgdat) in pgdat_is_empty() argument
872 return !pgdat->node_start_pfn && !pgdat->node_spanned_pages; in pgdat_is_empty()
877 void build_all_zonelists(pg_data_t *pgdat);
[all …]
Dvmstat.h179 static inline void node_page_state_add(long x, struct pglist_data *pgdat, in node_page_state_add() argument
182 atomic_long_add(x, &pgdat->vm_stat[item]); in node_page_state_add()
252 extern unsigned long node_page_state(struct pglist_data *pgdat,
254 extern unsigned long node_page_state_pages(struct pglist_data *pgdat,
298 void set_pgdat_percpu_threshold(pg_data_t *pgdat,
312 static inline void __mod_node_page_state(struct pglist_data *pgdat, in __mod_node_page_state() argument
320 node_page_state_add(delta, pgdat, item); in __mod_node_page_state()
329 static inline void __inc_node_state(struct pglist_data *pgdat, enum node_stat_item item) in __inc_node_state() argument
331 atomic_long_inc(&pgdat->vm_stat[item]); in __inc_node_state()
341 static inline void __dec_node_state(struct pglist_data *pgdat, enum node_stat_item item) in __dec_node_state() argument
[all …]
Dcompaction.h97 extern void reset_isolation_suitable(pg_data_t *pgdat);
184 extern void wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx);
187 static inline void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() argument
234 static inline void wakeup_kcompactd(pg_data_t *pgdat, in wakeup_kcompactd() argument
Dmemcontrol.h53 pg_data_t *pgdat; member
480 struct pglist_data *pgdat) in mem_cgroup_lruvec() argument
486 lruvec = &pgdat->__lruvec; in mem_cgroup_lruvec()
493 mz = mem_cgroup_nodeinfo(memcg, pgdat->node_id); in mem_cgroup_lruvec()
501 if (unlikely(lruvec->pgdat != pgdat)) in mem_cgroup_lruvec()
502 lruvec->pgdat = pgdat; in mem_cgroup_lruvec()
880 pg_data_t *pgdat = page_pgdat(page); in __mod_lruvec_page_state() local
885 __mod_node_page_state(pgdat, idx, val); in __mod_lruvec_page_state()
893 __mod_node_page_state(pgdat, idx, val); in __mod_lruvec_page_state()
897 lruvec = mem_cgroup_lruvec(head->mem_cgroup, pgdat); in __mod_lruvec_page_state()
[all …]
Dnode.h124 struct pglist_data *pgdat = NODE_DATA(nid); in register_one_node() local
125 unsigned long start_pfn = pgdat->node_start_pfn; in register_one_node()
126 unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages; in register_one_node()
Dmemcg_policy.h21 void shrink_anon_memcg(struct pglist_data *pgdat,
24 bool shrink_node_hyperhold(struct pglist_data *pgdat, struct scan_control *sc);
/kernel/linux/linux-5.10/arch/m68k/include/asm/
Dpage_mm.h158 struct pglist_data *pgdat; \
159 pgdat = __virt_to_node((unsigned long)pfn_to_virt(__pfn)); \
160 pgdat->node_mem_map + (__pfn - pgdat->node_start_pfn); \
164 struct pglist_data *pgdat; \
165 pgdat = &pg_data_map[page_to_nid(__p)]; \
166 ((__p) - pgdat->node_mem_map) + pgdat->node_start_pfn; \
/kernel/linux/linux-5.10/drivers/base/
Dnode.c370 struct pglist_data *pgdat = NODE_DATA(nid); in node_read_meminfo() local
375 sreclaimable = node_page_state_pages(pgdat, NR_SLAB_RECLAIMABLE_B); in node_read_meminfo()
376 sunreclaimable = node_page_state_pages(pgdat, NR_SLAB_UNRECLAIMABLE_B); in node_read_meminfo()
392 nid, K(node_page_state(pgdat, NR_ACTIVE_ANON) + in node_read_meminfo()
393 node_page_state(pgdat, NR_ACTIVE_FILE)), in node_read_meminfo()
394 nid, K(node_page_state(pgdat, NR_INACTIVE_ANON) + in node_read_meminfo()
395 node_page_state(pgdat, NR_INACTIVE_FILE)), in node_read_meminfo()
396 nid, K(node_page_state(pgdat, NR_ACTIVE_ANON)), in node_read_meminfo()
397 nid, K(node_page_state(pgdat, NR_INACTIVE_ANON)), in node_read_meminfo()
398 nid, K(node_page_state(pgdat, NR_ACTIVE_FILE)), in node_read_meminfo()
[all …]
/kernel/linux/linux-5.10/lib/
Dshow_mem.c13 pg_data_t *pgdat; in show_mem() local
19 for_each_online_pgdat(pgdat) { in show_mem()
23 struct zone *zone = &pgdat->node_zones[zoneid]; in show_mem()
/kernel/linux/linux-5.10/arch/ia64/include/asm/
Dnodedata.h57 #define LOCAL_DATA_ADDR(pgdat) \ argument
58 ((struct ia64_node_data *)((u64)(pgdat) + \

12