| /kernel/linux/linux-5.10/include/linux/ |
| D | oom.h | 13 struct zonelist; 31 struct zonelist *zonelist; member
|
| D | mmzone.h | 712 struct zonelist { struct 750 struct zonelist node_zonelists[MAX_ZONELISTS]; 1136 static inline struct zoneref *first_zones_zonelist(struct zonelist *zonelist, in first_zones_zonelist() argument 1140 return next_zones_zonelist(zonelist->_zonerefs, in first_zones_zonelist()
|
| D | swap.h | 359 extern unsigned long try_to_free_pages(struct zonelist *zonelist, int order,
|
| D | gfp.h | 497 static inline struct zonelist *node_zonelist(int nid, gfp_t flags) in node_zonelist()
|
| /kernel/linux/linux-5.10/mm/ |
| D | mm_init.c | 38 struct zonelist *zonelist; in mminit_verify_zonelist() local 47 zonelist = &pgdat->node_zonelists[listid]; in mminit_verify_zonelist() 58 for_each_zone_zonelist(zone, z, zonelist, zoneid) in mminit_verify_zonelist()
|
| D | vmscan.c | 2888 static void shrink_zones(struct zonelist *zonelist, struct scan_control *sc) in shrink_zones() argument 2908 for_each_zone_zonelist_nodemask(zone, z, zonelist, in shrink_zones() 3013 static unsigned long do_try_to_free_pages(struct zonelist *zonelist, in do_try_to_free_pages() argument 3030 shrink_zones(zonelist, sc); in do_try_to_free_pages() 3047 for_each_zone_zonelist_nodemask(zone, z, zonelist, sc->reclaim_idx, in do_try_to_free_pages() 3150 static bool throttle_direct_reclaim(gfp_t gfp_mask, struct zonelist *zonelist, in throttle_direct_reclaim() argument 3188 for_each_zone_zonelist_nodemask(zone, z, zonelist, in throttle_direct_reclaim() 3234 unsigned long try_to_free_pages(struct zonelist *zonelist, int order, in try_to_free_pages() argument 3263 if (throttle_direct_reclaim(sc.gfp_mask, zonelist, nodemask)) in try_to_free_pages() 3269 nr_reclaimed = do_try_to_free_pages(zonelist, &sc); in try_to_free_pages() [all …]
|
| D | page_alloc.c | 2687 struct zonelist *zonelist = ac->zonelist; in unreserve_highatomic_pageblock() local 2695 for_each_zone_zonelist_nodemask(zone, z, zonelist, ac->highest_zoneidx, in unreserve_highatomic_pageblock() 4047 .zonelist = ac->zonelist, in __alloc_pages_may_oom() 4296 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in should_compact_retry() 4397 progress = try_to_free_pages(ac->zonelist, order, gfp_mask, in __perform_reclaim() 4454 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, highest_zoneidx, in wake_all_kswapds() 4590 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in should_reclaim_retry() 4728 ac->preferred_zoneref = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() 4815 ac->preferred_zoneref = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() 4965 ac->zonelist = node_zonelist(preferred_nid, gfp_mask); in prepare_alloc_pages() [all …]
|
| D | internal.h | 270 struct zonelist *zonelist; member
|
| D | oom_kill.c | 270 if (!oc->zonelist) in constrained_alloc() 294 for_each_zone_zonelist_nodemask(zone, z, oc->zonelist, in constrained_alloc()
|
| D | mempolicy.c | 1941 struct zonelist *zonelist; in mempolicy_slab_node() local 1943 zonelist = &NODE_DATA(node)->node_zonelists[ZONELIST_FALLBACK]; in mempolicy_slab_node() 1944 z = first_zones_zonelist(zonelist, highest_zoneidx, in mempolicy_slab_node()
|
| D | slab.c | 3097 struct zonelist *zonelist; in fallback_alloc() local 3111 zonelist = node_zonelist(mempolicy_slab_node(), flags); in fallback_alloc() 3118 for_each_zone_zonelist(zone, z, zonelist, highest_zoneidx) { in fallback_alloc()
|
| D | compaction.c | 2169 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in compaction_zonelist_suitable() 2484 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in try_to_compact_pages()
|
| D | hugetlb.c | 1104 struct zonelist *zonelist; in dequeue_huge_page_nodemask() local 1109 zonelist = node_zonelist(nid, gfp_mask); in dequeue_huge_page_nodemask() 1113 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nmask) { in dequeue_huge_page_nodemask()
|
| D | slub.c | 2009 struct zonelist *zonelist; in get_any_partial() local 2040 zonelist = node_zonelist(mempolicy_slab_node(), flags); in get_any_partial() 2041 for_each_zone_zonelist(zone, z, zonelist, highest_zoneidx) { in get_any_partial()
|
| D | memcontrol.c | 1743 .zonelist = NULL, in mem_cgroup_out_of_memory()
|
| /kernel/linux/linux-5.10/Documentation/vm/ |
| D | numa.rst | 76 an ordered "zonelist". A zonelist specifies the zones/nodes to visit when a 86 a default Node ordered zonelist. This means it tries to fallback to other zones 91 Linux will attempt to allocate from the first node in the appropriate zonelist 94 nodes' zones in the selected zonelist looking for the first zone in the list 122 zonelist--will not be the node itself. Rather, it will be the node that the
|
| /kernel/linux/linux-5.10/drivers/tty/ |
| D | sysrq.c | 386 .zonelist = node_zonelist(first_memory_node, gfp_mask), in moom_callback()
|
| /kernel/linux/linux-5.10/Documentation/admin-guide/sysctl/ |
| D | vm.rst | 609 In non-NUMA case, a zonelist for GFP_KERNEL is ordered as following. 615 Assume 2 node NUMA and below is zonelist of Node(0)'s GFP_KERNEL::
|
| /kernel/linux/linux-5.10/Documentation/admin-guide/mm/ |
| D | numa_memory_policy.rst | 236 node zonelist.
|
| /kernel/linux/linux-5.10/Documentation/admin-guide/ |
| D | kernel-parameters.txt | 3517 numa_zonelist_order= [KNL, BOOT] Select zonelist order for NUMA.
|