| /mm/ |
| D | sparse-vmemmap.c | 40 static void * __ref __earlyonly_bootmem_alloc(int node, in __earlyonly_bootmem_alloc() 49 void * __meminit vmemmap_alloc_block(unsigned long size, int node) in vmemmap_alloc_block() 77 void * __meminit vmemmap_alloc_block_buf(unsigned long size, int node, in vmemmap_alloc_block_buf() 133 void __meminit vmemmap_verify(pte_t *pte, int node, in vmemmap_verify() 144 pte_t * __meminit vmemmap_pte_populate(pmd_t *pmd, unsigned long addr, int node, in vmemmap_pte_populate() 176 static void * __meminit vmemmap_alloc_block_zero(unsigned long size, int node) in vmemmap_alloc_block_zero() 191 pmd_t * __meminit vmemmap_pmd_populate(pud_t *pud, unsigned long addr, int node) in vmemmap_pmd_populate() 208 pud_t * __meminit vmemmap_pud_populate(p4d_t *p4d, unsigned long addr, int node) in vmemmap_pud_populate() 225 p4d_t * __meminit vmemmap_p4d_populate(pgd_t *pgd, unsigned long addr, int node) in vmemmap_p4d_populate() 238 pgd_t * __meminit vmemmap_pgd_populate(unsigned long addr, int node) in vmemmap_pgd_populate() [all …]
|
| D | memory-tiers.c | 260 static struct memory_tier *__node_get_memory_tier(int node) in __node_get_memory_tier() 277 bool node_is_toptier(int node) in node_is_toptier() 329 int next_demotion_node(int node) in next_demotion_node() 368 int node; in disable_all_demotion_targets() local 391 int node; in dump_demotion_targets() local 418 int target = NUMA_NO_NODE, node; in establish_demotion_targets() local 514 static inline void __init_node_memory_type(int node, struct memory_dev_type *memtype) in __init_node_memory_type() 532 static struct memory_tier *set_node_memory_tier(int node) in set_node_memory_tier() 570 static bool clear_node_memory_tier(int node) in clear_node_memory_tier() 636 void init_node_memory_type(int node, struct memory_dev_type *memtype) in init_node_memory_type() [all …]
|
| D | interval_tree.c | 28 void vma_interval_tree_insert_after(struct vm_area_struct *node, in vma_interval_tree_insert_after() 85 void anon_vma_interval_tree_remove(struct anon_vma_chain *node, in anon_vma_interval_tree_remove() 99 anon_vma_interval_tree_iter_next(struct anon_vma_chain *node, in anon_vma_interval_tree_iter_next() 106 void anon_vma_interval_tree_verify(struct anon_vma_chain *node) in anon_vma_interval_tree_verify()
|
| D | shmem_quota.c | 47 struct rb_node node; member 98 struct rb_node *node; in shmem_free_file_info() local 117 struct rb_node *node; in shmem_get_next_id() local 268 struct rb_node *node; in shmem_release_dquot() local
|
| D | slub.c | 420 static inline struct kmem_cache_node *get_node(struct kmem_cache *s, int node) in get_node() 928 int node; in get_each_kmemcache_object() local 1588 static inline void inc_slabs_node(struct kmem_cache *s, int node, int objects) in inc_slabs_node() 1595 static inline void dec_slabs_node(struct kmem_cache *s, int node, int objects) in dec_slabs_node() 1946 static inline void inc_slabs_node(struct kmem_cache *s, int node, in inc_slabs_node() 1948 static inline void dec_slabs_node(struct kmem_cache *s, int node, in dec_slabs_node() 2485 static inline struct slab *alloc_slab_page(gfp_t flags, int node, in alloc_slab_page() 2644 static struct slab *allocate_slab(struct kmem_cache *s, gfp_t flags, int node) in allocate_slab() 2712 static struct slab *new_slab(struct kmem_cache *s, gfp_t flags, int node) in new_slab() 3020 static struct slab *get_partial(struct kmem_cache *s, int node, in get_partial() [all …]
|
| D | migrate.c | 2182 static int do_move_pages_to_node(struct list_head *pagelist, int node) in do_move_pages_to_node() 2198 static int __add_folio_for_migration(struct folio *folio, int node, in __add_folio_for_migration() 2236 int node, struct list_head *pagelist, bool migrate_all) in add_folio_for_migration() 2262 static int move_pages_and_store_status(int node, in move_pages_and_store_status() 2308 int node; in do_pages_move() local 2624 struct vm_area_struct *vma, int node) in migrate_misplaced_folio_prepare() 2690 int node) in migrate_misplaced_folio()
|
| D | hugetlb.c | 1357 int node = NUMA_NO_NODE; in dequeue_hugetlb_folio_nodemask() local 1499 #define for_each_node_mask_to_alloc(next_node, nr_nodes, node, mask) \ argument 1505 #define for_each_node_mask_to_free(hs, nr_nodes, node, mask) \ argument 1526 int node; in alloc_gigantic_folio() local 1721 struct llist_node *node; in free_hpage_workfn() local 2126 int nr_nodes, node; in alloc_pool_huge_folio() local 2150 int nr_nodes, node; in remove_pool_hugetlb_folio() local 2459 int node; in gather_surplus_pages() local 3114 int nr_nodes, node = nid; in __alloc_bootmem_huge_page() local 3562 int nr_nodes, node; in adjust_pool_surplus() local [all …]
|
| D | nommu.c | 137 pgprot_t prot, unsigned long vm_flags, int node, in __vmalloc_node_range_noprof() 144 int node, const void *caller) in __vmalloc_node_noprof() 242 void *vmalloc_node_noprof(unsigned long size, int node) in vmalloc_node_noprof() 260 void *vzalloc_node_noprof(unsigned long size, int node) in vzalloc_node_noprof() 312 void *vm_map_ram(struct page **pages, unsigned int count, int node) in vm_map_ram()
|
| D | vmalloc.c | 988 get_subtree_max_size(struct rb_node *node) in get_subtree_max_size() 1508 struct rb_node *node; in find_vmap_lowest_match() local 1829 preload_this_cpu_lock(spinlock_t *lock, gfp_t gfp_mask, int node) in preload_this_cpu_lock() 1968 int node, gfp_t gfp_mask, in alloc_vmap_area() 2625 int node, err; in new_vmap_block() local 2987 void *vm_map_ram(struct page **pages, unsigned int count, int node) in vm_map_ram() 3119 unsigned long start, unsigned long end, int node, in __get_vm_area_node() 3638 int node) in __vmalloc_area_node() 3781 pgprot_t prot, unsigned long vm_flags, int node, in __vmalloc_node_range_noprof() 3927 gfp_t gfp_mask, int node, const void *caller) in __vmalloc_node_noprof() [all …]
|
| D | workingset.c | 624 void workingset_update_node(struct xa_node *node) in workingset_update_node() 716 struct xa_node *node = container_of(item, struct xa_node, private_list); in shadow_lru_isolate() local
|
| D | bootmem_info.c | 105 int node = pgdat->node_id; in register_page_bootmem_info_node() local
|
| D | vmstat.c | 987 unsigned long sum_zone_node_page_state(int node, in sum_zone_node_page_state() 1001 unsigned long sum_zone_numa_event_state(int node, in sum_zone_numa_event_state() 1474 loff_t node = *pos; in frag_start() local 2127 int node; in init_cpu_node_state() local 2155 int node; in vmstat_cpu_dead() local
|
| D | mempolicy.c | 152 static u8 get_il_weight(int node) in get_il_weight() 176 int numa_nearest_node(int node, unsigned int state) in numa_nearest_node() 202 int node; in get_task_policy() local 1883 unsigned int node; in weighted_interleave_nodes() local 1927 int node = numa_mem_id(); in mempolicy_slab_node() local 2409 int nnodes, node; in alloc_pages_bulk_array_weighted_interleave() local 2865 static void sp_node_init(struct sp_node *node, unsigned long start, in sp_node_init()
|
| D | mmu_notifier.c | 100 struct interval_tree_node *node; in mn_itree_inv_start_range() local 122 struct interval_tree_node *node; in mn_itree_inv_next() local
|
| D | cma_debug.c | 19 struct hlist_node node; member
|
| D | zswap.c | 165 struct hlist_node node; member 814 static int zswap_cpu_comp_prepare(unsigned int cpu, struct hlist_node *node) in zswap_cpu_comp_prepare() 875 static int zswap_cpu_comp_dead(unsigned int cpu, struct hlist_node *node) in zswap_cpu_comp_dead()
|
| D | vmpressure.c | 154 struct list_head node; member
|
| D | hugetlb_cgroup.c | 138 int node; in hugetlb_cgroup_free() local 150 int node; in hugetlb_cgroup_css_alloc() local
|
| D | memory_hotplug.c | 690 static void node_states_set_node(int node, struct memory_notify *arg) in node_states_set_node() 1951 static void node_states_clear_node(int node, struct memory_notify *arg) in node_states_clear_node() 1977 const int node = zone_to_nid(zone); in offline_pages() local
|
| D | khugepaged.c | 1073 int node = hpage_collapse_find_target_node(cc); in alloc_charge_folio() local 1272 int node = NUMA_NO_NODE, unmapped = 0; in hpage_collapse_scan_pmd() local 2254 int node = NUMA_NO_NODE; in hpage_collapse_scan_file() local
|
| D | ksm.c | 162 struct rb_node node; /* when node of stable tree */ member 216 struct rb_node node; /* when node of unstable tree */ member 3192 struct rb_node *node; in ksm_check_stable_tree() local
|
| D | memcontrol.c | 3488 static bool alloc_mem_cgroup_per_node_info(struct mem_cgroup *memcg, int node) in alloc_mem_cgroup_per_node_info() 3518 static void free_mem_cgroup_per_node_info(struct mem_cgroup *memcg, int node) in free_mem_cgroup_per_node_info() 3532 int node; in __mem_cgroup_free() local 3556 int node, cpu; in mem_cgroup_alloc() local
|
| D | compaction.c | 3085 int compaction_register_node(struct node *node) in compaction_register_node() 3090 void compaction_unregister_node(struct node *node) in compaction_unregister_node()
|
| D | mm_init.c | 819 int zone, int node) in init_unavailable_range() 946 struct pglist_data *node = NODE_DATA(nid); in memmap_init() local
|
| /mm/kasan/ |
| D | init.c | 80 static __init void *early_alloc(size_t size, int node) in early_alloc()
|