/mm/ |
D | ioremap.c | 66 pgtbl_mod_mask *mask) in ioremap_pte_range() argument 72 pte = pte_alloc_kernel_track(pmd, addr, mask); in ioremap_pte_range() 80 *mask |= PGTBL_PTE_MODIFIED; in ioremap_pte_range() 108 pgtbl_mod_mask *mask) in ioremap_pmd_range() argument 113 pmd = pmd_alloc_track(&init_mm, pud, addr, mask); in ioremap_pmd_range() 120 *mask |= PGTBL_PMD_MODIFIED; in ioremap_pmd_range() 124 if (ioremap_pte_range(pmd, addr, next, phys_addr, prot, mask)) in ioremap_pmd_range() 154 pgtbl_mod_mask *mask) in ioremap_pud_range() argument 159 pud = pud_alloc_track(&init_mm, p4d, addr, mask); in ioremap_pud_range() 166 *mask |= PGTBL_PUD_MODIFIED; in ioremap_pud_range() [all …]
|
D | vmalloc.c | 74 pgtbl_mod_mask *mask) in vunmap_pte_range() argument 83 *mask |= PGTBL_PTE_MODIFIED; in vunmap_pte_range() 87 pgtbl_mod_mask *mask) in vunmap_pmd_range() argument 99 *mask |= PGTBL_PMD_MODIFIED; in vunmap_pmd_range() 105 vunmap_pte_range(pmd, addr, next, mask); in vunmap_pmd_range() 112 pgtbl_mod_mask *mask) in vunmap_pud_range() argument 124 *mask |= PGTBL_PUD_MODIFIED; in vunmap_pud_range() 130 vunmap_pmd_range(pud, addr, next, mask); in vunmap_pud_range() 135 pgtbl_mod_mask *mask) in vunmap_p4d_range() argument 147 *mask |= PGTBL_P4D_MODIFIED; in vunmap_p4d_range() [all …]
|
D | pgalloc-track.h | 46 #define pte_alloc_kernel_track(pmd, address, mask) \ argument 48 (__pte_alloc_kernel(pmd) || ({*(mask)|=PGTBL_PMD_MODIFIED;0;})))?\
|
D | cma.c | 440 unsigned long mask, offset; in cma_alloc() local 465 mask = cma_bitmap_aligned_mask(cma, align); in cma_alloc() 479 bitmap_maxno, start, bitmap_count, mask, in cma_alloc() 551 start = bitmap_no + mask + 1; in cma_alloc()
|
D | swap_state.c | 661 unsigned long mask; in swap_cluster_readahead() local 668 mask = swapin_nr_pages(offset) - 1; in swap_cluster_readahead() 669 if (!mask) in swap_cluster_readahead() 681 start_offset = offset & ~mask; in swap_cluster_readahead() 682 end_offset = offset | mask; in swap_cluster_readahead()
|
D | mremap.c | 394 unsigned long next, extent, mask, size; in get_extent() local 399 mask = PMD_MASK; in get_extent() 403 mask = PUD_MASK; in get_extent() 411 next = (old_addr + size) & mask; in get_extent() 416 next = (new_addr + size) & mask; in get_extent()
|
D | readahead.c | 120 gfp_t mask = mapping_gfp_mask(x) | __GFP_NORETRY | __GFP_NOWARN; in readahead_gfp_mask() local 122 trace_android_rvh_set_readahead_gfp_mask(&mask); in readahead_gfp_mask() 123 return mask; in readahead_gfp_mask()
|
D | mempolicy.c | 1443 static int copy_nodes_to_user(unsigned long __user *mask, unsigned long maxnode, in copy_nodes_to_user() argument 1452 if (clear_user((char __user *)mask + nbytes, copy - nbytes)) in copy_nodes_to_user() 1456 return copy_to_user(mask, nodes_addr(*nodes), copy) ? -EFAULT : 0; in copy_nodes_to_user() 2064 bool init_nodemask_of_mempolicy(nodemask_t *mask) in init_nodemask_of_mempolicy() argument 2069 if (!(mask && current->mempolicy)) in init_nodemask_of_mempolicy() 2080 init_nodemask_of_node(mask, nid); in init_nodemask_of_mempolicy() 2085 *mask = mempolicy->v.nodes; in init_nodemask_of_mempolicy() 2108 const nodemask_t *mask) in mempolicy_nodemask_intersects() argument 2113 if (!mask) in mempolicy_nodemask_intersects() 2131 ret = nodes_intersects(mempolicy->v.nodes, *mask); in mempolicy_nodemask_intersects()
|
D | memory.c | 2470 pgtbl_mod_mask *mask) in apply_to_pte_range() argument 2478 pte_alloc_kernel_track(pmd, addr, mask) : in apply_to_pte_range() 2501 *mask |= PGTBL_PTE_MODIFIED; in apply_to_pte_range() 2513 pgtbl_mod_mask *mask) in apply_to_pmd_range() argument 2522 pmd = pmd_alloc_track(mm, pud, addr, mask); in apply_to_pmd_range() 2532 create, mask); in apply_to_pmd_range() 2543 pgtbl_mod_mask *mask) in apply_to_pud_range() argument 2550 pud = pud_alloc_track(mm, p4d, addr, mask); in apply_to_pud_range() 2560 create, mask); in apply_to_pud_range() 2571 pgtbl_mod_mask *mask) in apply_to_p4d_range() argument [all …]
|
D | oom_kill.c | 98 const nodemask_t *mask = oc->nodemask; in oom_cpuset_eligible() local 105 if (mask) { in oom_cpuset_eligible() 112 ret = mempolicy_nodemask_intersects(tsk, mask); in oom_cpuset_eligible()
|
D | memory-failure.c | 862 unsigned long mask; member 1119 if ((p->flags & ps->mask) == ps->res) in identify_page_state() 1124 if (!ps->mask) in identify_page_state() 1126 if ((page_flags & ps->mask) == ps->res) in identify_page_state()
|
D | page_alloc.c | 517 unsigned long mask) in __get_pfnblock_flags_mask() argument 529 return (word >> bitidx) & mask; in __get_pfnblock_flags_mask() 533 unsigned long mask) in get_pfnblock_flags_mask() argument 535 return __get_pfnblock_flags_mask(page, pfn, mask); in get_pfnblock_flags_mask() 570 unsigned long mask) in set_pfnblock_flags_mask() argument 586 mask <<= bitidx; in set_pfnblock_flags_mask() 591 old_word = cmpxchg(&bitmap[word_bitidx], word, (word & ~mask) | flags); in set_pfnblock_flags_mask() 7413 unsigned long start, end, mask; in node_map_pfn_alignment() local 7429 mask = ~((1 << __ffs(start)) - 1); in node_map_pfn_alignment() 7430 while (mask && last_end <= (start & (mask << 1))) in node_map_pfn_alignment() [all …]
|
D | compaction.c | 2978 const struct cpumask *mask; in kcompactd_cpu_online() local 2980 mask = cpumask_of_node(pgdat->node_id); in kcompactd_cpu_online() 2982 if (cpumask_any_and(cpu_online_mask, mask) < nr_cpu_ids) in kcompactd_cpu_online() 2984 set_cpus_allowed_ptr(pgdat->kcompactd, mask); in kcompactd_cpu_online()
|
D | hugetlb.c | 1236 #define for_each_node_mask_to_alloc(hs, nr_nodes, node, mask) \ argument 1237 for (nr_nodes = nodes_weight(*mask); \ 1239 ((node = hstate_next_node_to_alloc(hs, mask)) || 1); \ 1242 #define for_each_node_mask_to_free(hs, nr_nodes, node, mask) \ argument 1243 for (nr_nodes = nodes_weight(*mask); \ 1245 ((node = hstate_next_node_to_free(hs, mask)) || 1); \ 3259 h->mask = ~((1ULL << (order + PAGE_SHIFT)) - 1); in hugetlb_add_hstate()
|
D | internal.h | 564 static inline int node_reclaim(struct pglist_data *pgdat, gfp_t mask, in node_reclaim() argument
|
D | slab.c | 956 const struct cpumask *mask = cpumask_of_node(node); in cpuup_canceled() local 978 if (!cpumask_empty(mask)) { in cpuup_canceled()
|
D | memcontrol.c | 1956 static enum oom_status mem_cgroup_oom(struct mem_cgroup *memcg, gfp_t mask, int order) in mem_cgroup_oom() argument 1989 current->memcg_oom_gfp_mask = mask; in mem_cgroup_oom() 2003 if (mem_cgroup_out_of_memory(memcg, mask, order)) in mem_cgroup_oom()
|