Home
last modified time | relevance | path

Searched refs:mask (Results 1 – 17 of 17) sorted by relevance

/mm/
Dioremap.c66 pgtbl_mod_mask *mask) in ioremap_pte_range() argument
72 pte = pte_alloc_kernel_track(pmd, addr, mask); in ioremap_pte_range()
80 *mask |= PGTBL_PTE_MODIFIED; in ioremap_pte_range()
108 pgtbl_mod_mask *mask) in ioremap_pmd_range() argument
113 pmd = pmd_alloc_track(&init_mm, pud, addr, mask); in ioremap_pmd_range()
120 *mask |= PGTBL_PMD_MODIFIED; in ioremap_pmd_range()
124 if (ioremap_pte_range(pmd, addr, next, phys_addr, prot, mask)) in ioremap_pmd_range()
154 pgtbl_mod_mask *mask) in ioremap_pud_range() argument
159 pud = pud_alloc_track(&init_mm, p4d, addr, mask); in ioremap_pud_range()
166 *mask |= PGTBL_PUD_MODIFIED; in ioremap_pud_range()
[all …]
Dvmalloc.c74 pgtbl_mod_mask *mask) in vunmap_pte_range() argument
83 *mask |= PGTBL_PTE_MODIFIED; in vunmap_pte_range()
87 pgtbl_mod_mask *mask) in vunmap_pmd_range() argument
99 *mask |= PGTBL_PMD_MODIFIED; in vunmap_pmd_range()
105 vunmap_pte_range(pmd, addr, next, mask); in vunmap_pmd_range()
112 pgtbl_mod_mask *mask) in vunmap_pud_range() argument
124 *mask |= PGTBL_PUD_MODIFIED; in vunmap_pud_range()
130 vunmap_pmd_range(pud, addr, next, mask); in vunmap_pud_range()
135 pgtbl_mod_mask *mask) in vunmap_p4d_range() argument
147 *mask |= PGTBL_P4D_MODIFIED; in vunmap_p4d_range()
[all …]
Dpgalloc-track.h46 #define pte_alloc_kernel_track(pmd, address, mask) \ argument
48 (__pte_alloc_kernel(pmd) || ({*(mask)|=PGTBL_PMD_MODIFIED;0;})))?\
Dcma.c440 unsigned long mask, offset; in cma_alloc() local
465 mask = cma_bitmap_aligned_mask(cma, align); in cma_alloc()
479 bitmap_maxno, start, bitmap_count, mask, in cma_alloc()
551 start = bitmap_no + mask + 1; in cma_alloc()
Dswap_state.c661 unsigned long mask; in swap_cluster_readahead() local
668 mask = swapin_nr_pages(offset) - 1; in swap_cluster_readahead()
669 if (!mask) in swap_cluster_readahead()
681 start_offset = offset & ~mask; in swap_cluster_readahead()
682 end_offset = offset | mask; in swap_cluster_readahead()
Dmremap.c394 unsigned long next, extent, mask, size; in get_extent() local
399 mask = PMD_MASK; in get_extent()
403 mask = PUD_MASK; in get_extent()
411 next = (old_addr + size) & mask; in get_extent()
416 next = (new_addr + size) & mask; in get_extent()
Dreadahead.c120 gfp_t mask = mapping_gfp_mask(x) | __GFP_NORETRY | __GFP_NOWARN; in readahead_gfp_mask() local
122 trace_android_rvh_set_readahead_gfp_mask(&mask); in readahead_gfp_mask()
123 return mask; in readahead_gfp_mask()
Dmempolicy.c1443 static int copy_nodes_to_user(unsigned long __user *mask, unsigned long maxnode, in copy_nodes_to_user() argument
1452 if (clear_user((char __user *)mask + nbytes, copy - nbytes)) in copy_nodes_to_user()
1456 return copy_to_user(mask, nodes_addr(*nodes), copy) ? -EFAULT : 0; in copy_nodes_to_user()
2064 bool init_nodemask_of_mempolicy(nodemask_t *mask) in init_nodemask_of_mempolicy() argument
2069 if (!(mask && current->mempolicy)) in init_nodemask_of_mempolicy()
2080 init_nodemask_of_node(mask, nid); in init_nodemask_of_mempolicy()
2085 *mask = mempolicy->v.nodes; in init_nodemask_of_mempolicy()
2108 const nodemask_t *mask) in mempolicy_nodemask_intersects() argument
2113 if (!mask) in mempolicy_nodemask_intersects()
2131 ret = nodes_intersects(mempolicy->v.nodes, *mask); in mempolicy_nodemask_intersects()
Dmemory.c2470 pgtbl_mod_mask *mask) in apply_to_pte_range() argument
2478 pte_alloc_kernel_track(pmd, addr, mask) : in apply_to_pte_range()
2501 *mask |= PGTBL_PTE_MODIFIED; in apply_to_pte_range()
2513 pgtbl_mod_mask *mask) in apply_to_pmd_range() argument
2522 pmd = pmd_alloc_track(mm, pud, addr, mask); in apply_to_pmd_range()
2532 create, mask); in apply_to_pmd_range()
2543 pgtbl_mod_mask *mask) in apply_to_pud_range() argument
2550 pud = pud_alloc_track(mm, p4d, addr, mask); in apply_to_pud_range()
2560 create, mask); in apply_to_pud_range()
2571 pgtbl_mod_mask *mask) in apply_to_p4d_range() argument
[all …]
Doom_kill.c98 const nodemask_t *mask = oc->nodemask; in oom_cpuset_eligible() local
105 if (mask) { in oom_cpuset_eligible()
112 ret = mempolicy_nodemask_intersects(tsk, mask); in oom_cpuset_eligible()
Dmemory-failure.c862 unsigned long mask; member
1119 if ((p->flags & ps->mask) == ps->res) in identify_page_state()
1124 if (!ps->mask) in identify_page_state()
1126 if ((page_flags & ps->mask) == ps->res) in identify_page_state()
Dpage_alloc.c517 unsigned long mask) in __get_pfnblock_flags_mask() argument
529 return (word >> bitidx) & mask; in __get_pfnblock_flags_mask()
533 unsigned long mask) in get_pfnblock_flags_mask() argument
535 return __get_pfnblock_flags_mask(page, pfn, mask); in get_pfnblock_flags_mask()
570 unsigned long mask) in set_pfnblock_flags_mask() argument
586 mask <<= bitidx; in set_pfnblock_flags_mask()
591 old_word = cmpxchg(&bitmap[word_bitidx], word, (word & ~mask) | flags); in set_pfnblock_flags_mask()
7413 unsigned long start, end, mask; in node_map_pfn_alignment() local
7429 mask = ~((1 << __ffs(start)) - 1); in node_map_pfn_alignment()
7430 while (mask && last_end <= (start & (mask << 1))) in node_map_pfn_alignment()
[all …]
Dcompaction.c2978 const struct cpumask *mask; in kcompactd_cpu_online() local
2980 mask = cpumask_of_node(pgdat->node_id); in kcompactd_cpu_online()
2982 if (cpumask_any_and(cpu_online_mask, mask) < nr_cpu_ids) in kcompactd_cpu_online()
2984 set_cpus_allowed_ptr(pgdat->kcompactd, mask); in kcompactd_cpu_online()
Dhugetlb.c1236 #define for_each_node_mask_to_alloc(hs, nr_nodes, node, mask) \ argument
1237 for (nr_nodes = nodes_weight(*mask); \
1239 ((node = hstate_next_node_to_alloc(hs, mask)) || 1); \
1242 #define for_each_node_mask_to_free(hs, nr_nodes, node, mask) \ argument
1243 for (nr_nodes = nodes_weight(*mask); \
1245 ((node = hstate_next_node_to_free(hs, mask)) || 1); \
3259 h->mask = ~((1ULL << (order + PAGE_SHIFT)) - 1); in hugetlb_add_hstate()
Dinternal.h564 static inline int node_reclaim(struct pglist_data *pgdat, gfp_t mask, in node_reclaim() argument
Dslab.c956 const struct cpumask *mask = cpumask_of_node(node); in cpuup_canceled() local
978 if (!cpumask_empty(mask)) { in cpuup_canceled()
Dmemcontrol.c1956 static enum oom_status mem_cgroup_oom(struct mem_cgroup *memcg, gfp_t mask, int order) in mem_cgroup_oom() argument
1989 current->memcg_oom_gfp_mask = mask; in mem_cgroup_oom()
2003 if (mem_cgroup_out_of_memory(memcg, mask, order)) in mem_cgroup_oom()