/mm/ |
D | cma.c | 420 unsigned long mask, offset; in cma_alloc() local 436 mask = cma_bitmap_aligned_mask(cma, align); in cma_alloc() 447 bitmap_maxno, start, bitmap_count, mask, in cma_alloc() 478 start = bitmap_no + mask + 1; in cma_alloc()
|
D | swap_state.c | 562 unsigned long mask; in swapin_readahead() local 566 mask = swapin_nr_pages(offset) - 1; in swapin_readahead() 567 if (!mask) in swapin_readahead() 572 start_offset = offset & ~mask; in swapin_readahead() 573 end_offset = offset | mask; in swapin_readahead()
|
D | oom_kill.c | 68 const nodemask_t *mask) in has_intersects_mems_allowed() argument 75 if (mask) { in has_intersects_mems_allowed() 82 ret = mempolicy_nodemask_intersects(tsk, mask); in has_intersects_mems_allowed() 99 const nodemask_t *mask) in has_intersects_mems_allowed() argument
|
D | mempolicy.c | 1351 static int copy_nodes_to_user(unsigned long __user *mask, unsigned long maxnode, in copy_nodes_to_user() argument 1360 if (clear_user((char __user *)mask + nbytes, copy - nbytes)) in copy_nodes_to_user() 1364 return copy_to_user(mask, nodes_addr(*nodes), copy) ? -EFAULT : 0; in copy_nodes_to_user() 1885 bool init_nodemask_of_mempolicy(nodemask_t *mask) in init_nodemask_of_mempolicy() argument 1890 if (!(mask && current->mempolicy)) in init_nodemask_of_mempolicy() 1901 init_nodemask_of_node(mask, nid); in init_nodemask_of_mempolicy() 1907 *mask = mempolicy->v.nodes; in init_nodemask_of_mempolicy() 1930 const nodemask_t *mask) in mempolicy_nodemask_intersects() argument 1935 if (!mask) in mempolicy_nodemask_intersects() 1953 ret = nodes_intersects(mempolicy->v.nodes, *mask); in mempolicy_nodemask_intersects()
|
D | memory-failure.c | 802 unsigned long mask; member 1038 if ((p->flags & ps->mask) == ps->res) in identify_page_state() 1043 if (!ps->mask) in identify_page_state() 1045 if ((page_flags & ps->mask) == ps->res) in identify_page_state()
|
D | compaction.c | 2094 const struct cpumask *mask; in kcompactd_cpu_online() local 2096 mask = cpumask_of_node(pgdat->node_id); in kcompactd_cpu_online() 2098 if (cpumask_any_and(cpu_online_mask, mask) < nr_cpu_ids) in kcompactd_cpu_online() 2100 set_cpus_allowed_ptr(pgdat->kcompactd, mask); in kcompactd_cpu_online()
|
D | page_alloc.c | 422 unsigned long mask) in __get_pfnblock_flags_mask() argument 435 return (word >> (BITS_PER_LONG - bitidx - 1)) & mask; in __get_pfnblock_flags_mask() 440 unsigned long mask) in get_pfnblock_flags_mask() argument 442 return __get_pfnblock_flags_mask(page, pfn, end_bitidx, mask); in get_pfnblock_flags_mask() 461 unsigned long mask) in set_pfnblock_flags_mask() argument 477 mask <<= (BITS_PER_LONG - bitidx - 1); in set_pfnblock_flags_mask() 482 old_word = cmpxchg(&bitmap[word_bitidx], word, (word & ~mask) | flags); in set_pfnblock_flags_mask() 6248 unsigned long start, end, mask; in node_map_pfn_alignment() local 6264 mask = ~((1 << __ffs(start)) - 1); in node_map_pfn_alignment() 6265 while (mask && last_end <= (start & (mask << 1))) in node_map_pfn_alignment() [all …]
|
D | hugetlb.c | 1034 #define for_each_node_mask_to_alloc(hs, nr_nodes, node, mask) \ argument 1035 for (nr_nodes = nodes_weight(*mask); \ 1037 ((node = hstate_next_node_to_alloc(hs, mask)) || 1); \ 1040 #define for_each_node_mask_to_free(hs, nr_nodes, node, mask) \ argument 1041 for (nr_nodes = nodes_weight(*mask); \ 1043 ((node = hstate_next_node_to_free(hs, mask)) || 1); \ 2839 h->mask = ~((1ULL << (order + PAGE_SHIFT)) - 1); in hugetlb_add_hstate()
|
D | internal.h | 461 static inline int node_reclaim(struct pglist_data *pgdat, gfp_t mask, in node_reclaim() argument
|
D | vmscan.c | 3731 const struct cpumask *mask; in kswapd_cpu_online() local 3733 mask = cpumask_of_node(pgdat->node_id); in kswapd_cpu_online() 3735 if (cpumask_any_and(cpu_online_mask, mask) < nr_cpu_ids) in kswapd_cpu_online() 3737 set_cpus_allowed_ptr(pgdat->kswapd, mask); in kswapd_cpu_online()
|
D | memory.c | 3564 unsigned long address = vmf->address, nr_pages, mask; in do_fault_around() local 3570 mask = ~(nr_pages * PAGE_SIZE - 1) & PAGE_MASK; in do_fault_around() 3572 vmf->address = max(address & mask, vmf->vma->vm_start); in do_fault_around()
|
D | slab.c | 988 const struct cpumask *mask = cpumask_of_node(node); in cpuup_canceled() local 1012 if (!cpumask_empty(mask)) { in cpuup_canceled()
|
D | memcontrol.c | 1548 static void mem_cgroup_oom(struct mem_cgroup *memcg, gfp_t mask, int order) in mem_cgroup_oom() argument 1568 current->memcg_oom_gfp_mask = mask; in mem_cgroup_oom()
|