/mm/ |
D | percpu.c | 310 static unsigned long pcpu_off_to_block_index(int off) in pcpu_off_to_block_index() argument 312 return off / PCPU_BITMAP_BLOCK_BITS; in pcpu_off_to_block_index() 315 static unsigned long pcpu_off_to_block_off(int off) in pcpu_off_to_block_off() argument 317 return off & (PCPU_BITMAP_BLOCK_BITS - 1); in pcpu_off_to_block_off() 320 static unsigned long pcpu_block_off_to_off(int index, int off) in pcpu_block_off_to_off() argument 322 return index * PCPU_BITMAP_BLOCK_BITS + off; in pcpu_block_off_to_off() 1243 static void pcpu_free_area(struct pcpu_chunk *chunk, int off) in pcpu_free_area() argument 1253 bit_off = off / PCPU_MIN_ALLOC_SIZE; in pcpu_free_area() 1596 int slot, off, cpu, ret; in pcpu_alloc() local 1639 off = pcpu_find_block_fit(chunk, bits, bit_align, is_atomic); in pcpu_alloc() [all …]
|
D | zsmalloc.c | 982 unsigned long off = 0; in init_zspage() local 990 set_first_obj_offset(page, off); in init_zspage() 993 link = (struct link_free *)vaddr + off / sizeof(*link); in init_zspage() 995 while ((off += class->size) < PAGE_SIZE) { in init_zspage() 1017 off %= PAGE_SIZE; in init_zspage() 1133 struct page *pages[2], int off, int size) in __zs_map_object() argument 1146 sizes[0] = PAGE_SIZE - off; in __zs_map_object() 1151 memcpy(buf, addr + off, sizes[0]); in __zs_map_object() 1161 struct page *pages[2], int off, int size) in __zs_unmap_object() argument 1174 off += ZS_HANDLE_SIZE; in __zs_unmap_object() [all …]
|
D | hugetlb_cgroup.c | 289 char *buf, size_t nbytes, loff_t off) in hugetlb_cgroup_write() argument 320 char *buf, size_t nbytes, loff_t off) in hugetlb_cgroup_reset() argument
|
D | huge_memory.c | 527 loff_t off, unsigned long flags, unsigned long size) in __thp_get_unmapped_area() argument 529 loff_t off_end = off + len; in __thp_get_unmapped_area() 530 loff_t off_align = round_up(off, size); in __thp_get_unmapped_area() 537 if (len_pad < len || (off + len_pad) < off) in __thp_get_unmapped_area() 541 off >> PAGE_SHIFT, flags); in __thp_get_unmapped_area() 557 ret += (off - ret) & (size - 1); in __thp_get_unmapped_area() 565 loff_t off = (loff_t)pgoff << PAGE_SHIFT; in thp_get_unmapped_area() local 570 ret = __thp_get_unmapped_area(filp, addr, len, off, flags, PMD_SIZE); in thp_get_unmapped_area()
|
D | slub.c | 681 unsigned int off; /* Offset of last byte */ in print_trailer() local 703 off = get_info_end(s); in print_trailer() 706 off += 2 * sizeof(struct track); in print_trailer() 708 off += kasan_metadata_size(s); in print_trailer() 710 if (off != size_from_object(s)) in print_trailer() 712 print_section(KERN_ERR, "Padding ", p + off, in print_trailer() 713 size_from_object(s) - off); in print_trailer() 828 unsigned long off = get_info_end(s); /* The end of info */ in check_pad_bytes() local 832 off += 2 * sizeof(struct track); in check_pad_bytes() 834 off += kasan_metadata_size(s); in check_pad_bytes() [all …]
|
D | memory.c | 634 unsigned long off; in vm_normal_page() local 635 off = (addr - vma->vm_start) >> PAGE_SHIFT; in vm_normal_page() 636 if (pfn == vma->vm_pgoff + off) in vm_normal_page() 677 unsigned long off; in vm_normal_page_pmd() local 678 off = (addr - vma->vm_start) >> PAGE_SHIFT; in vm_normal_page_pmd() 679 if (pfn == vma->vm_pgoff + off) in vm_normal_page_pmd() 3590 int off; in do_fault_around() local 3597 off = ((address - vmf->address) >> PAGE_SHIFT) & (PTRS_PER_PTE - 1); in do_fault_around() 3598 start_pgoff -= off; in do_fault_around()
|
D | memcontrol.c | 3372 loff_t off) in mem_cgroup_force_empty_write() argument 3692 char *buf, size_t nbytes, loff_t off) in mem_cgroup_write() argument 3736 size_t nbytes, loff_t off) in mem_cgroup_reset() argument 4713 char *buf, size_t nbytes, loff_t off) in memcg_write_event_control() argument 6122 char *buf, size_t nbytes, loff_t off) in memory_min_write() argument 6145 char *buf, size_t nbytes, loff_t off) in memory_low_write() argument 6167 char *buf, size_t nbytes, loff_t off) in memory_high_write() argument 6197 char *buf, size_t nbytes, loff_t off) in memory_max_write() argument 6294 char *buf, size_t nbytes, loff_t off) in memory_oom_group_write() argument 7278 char *buf, size_t nbytes, loff_t off) in swap_max_write() argument
|
D | mempolicy.c | 1905 unsigned long off; in interleave_nid() local 1915 off = vma->vm_pgoff >> (shift - PAGE_SHIFT); in interleave_nid() 1916 off += (addr - vma->vm_start) >> shift; in interleave_nid() 1917 return offset_il_node(pol, off); in interleave_nid()
|
D | shmem.c | 1821 pgoff_t off; in shmem_getpage_gfp() local 1825 off = round_up(index, HPAGE_PMD_NR); in shmem_getpage_gfp() 1828 i_size >> PAGE_SHIFT >= off) in shmem_getpage_gfp() 4017 pgoff_t off; in shmem_huge_enabled() local 4032 off = round_up(vma->vm_pgoff, HPAGE_PMD_NR); in shmem_huge_enabled() 4035 i_size >> PAGE_SHIFT >= off) in shmem_huge_enabled()
|
D | vmalloc.c | 2999 unsigned long off; in remap_vmalloc_range_partial() local 3002 if (check_shl_overflow(pgoff, PAGE_SHIFT, &off)) in remap_vmalloc_range_partial() 3017 if (check_add_overflow(size, off, &end_index) || in remap_vmalloc_range_partial() 3020 kaddr += off; in remap_vmalloc_range_partial()
|
D | Kconfig.debug | 46 can be overridden by debug_pagealloc=off|on.
|
D | Kconfig | 350 more than it requires. To deal with this, mmap() is able to trim off 353 If trimming is enabled, the excess is trimmed off and returned to the 616 by starting one-off "pgdatinitX" kernel thread for each node X. This
|
D | vmstat.c | 1725 unsigned long off = l - (unsigned long *)m->private; in vmstat_show() local 1727 seq_puts(m, vmstat_text[off]); in vmstat_show()
|