/mm/ |
D | percpu.c | 310 static unsigned long pcpu_off_to_block_index(int off) in pcpu_off_to_block_index() argument 312 return off / PCPU_BITMAP_BLOCK_BITS; in pcpu_off_to_block_index() 315 static unsigned long pcpu_off_to_block_off(int off) in pcpu_off_to_block_off() argument 317 return off & (PCPU_BITMAP_BLOCK_BITS - 1); in pcpu_off_to_block_off() 320 static unsigned long pcpu_block_off_to_off(int index, int off) in pcpu_block_off_to_off() argument 322 return index * PCPU_BITMAP_BLOCK_BITS + off; in pcpu_block_off_to_off() 1243 static void pcpu_free_area(struct pcpu_chunk *chunk, int off) in pcpu_free_area() argument 1253 bit_off = off / PCPU_MIN_ALLOC_SIZE; in pcpu_free_area() 1596 int slot, off, cpu, ret; in pcpu_alloc() local 1639 off = pcpu_find_block_fit(chunk, bits, bit_align, is_atomic); in pcpu_alloc() [all …]
|
D | zsmalloc.c | 986 unsigned long off = 0; in init_zspage() local 994 set_first_obj_offset(page, off); in init_zspage() 997 link = (struct link_free *)vaddr + off / sizeof(*link); in init_zspage() 999 while ((off += class->size) < PAGE_SIZE) { in init_zspage() 1021 off %= PAGE_SIZE; in init_zspage() 1139 struct page *pages[2], int off, int size) in __zs_map_object() argument 1143 return area->vm_addr + off; in __zs_map_object() 1147 struct page *pages[2], int off, int size) in __zs_unmap_object() argument 1177 struct page *pages[2], int off, int size) in __zs_map_object() argument 1190 sizes[0] = PAGE_SIZE - off; in __zs_map_object() [all …]
|
D | hugetlb_cgroup.c | 289 char *buf, size_t nbytes, loff_t off) in hugetlb_cgroup_write() argument 320 char *buf, size_t nbytes, loff_t off) in hugetlb_cgroup_reset() argument
|
D | huge_memory.c | 532 loff_t off, unsigned long flags, unsigned long size) in __thp_get_unmapped_area() argument 534 loff_t off_end = off + len; in __thp_get_unmapped_area() 535 loff_t off_align = round_up(off, size); in __thp_get_unmapped_area() 542 if (len_pad < len || (off + len_pad) < off) in __thp_get_unmapped_area() 546 off >> PAGE_SHIFT, flags); in __thp_get_unmapped_area() 562 ret += (off - ret) & (size - 1); in __thp_get_unmapped_area() 570 loff_t off = (loff_t)pgoff << PAGE_SHIFT; in thp_get_unmapped_area() local 575 ret = __thp_get_unmapped_area(filp, addr, len, off, flags, PMD_SIZE); in thp_get_unmapped_area()
|
D | slub.c | 649 unsigned int off; /* Offset of last byte */ in print_trailer() local 672 off = s->offset + sizeof(void *); in print_trailer() 674 off = s->inuse; in print_trailer() 677 off += 2 * sizeof(struct track); in print_trailer() 679 off += kasan_metadata_size(s); in print_trailer() 681 if (off != size_from_object(s)) in print_trailer() 683 print_section(KERN_ERR, "Padding ", p + off, in print_trailer() 684 size_from_object(s) - off); in print_trailer() 799 unsigned long off = s->inuse; /* The end of info */ in check_pad_bytes() local 803 off += sizeof(void *); in check_pad_bytes() [all …]
|
D | memory.c | 622 unsigned long off; in vm_normal_page() local 623 off = (addr - vma->vm_start) >> PAGE_SHIFT; in vm_normal_page() 624 if (pfn == vma->vm_pgoff + off) in vm_normal_page() 665 unsigned long off; in vm_normal_page_pmd() local 666 off = (addr - vma->vm_start) >> PAGE_SHIFT; in vm_normal_page_pmd() 667 if (pfn == vma->vm_pgoff + off) in vm_normal_page_pmd() 3458 int off; in do_fault_around() local 3465 off = ((address - vmf->address) >> PAGE_SHIFT) & (PTRS_PER_PTE - 1); in do_fault_around() 3466 start_pgoff -= off; in do_fault_around()
|
D | memcontrol.c | 3293 loff_t off) in mem_cgroup_force_empty_write() argument 3613 char *buf, size_t nbytes, loff_t off) in mem_cgroup_write() argument 3657 size_t nbytes, loff_t off) in mem_cgroup_reset() argument 4624 char *buf, size_t nbytes, loff_t off) in memcg_write_event_control() argument 6032 char *buf, size_t nbytes, loff_t off) in memory_min_write() argument 6055 char *buf, size_t nbytes, loff_t off) in memory_low_write() argument 6077 char *buf, size_t nbytes, loff_t off) in memory_high_write() argument 6107 char *buf, size_t nbytes, loff_t off) in memory_max_write() argument 6204 char *buf, size_t nbytes, loff_t off) in memory_oom_group_write() argument 7190 char *buf, size_t nbytes, loff_t off) in swap_max_write() argument
|
D | mempolicy.c | 1910 unsigned long off; in interleave_nid() local 1920 off = vma->vm_pgoff >> (shift - PAGE_SHIFT); in interleave_nid() 1921 off += (addr - vma->vm_start) >> shift; in interleave_nid() 1922 return offset_il_node(pol, off); in interleave_nid()
|
D | shmem.c | 1816 pgoff_t off; in shmem_getpage_gfp() local 1820 off = round_up(index, HPAGE_PMD_NR); in shmem_getpage_gfp() 1823 i_size >> PAGE_SHIFT >= off) in shmem_getpage_gfp() 3993 pgoff_t off; in shmem_huge_enabled() local 4008 off = round_up(vma->vm_pgoff, HPAGE_PMD_NR); in shmem_huge_enabled() 4011 i_size >> PAGE_SHIFT >= off) in shmem_huge_enabled()
|
D | Kconfig.debug | 46 can be overridden by debug_pagealloc=off|on.
|
D | Kconfig | 350 more than it requires. To deal with this, mmap() is able to trim off 353 If trimming is enabled, the excess is trimmed off and returned to the 629 by starting one-off "pgdatinitX" kernel thread for each node X. This
|
D | vmstat.c | 1729 unsigned long off = l - (unsigned long *)m->private; in vmstat_show() local 1731 seq_puts(m, vmstat_text[off]); in vmstat_show()
|