/mm/ |
D | percpu.c | 335 int off = chunk->map[i] & ~1; in pcpu_count_occupied_pages() local 338 if (!PAGE_ALIGNED(off) && i > 0) { in pcpu_count_occupied_pages() 341 if (!(prev & 1) && prev <= round_down(off, PAGE_SIZE)) in pcpu_count_occupied_pages() 342 off = round_down(off, PAGE_SIZE); in pcpu_count_occupied_pages() 353 return max_t(int, PFN_DOWN(end) - PFN_UP(off), 0); in pcpu_count_occupied_pages() 502 static int pcpu_fit_in_area(struct pcpu_chunk *chunk, int off, int this_size, in pcpu_fit_in_area() argument 505 int cand_off = off; in pcpu_fit_in_area() 508 int head = ALIGN(cand_off, align) - off; in pcpu_fit_in_area() 522 page_start = PFN_DOWN(head + off); in pcpu_fit_in_area() 523 page_end = PFN_UP(head + off + size); in pcpu_fit_in_area() [all …]
|
D | zsmalloc.c | 849 unsigned long off = 0; in obj_idx_to_offset() local 852 off = page->index; in obj_idx_to_offset() 854 return off + obj_idx * class_size; in obj_idx_to_offset() 914 unsigned long off = 0; in init_zspage() local 931 page->index = off; in init_zspage() 934 link = (struct link_free *)vaddr + off / sizeof(*link); in init_zspage() 936 while ((off += class->size) < PAGE_SIZE) { in init_zspage() 950 off %= PAGE_SIZE; in init_zspage() 1053 struct page *pages[2], int off, int size) in __zs_map_object() argument 1057 return area->vm_addr + off; in __zs_map_object() [all …]
|
D | hugetlb_cgroup.c | 273 char *buf, size_t nbytes, loff_t off) in hugetlb_cgroup_write() argument 303 char *buf, size_t nbytes, loff_t off) in hugetlb_cgroup_reset() argument
|
D | slub.c | 650 unsigned int off; /* Offset of last byte */ in print_trailer() local 672 off = s->offset + sizeof(void *); in print_trailer() 674 off = s->inuse; in print_trailer() 677 off += 2 * sizeof(struct track); in print_trailer() 679 off += kasan_metadata_size(s); in print_trailer() 681 if (off != size_from_object(s)) in print_trailer() 683 print_section("Padding ", p + off, size_from_object(s) - off); in print_trailer() 798 unsigned long off = s->inuse; /* The end of info */ in check_pad_bytes() local 802 off += sizeof(void *); in check_pad_bytes() 806 off += 2 * sizeof(struct track); in check_pad_bytes() [all …]
|
D | memory.c | 766 unsigned long off; in vm_normal_page() local 767 off = (addr - vma->vm_start) >> PAGE_SHIFT; in vm_normal_page() 768 if (pfn == vma->vm_pgoff + off) in vm_normal_page() 808 unsigned long off; in vm_normal_page_pmd() local 809 off = (addr - vma->vm_start) >> PAGE_SHIFT; in vm_normal_page_pmd() 810 if (pfn == vma->vm_pgoff + off) in vm_normal_page_pmd() 2952 int off; in do_fault_around() local 2958 off = ((address - start_addr) >> PAGE_SHIFT) & (PTRS_PER_PTE - 1); in do_fault_around() 2959 pte -= off; in do_fault_around() 2960 pgoff -= off; in do_fault_around()
|
D | mempolicy.c | 1752 struct vm_area_struct *vma, unsigned long off) in offset_il_node() argument 1761 target = (unsigned int)off % nnodes; in offset_il_node() 1775 unsigned long off; in interleave_nid() local 1785 off = vma->vm_pgoff >> (shift - PAGE_SHIFT); in interleave_nid() 1786 off += (addr - vma->vm_start) >> shift; in interleave_nid() 1787 return offset_il_node(pol, vma, off); in interleave_nid()
|
D | bootmem.c | 491 unsigned long off, unsigned long align) in align_off() argument 497 return ALIGN(base + off, align) - base; in align_off()
|
D | memcontrol.c | 2782 loff_t off) in mem_cgroup_force_empty_write() argument 3013 char *buf, size_t nbytes, loff_t off) in mem_cgroup_write() argument 3051 size_t nbytes, loff_t off) in mem_cgroup_reset() argument 3907 char *buf, size_t nbytes, loff_t off) in memcg_write_event_control() argument 5193 char *buf, size_t nbytes, loff_t off) in memory_low_write() argument 5223 char *buf, size_t nbytes, loff_t off) in memory_high_write() argument 5260 char *buf, size_t nbytes, loff_t off) in memory_max_write() argument
|
D | vmstat.c | 1354 unsigned long off = l - (unsigned long *)m->private; in vmstat_show() local 1356 seq_puts(m, vmstat_text[off]); in vmstat_show()
|
D | Kconfig | 373 more than it requires. To deal with this, mmap() is able to trim off 376 If trimming is enabled, the excess is trimmed off and returned to the
|