/mm/ |
D | swapfile.c | 129 unsigned long offset, unsigned long flags) in __try_to_reclaim_swap() argument 131 swp_entry_t entry = swp_entry(si->type, offset); in __try_to_reclaim_swap() 135 page = find_get_page(swap_address_space(entry), offset); in __try_to_reclaim_swap() 206 offset_to_swap_extent(struct swap_info_struct *sis, unsigned long offset) in offset_to_swap_extent() argument 214 if (offset < se->start_page) in offset_to_swap_extent() 216 else if (offset >= se->start_page + se->nr_pages) in offset_to_swap_extent() 230 pgoff_t offset; in swap_page_sector() local 232 offset = __page_file_index(page); in swap_page_sector() 233 se = offset_to_swap_extent(sis, offset); in swap_page_sector() 234 sector = se->start_block + (offset - se->start_page); in swap_page_sector() [all …]
|
D | frontswap.c | 216 pgoff_t offset) in __frontswap_test() argument 219 return test_bit(offset, sis->frontswap_map); in __frontswap_test() 225 pgoff_t offset) in __frontswap_set() argument 227 set_bit(offset, sis->frontswap_map); in __frontswap_set() 232 pgoff_t offset) in __frontswap_clear() argument 234 clear_bit(offset, sis->frontswap_map); in __frontswap_clear() 251 pgoff_t offset = swp_offset(entry); in __frontswap_store() local 264 if (__frontswap_test(sis, offset)) { in __frontswap_store() 265 __frontswap_clear(sis, offset); in __frontswap_store() 267 ops->invalidate_page(type, offset); in __frontswap_store() [all …]
|
D | fadvise.c | 32 int generic_fadvise(struct file *file, loff_t offset, loff_t len, int advice) in generic_fadvise() argument 73 endbyte = (u64)offset + (u64)len; in generic_fadvise() 99 start_index = offset >> PAGE_SHIFT; in generic_fadvise() 113 __filemap_fdatawrite_range(mapping, offset, endbyte, in generic_fadvise() 121 start_index = (offset+(PAGE_SIZE-1)) >> PAGE_SHIFT; in generic_fadvise() 181 int vfs_fadvise(struct file *file, loff_t offset, loff_t len, int advice) in vfs_fadvise() argument 184 return file->f_op->fadvise(file, offset, len, advice); in vfs_fadvise() 186 return generic_fadvise(file, offset, len, advice); in vfs_fadvise() 192 int ksys_fadvise64_64(int fd, loff_t offset, loff_t len, int advice) in ksys_fadvise64_64() argument 200 ret = vfs_fadvise(f.file, offset, len, advice); in ksys_fadvise64_64() [all …]
|
D | dmapool.c | 58 unsigned int offset; member 205 unsigned int offset = 0; in pool_initialise_page() local 209 unsigned int next = offset + pool->size; in pool_initialise_page() 214 *(int *)(page->vaddr + offset) = next; in pool_initialise_page() 215 offset = next; in pool_initialise_page() 216 } while (offset < pool->allocation); in pool_initialise_page() 234 page->offset = 0; in pool_alloc_page() 319 size_t offset; in dma_pool_alloc() local 326 if (page->offset < pool->allocation) in dma_pool_alloc() 342 offset = page->offset; in dma_pool_alloc() [all …]
|
D | swap_cgroup.c | 66 pgoff_t offset) in __lookup_swap_cgroup() argument 71 mappage = ctrl->map[offset / SC_PER_PAGE]; in __lookup_swap_cgroup() 73 return sc + offset % SC_PER_PAGE; in __lookup_swap_cgroup() 79 pgoff_t offset = swp_offset(ent); in lookup_swap_cgroup() local 85 return __lookup_swap_cgroup(ctrl, offset); in lookup_swap_cgroup() 133 pgoff_t offset = swp_offset(ent); in swap_cgroup_record() local 134 pgoff_t end = offset + nr_ents; in swap_cgroup_record() 143 offset++; in swap_cgroup_record() 144 if (offset == end) in swap_cgroup_record() 146 if (offset % SC_PER_PAGE) in swap_cgroup_record() [all …]
|
D | zswap.c | 163 pgoff_t offset; member 284 static struct zswap_entry *zswap_rb_search(struct rb_root *root, pgoff_t offset) in zswap_rb_search() argument 291 if (entry->offset > offset) in zswap_rb_search() 293 else if (entry->offset < offset) in zswap_rb_search() 314 if (myentry->offset > entry->offset) in zswap_rb_insert() 316 else if (myentry->offset < entry->offset) in zswap_rb_insert() 376 pgoff_t offset) in zswap_entry_find_get() argument 380 entry = zswap_rb_search(root, offset); in zswap_entry_find_get() 876 pgoff_t offset; in zswap_writeback_entry() local 891 offset = swp_offset(swpentry); in zswap_writeback_entry() [all …]
|
D | page_reporting.c | 111 struct scatterlist *sgl, unsigned int *offset) in page_reporting_cycle() argument 163 if (*offset) { in page_reporting_cycle() 170 --(*offset); in page_reporting_cycle() 171 sg_set_page(&sgl[*offset], page, page_len, 0); in page_reporting_cycle() 191 *offset = PAGE_REPORTING_CAPACITY; in page_reporting_cycle() 226 unsigned int order, mt, leftover, offset = PAGE_REPORTING_CAPACITY; in page_reporting_process_zone() local 249 sgl, &offset); in page_reporting_process_zone() 256 leftover = PAGE_REPORTING_CAPACITY - offset; in page_reporting_process_zone() 258 sgl = &sgl[offset]; in page_reporting_process_zone()
|
D | swap_state.c | 572 unsigned long offset, in __swapin_nr_pages() argument 591 if (offset != prev_offset + 1 && offset != prev_offset - 1) in __swapin_nr_pages() 611 static unsigned long swapin_nr_pages(unsigned long offset) in swapin_nr_pages() argument 622 pages = __swapin_nr_pages(READ_ONCE(prev_offset), offset, hits, in swapin_nr_pages() 626 WRITE_ONCE(prev_offset, offset); in swapin_nr_pages() 659 unsigned long offset = entry_offset; in swap_cluster_readahead() local 668 mask = swapin_nr_pages(offset) - 1; in swap_cluster_readahead() 681 start_offset = offset & ~mask; in swap_cluster_readahead() 682 end_offset = offset | mask; in swap_cluster_readahead() 689 for (offset = start_offset; offset <= end_offset ; offset++) { in swap_cluster_readahead() [all …]
|
D | early_ioremap.c | 108 unsigned long offset; in __early_ioremap() local 137 offset = offset_in_page(phys_addr); in __early_ioremap() 162 __func__, &phys_addr, size, slot, offset, slot_virt[slot]); in __early_ioremap() 164 prev_map[slot] = (void __iomem *)(offset + slot_virt[slot]); in __early_ioremap() 171 unsigned long offset; in early_iounmap() local 200 offset = offset_in_page(virt_addr); in early_iounmap() 201 nrpages = PAGE_ALIGN(offset + size) >> PAGE_SHIFT; in early_iounmap()
|
D | filemap.c | 799 pgoff_t offset = old->index; in replace_page_cache_page() local 800 XA_STATE(xas, &mapping->i_pages, offset); in replace_page_cache_page() 809 new->index = offset; in replace_page_cache_page() 837 pgoff_t offset, gfp_t gfp, in __add_to_page_cache_locked() argument 840 XA_STATE(xas, &mapping->i_pages, offset); in __add_to_page_cache_locked() 851 page->index = offset; in __add_to_page_cache_locked() 934 pgoff_t offset, gfp_t gfp_mask) in add_to_page_cache_locked() argument 936 return __add_to_page_cache_locked(page, mapping, offset, in add_to_page_cache_locked() 942 pgoff_t offset, gfp_t gfp_mask) in add_to_page_cache_lru() argument 948 ret = __add_to_page_cache_locked(page, mapping, offset, in add_to_page_cache_lru() [all …]
|
D | shmem.c | 2155 unsigned long offset; in shmem_get_unmapped_area() local 2209 offset = (pgoff << PAGE_SHIFT) & (HPAGE_PMD_SIZE-1); in shmem_get_unmapped_area() 2210 if (offset && offset + len < 2 * HPAGE_PMD_SIZE) in shmem_get_unmapped_area() 2212 if ((addr & (HPAGE_PMD_SIZE-1)) == offset) in shmem_get_unmapped_area() 2228 inflated_addr += offset - inflated_offset; in shmem_get_unmapped_area() 2229 if (inflated_offset > offset) in shmem_get_unmapped_area() 2547 unsigned long offset; in shmem_file_read_iter() local 2562 offset = *ppos & ~PAGE_MASK; in shmem_file_read_iter() 2575 if (nr <= offset) in shmem_file_read_iter() 2600 if (nr <= offset) { in shmem_file_read_iter() [all …]
|
D | internal.h | 483 static inline struct page *mem_map_offset(struct page *base, int offset) in mem_map_offset() argument 485 if (unlikely(offset >= MAX_ORDER_NR_PAGES)) in mem_map_offset() 486 return nth_page(base, offset); in mem_map_offset() 487 return base + offset; in mem_map_offset() 495 struct page *base, int offset) in mem_map_next() argument 497 if (unlikely((offset & (MAX_ORDER_NR_PAGES - 1)) == 0)) { in mem_map_next() 498 unsigned long pfn = page_to_pfn(base) + offset; in mem_map_next()
|
D | process_vm_access.c | 29 unsigned offset, in process_vm_rw_pages() argument 37 size_t copy = PAGE_SIZE - offset; in process_vm_rw_pages() 44 copied = copy_page_from_iter(page, offset, copy, iter); in process_vm_rw_pages() 46 copied = copy_page_to_iter(page, offset, copy, iter); in process_vm_rw_pages() 51 offset = 0; in process_vm_rw_pages()
|
D | ioremap.c | 255 unsigned long offset, vaddr; in ioremap_prot() local 265 offset = addr & (~PAGE_MASK); in ioremap_prot() 266 addr -= offset; in ioremap_prot() 267 size = PAGE_ALIGN(size + offset); in ioremap_prot() 280 return (void __iomem *)(vaddr + offset); in ioremap_prot()
|
D | usercopy.c | 73 unsigned long offset, unsigned long len) in usercopy_warn() argument 80 offset, len); in usercopy_warn() 84 bool to_user, unsigned long offset, in usercopy_abort() argument 92 offset, len); in usercopy_abort()
|
D | zsmalloc.c | 495 static inline void set_first_obj_offset(struct page *page, int offset) in set_first_obj_offset() argument 497 page->units = offset; in set_first_obj_offset() 1364 int i, nr_page, offset; in obj_malloc() local 1375 offset = obj * class->size; in obj_malloc() 1376 nr_page = offset >> PAGE_SHIFT; in obj_malloc() 1377 m_offset = offset & ~PAGE_MASK; in obj_malloc() 1610 int offset = 0; in find_alloced_obj() local 1615 offset = get_first_obj_offset(page); in find_alloced_obj() 1616 offset += class->size * index; in find_alloced_obj() 1618 while (offset < PAGE_SIZE) { in find_alloced_obj() [all …]
|
D | userfaultfd.c | 69 pgoff_t offset, max_off; in mfill_atomic_install_pte() local 88 offset = linear_page_index(dst_vma, dst_addr); in mfill_atomic_install_pte() 91 if (unlikely(offset >= max_off)) in mfill_atomic_install_pte() 191 pgoff_t offset, max_off; in mfill_zeropage_pte() local 200 offset = linear_page_index(dst_vma, dst_addr); in mfill_zeropage_pte() 203 if (unlikely(offset >= max_off)) in mfill_zeropage_pte()
|
D | slab.c | 1456 static void dump_line(char *data, int offset, int limit) in dump_line() argument 1462 pr_err("%03x: ", offset); in dump_line() 1464 if (data[offset + i] != POISON_FREE) { in dump_line() 1465 error = data[offset + i]; in dump_line() 1470 &data[offset], limit, 1); in dump_line() 2568 size_t offset; in cache_grow_begin() local 2604 offset = n->colour_next; in cache_grow_begin() 2605 if (offset >= cachep->colour) in cache_grow_begin() 2606 offset = 0; in cache_grow_begin() 2608 offset *= cachep->colour_off; in cache_grow_begin() [all …]
|
D | readahead.c | 623 ssize_t ksys_readahead(int fd, loff_t offset, size_t count) in ksys_readahead() argument 644 ret = vfs_fadvise(f.file, offset, count, POSIX_FADV_WILLNEED); in ksys_readahead() 650 SYSCALL_DEFINE3(readahead, int, fd, loff_t, offset, size_t, count) in SYSCALL_DEFINE3() argument 652 return ksys_readahead(fd, offset, count); in SYSCALL_DEFINE3()
|
D | util.c | 557 unsigned long flag, unsigned long offset) in vm_mmap() argument 559 if (unlikely(offset + PAGE_ALIGN(len) < offset)) in vm_mmap() 561 if (unlikely(offset_in_page(offset))) in vm_mmap() 564 return vm_mmap_pgoff(file, addr, len, prot, flag, offset >> PAGE_SHIFT); in vm_mmap()
|
D | madvise.c | 268 loff_t offset; in madvise_willneed() local 301 offset = (loff_t)(start - vma->vm_start) in madvise_willneed() 304 vfs_fadvise(file, offset, end - start, POSIX_FADV_WILLNEED); in madvise_willneed() 865 loff_t offset; in madvise_remove() local 884 offset = (loff_t)(start - vma->vm_start) in madvise_remove() 900 offset, end - start); in madvise_remove()
|
D | vmalloc.c | 1707 unsigned long offset; in vb_free() local 1717 offset = (addr & (VMAP_BLOCK_SIZE - 1)) >> PAGE_SHIFT; in vb_free() 1728 vb->dirty_min = min(vb->dirty_min, offset); in vb_free() 1729 vb->dirty_max = max(vb->dirty_max, offset + (1UL << order)); in vb_free() 2788 unsigned long offset, length; in aligned_vread() local 2790 offset = offset_in_page(addr); in aligned_vread() 2791 length = PAGE_SIZE - offset; in aligned_vread() 2808 memcpy(buf, map + offset, length); in aligned_vread() 2827 unsigned long offset, length; in aligned_vwrite() local 2829 offset = offset_in_page(addr); in aligned_vwrite() [all …]
|
D | slub.c | 273 return freelist_dereference(s, object + s->offset); in get_freepointer() 278 prefetch(object + s->offset); in prefetch_freepointer() 290 freepointer_addr = (unsigned long)object + s->offset; in get_freepointer_safe() 297 unsigned long freeptr_addr = (unsigned long)object + s->offset; in set_freepointer() 554 return s->offset >= s->inuse; in freeptr_outside_object() 2853 memset((void *)((char *)kasan_reset_tag(obj) + s->offset), in maybe_wipe_obj_freeptr() 3777 s->offset = size; in calculate_sizes() 3785 s->offset = ALIGN_DOWN(s->object_size / 2, sizeof(void *)); in calculate_sizes() 3871 s->offset = 0; in kmem_cache_open() 4110 unsigned int offset; in __check_heap_object() local [all …]
|
D | cma.c | 440 unsigned long mask, offset; in cma_alloc() local 466 offset = cma_bitmap_aligned_offset(cma, align); in cma_alloc() 480 offset); in cma_alloc()
|
/mm/kasan/ |
D | report_generic.c | 187 unsigned long offset; in print_decoded_frame_descr() local 192 &offset)) in print_decoded_frame_descr() 210 pr_err(" [%lu, %lu) '%s'", offset, offset + size, token); in print_decoded_frame_descr() 215 unsigned long *offset, in get_address_stack_frame_info() argument 259 *offset = (unsigned long)addr - (unsigned long)frame; in get_address_stack_frame_info() 268 unsigned long offset; in kasan_print_address_stack_frame() local 272 if (!get_address_stack_frame_info(addr, &offset, &frame_descr, in kasan_print_address_stack_frame() 282 addr, current->comm, task_pid_nr(current), offset); in kasan_print_address_stack_frame()
|