/mm/ |
D | swapfile.c | 108 __try_to_reclaim_swap(struct swap_info_struct *si, unsigned long offset) in __try_to_reclaim_swap() argument 110 swp_entry_t entry = swp_entry(si->type, offset); in __try_to_reclaim_swap() 182 pgoff_t offset = start_page - se->start_page; in discard_swap_cluster() local 183 sector_t start_block = se->start_block + offset; in discard_swap_cluster() 184 sector_t nr_blocks = se->nr_pages - offset; in discard_swap_cluster() 281 unsigned long offset) in lock_cluster() argument 287 ci += offset / SWAPFILE_CLUSTER; in lock_cluster() 301 unsigned long offset) in lock_cluster_or_swap_info() argument 305 ci = lock_cluster(si, offset); in lock_cluster_or_swap_info() 514 unsigned long offset) in scan_swap_map_ssd_cluster_conflict() argument [all …]
|
D | readahead.c | 151 pgoff_t offset, unsigned long nr_to_read, in __do_page_cache_readahead() argument 172 pgoff_t page_offset = offset + page_idx; in __do_page_cache_readahead() 210 pgoff_t offset, unsigned long nr_to_read) in force_page_cache_readahead() argument 233 offset, this_chunk, 0); in force_page_cache_readahead() 237 offset += this_chunk; in force_page_cache_readahead() 327 pgoff_t offset, unsigned long max) in count_history_pages() argument 332 head = page_cache_prev_hole(mapping, offset - 1, max); in count_history_pages() 335 return offset - 1 - head; in count_history_pages() 343 pgoff_t offset, in try_context_readahead() argument 349 size = count_history_pages(mapping, offset, max); in try_context_readahead() [all …]
|
D | frontswap.c | 217 pgoff_t offset) in __frontswap_test() argument 220 return test_bit(offset, sis->frontswap_map); in __frontswap_test() 226 pgoff_t offset) in __frontswap_set() argument 228 set_bit(offset, sis->frontswap_map); in __frontswap_set() 233 pgoff_t offset) in __frontswap_clear() argument 235 clear_bit(offset, sis->frontswap_map); in __frontswap_clear() 252 pgoff_t offset = swp_offset(entry); in __frontswap_store() local 265 if (__frontswap_test(sis, offset)) { in __frontswap_store() 266 __frontswap_clear(sis, offset); in __frontswap_store() 268 ops->invalidate_page(type, offset); in __frontswap_store() [all …]
|
D | dmapool.c | 61 unsigned int offset; member 208 unsigned int offset = 0; in pool_initialise_page() local 212 unsigned int next = offset + pool->size; in pool_initialise_page() 217 *(int *)(page->vaddr + offset) = next; in pool_initialise_page() 218 offset = next; in pool_initialise_page() 219 } while (offset < pool->allocation); in pool_initialise_page() 237 page->offset = 0; in pool_alloc_page() 325 size_t offset; in dma_pool_alloc() local 332 if (page->offset < pool->allocation) in dma_pool_alloc() 348 offset = page->offset; in dma_pool_alloc() [all …]
|
D | swap_cgroup.c | 66 pgoff_t offset) in __lookup_swap_cgroup() argument 71 mappage = ctrl->map[offset / SC_PER_PAGE]; in __lookup_swap_cgroup() 73 return sc + offset % SC_PER_PAGE; in __lookup_swap_cgroup() 79 pgoff_t offset = swp_offset(ent); in lookup_swap_cgroup() local 85 return __lookup_swap_cgroup(ctrl, offset); in lookup_swap_cgroup() 133 pgoff_t offset = swp_offset(ent); in swap_cgroup_record() local 134 pgoff_t end = offset + nr_ents; in swap_cgroup_record() 143 offset++; in swap_cgroup_record() 144 if (offset == end) in swap_cgroup_record() 146 if (offset % SC_PER_PAGE) in swap_cgroup_record() [all …]
|
D | zswap.c | 154 pgoff_t offset; member 265 static struct zswap_entry *zswap_rb_search(struct rb_root *root, pgoff_t offset) in zswap_rb_search() argument 272 if (entry->offset > offset) in zswap_rb_search() 274 else if (entry->offset < offset) in zswap_rb_search() 295 if (myentry->offset > entry->offset) in zswap_rb_insert() 297 else if (myentry->offset < entry->offset) in zswap_rb_insert() 353 pgoff_t offset) in zswap_entry_find_get() argument 357 entry = zswap_rb_search(root, offset); in zswap_entry_find_get() 839 pgoff_t offset; in zswap_writeback_entry() local 855 offset = swp_offset(swpentry); in zswap_writeback_entry() [all …]
|
D | swap_state.c | 477 unsigned long offset, in __swapin_nr_pages() argument 496 if (offset != prev_offset + 1 && offset != prev_offset - 1) in __swapin_nr_pages() 516 static unsigned long swapin_nr_pages(unsigned long offset) in swapin_nr_pages() argument 527 pages = __swapin_nr_pages(prev_offset, offset, hits, max_pages, in swapin_nr_pages() 530 prev_offset = offset; in swapin_nr_pages() 560 unsigned long offset = entry_offset; in swapin_readahead() local 566 mask = swapin_nr_pages(offset) - 1; in swapin_readahead() 572 start_offset = offset & ~mask; in swapin_readahead() 573 end_offset = offset | mask; in swapin_readahead() 578 for (offset = start_offset; offset <= end_offset ; offset++) { in swapin_readahead() [all …]
|
D | filemap.c | 720 pgoff_t offset = old->index; in replace_page_cache_page() local 725 new->index = offset; in replace_page_cache_page() 753 pgoff_t offset, gfp_t gfp_mask, in __add_to_page_cache_locked() argument 779 page->index = offset; in __add_to_page_cache_locked() 816 pgoff_t offset, gfp_t gfp_mask) in add_to_page_cache_locked() argument 818 return __add_to_page_cache_locked(page, mapping, offset, in add_to_page_cache_locked() 824 pgoff_t offset, gfp_t gfp_mask) in add_to_page_cache_lru() argument 830 ret = __add_to_page_cache_locked(page, mapping, offset, in add_to_page_cache_lru() 1360 struct page *find_get_entry(struct address_space *mapping, pgoff_t offset) in find_get_entry() argument 1368 pagep = radix_tree_lookup_slot(&mapping->page_tree, offset); in find_get_entry() [all …]
|
D | fadvise.c | 29 SYSCALL_DEFINE4(fadvise64_64, int, fd, loff_t, offset, loff_t, len, int, advice) in SYSCALL_DEFINE4() argument 79 endbyte = (u64)offset + (u64)len; in SYSCALL_DEFINE4() 105 start_index = offset >> PAGE_SHIFT; in SYSCALL_DEFINE4() 124 __filemap_fdatawrite_range(mapping, offset, endbyte, in SYSCALL_DEFINE4() 132 start_index = (offset+(PAGE_SIZE-1)) >> PAGE_SHIFT; in SYSCALL_DEFINE4() 194 SYSCALL_DEFINE4(fadvise64, int, fd, loff_t, offset, size_t, len, int, advice) in SYSCALL_DEFINE4() argument 196 return sys_fadvise64_64(fd, offset, len, advice); in SYSCALL_DEFINE4()
|
D | early_ioremap.c | 108 unsigned long offset; in __early_ioremap() local 137 offset = offset_in_page(phys_addr); in __early_ioremap() 162 __func__, (u64)phys_addr, size, slot, offset, slot_virt[slot]); in __early_ioremap() 164 prev_map[slot] = (void __iomem *)(offset + slot_virt[slot]); in __early_ioremap() 171 unsigned long offset; in early_iounmap() local 200 offset = offset_in_page(virt_addr); in early_iounmap() 201 nrpages = PAGE_ALIGN(offset + size) >> PAGE_SHIFT; in early_iounmap()
|
D | internal.h | 57 struct file *filp, pgoff_t offset, unsigned long nr_to_read, 368 static inline struct page *mem_map_offset(struct page *base, int offset) in mem_map_offset() argument 370 if (unlikely(offset >= MAX_ORDER_NR_PAGES)) in mem_map_offset() 371 return nth_page(base, offset); in mem_map_offset() 372 return base + offset; in mem_map_offset() 380 struct page *base, int offset) in mem_map_next() argument 382 if (unlikely((offset & (MAX_ORDER_NR_PAGES - 1)) == 0)) { in mem_map_next() 383 unsigned long pfn = page_to_pfn(base) + offset; in mem_map_next()
|
D | shmem.c | 2026 unsigned long offset; in shmem_get_unmapped_area() local 2080 offset = (pgoff << PAGE_SHIFT) & (HPAGE_PMD_SIZE-1); in shmem_get_unmapped_area() 2081 if (offset && offset + len < 2 * HPAGE_PMD_SIZE) in shmem_get_unmapped_area() 2083 if ((addr & (HPAGE_PMD_SIZE-1)) == offset) in shmem_get_unmapped_area() 2099 inflated_addr += offset - inflated_offset; in shmem_get_unmapped_area() 2100 if (inflated_offset > offset) in shmem_get_unmapped_area() 2267 pgoff_t offset, max_off; in shmem_mfill_atomic_pte() local 2306 offset = linear_page_index(dst_vma, dst_addr); in shmem_mfill_atomic_pte() 2308 if (unlikely(offset >= max_off)) in shmem_mfill_atomic_pte() 2343 if (unlikely(offset >= max_off)) in shmem_mfill_atomic_pte() [all …]
|
D | process_vm_access.c | 35 unsigned offset, in process_vm_rw_pages() argument 43 size_t copy = PAGE_SIZE - offset; in process_vm_rw_pages() 50 copied = copy_page_from_iter(page, offset, copy, iter); in process_vm_rw_pages() 53 copied = copy_page_to_iter(page, offset, copy, iter); in process_vm_rw_pages() 58 offset = 0; in process_vm_rw_pages()
|
D | zsmalloc.c | 509 static inline void set_first_obj_offset(struct page *page, int offset) in set_first_obj_offset() argument 511 page->units = offset; in set_first_obj_offset() 1457 int i, nr_page, offset; in obj_malloc() local 1468 offset = obj * class->size; in obj_malloc() 1469 nr_page = offset >> PAGE_SHIFT; in obj_malloc() 1470 m_offset = offset & ~PAGE_MASK; in obj_malloc() 1703 int offset = 0; in find_alloced_obj() local 1708 offset = get_first_obj_offset(page); in find_alloced_obj() 1709 offset += class->size * index; in find_alloced_obj() 1711 while (offset < PAGE_SIZE) { in find_alloced_obj() [all …]
|
D | userfaultfd.c | 37 pgoff_t offset, max_off; in mcopy_atomic_pte() local 83 offset = linear_page_index(dst_vma, dst_addr); in mcopy_atomic_pte() 86 if (unlikely(offset >= max_off)) in mcopy_atomic_pte() 123 pgoff_t offset, max_off; in mfill_zeropage_pte() local 132 offset = linear_page_index(dst_vma, dst_addr); in mfill_zeropage_pte() 135 if (unlikely(offset >= max_off)) in mfill_zeropage_pte()
|
D | util.c | 345 unsigned long flag, unsigned long offset) in vm_mmap() argument 347 if (unlikely(offset + PAGE_ALIGN(len) < offset)) in vm_mmap() 349 if (unlikely(offset_in_page(offset))) in vm_mmap() 352 return vm_mmap_pgoff(file, addr, len, prot, flag, offset >> PAGE_SHIFT); in vm_mmap()
|
D | vmalloc.c | 1037 unsigned long offset; in vb_free() local 1049 offset = (unsigned long)addr & (VMAP_BLOCK_SIZE - 1); in vb_free() 1050 offset >>= PAGE_SHIFT; in vb_free() 1063 vb->dirty_min = min(vb->dirty_min, offset); in vb_free() 1064 vb->dirty_max = max(vb->dirty_max, offset + (1UL << order)); in vb_free() 2015 unsigned long offset, length; in aligned_vread() local 2017 offset = offset_in_page(addr); in aligned_vread() 2018 length = PAGE_SIZE - offset; in aligned_vread() 2035 memcpy(buf, map + offset, length); in aligned_vread() 2054 unsigned long offset, length; in aligned_vwrite() local [all …]
|
D | slab.c | 418 u32 offset = (obj - page->s_mem); in obj_to_index() local 419 return reciprocal_divide(offset, cache->reciprocal_buffer_size); in obj_to_index() 1550 static void dump_line(char *data, int offset, int limit) in dump_line() argument 1556 pr_err("%03x: ", offset); in dump_line() 1558 if (data[offset + i] != POISON_FREE) { in dump_line() 1559 error = data[offset + i]; in dump_line() 1564 &data[offset], limit, 1); in dump_line() 2646 size_t offset; in cache_grow_begin() local 2685 offset = n->colour_next; in cache_grow_begin() 2686 if (offset >= cachep->colour) in cache_grow_begin() [all …]
|
D | page_io.c | 112 unsigned long offset; in swap_slot_free_notify() local 115 offset = swp_offset(entry); in swap_slot_free_notify() 119 offset); in swap_slot_free_notify()
|
D | cma.c | 420 unsigned long mask, offset; in cma_alloc() local 437 offset = cma_bitmap_aligned_offset(cma, align); in cma_alloc() 448 offset); in cma_alloc()
|
D | slub.c | 267 return freelist_dereference(s, object + s->offset); in get_freepointer() 273 prefetch(freelist_dereference(s, object + s->offset)); in prefetch_freepointer() 284 freepointer_addr = (unsigned long)object + s->offset; in get_freepointer_safe() 291 unsigned long freeptr_addr = (unsigned long)object + s->offset; in set_freepointer() 548 if (s->offset) in get_track() 549 p = object + s->offset + sizeof(void *); in get_track() 686 if (s->offset) in print_trailer() 687 off = s->offset + sizeof(void *); in print_trailer() 816 if (s->offset) in check_pad_bytes() 902 if (!s->offset && val == SLUB_RED_ACTIVE) in check_object() [all …]
|
D | slob.c | 149 slobidx_t offset = next - base; in set_slob() local 153 s[1].units = offset; in set_slob() 155 s[0].units = -offset; in set_slob()
|
D | madvise.c | 578 loff_t offset; in madvise_remove() local 596 offset = (loff_t)(start - vma->vm_start) in madvise_remove() 612 offset, end - start); in madvise_remove()
|
D | page_alloc.c | 4334 int offset; in page_frag_alloc() local 4354 nc->offset = size; in page_frag_alloc() 4357 offset = nc->offset - fragsz; in page_frag_alloc() 4358 if (unlikely(offset < 0)) { in page_frag_alloc() 4373 offset = size - fragsz; in page_frag_alloc() 4377 nc->offset = offset; in page_frag_alloc() 4379 return nc->va + offset; in page_frag_alloc() 4482 static unsigned long nr_free_zone_pages(int offset) in nr_free_zone_pages() argument 4492 for_each_zone_zonelist(zone, z, zonelist, offset) { in nr_free_zone_pages() 6133 unsigned long __maybe_unused offset = 0; in alloc_node_mem_map() local [all …]
|
D | truncate.c | 113 void do_invalidatepage(struct page *page, unsigned int offset, in do_invalidatepage() argument 124 (*invalidatepage)(page, offset, length); in do_invalidatepage()
|