/mm/ |
D | readahead.c | 99 if (add_to_page_cache_lru(page, mapping, page->index, in read_cache_pages() 188 unsigned long index = readahead_index(ractl); in page_cache_ra_unbounded() local 210 struct page *page = xa_load(&mapping->i_pages, index + i); in page_cache_ra_unbounded() 222 i = ractl->_index + ractl->_nr_pages - index - 1; in page_cache_ra_unbounded() 230 page->index = index + i; in page_cache_ra_unbounded() 232 } else if (add_to_page_cache_lru(page, mapping, index + i, in page_cache_ra_unbounded() 236 i = ractl->_index + ractl->_nr_pages - index - 1; in page_cache_ra_unbounded() 265 unsigned long index = readahead_index(ractl); in do_page_cache_ra() local 273 if (index > end_index) in do_page_cache_ra() 276 if (nr_to_read > end_index - index) in do_page_cache_ra() [all …]
|
D | truncate.c | 35 pgoff_t index, void *entry) in __clear_shadow_entry() argument 37 XA_STATE(xas, &mapping->i_pages, index); in __clear_shadow_entry() 45 static void clear_shadow_entry(struct address_space *mapping, pgoff_t index, in clear_shadow_entry() argument 49 __clear_shadow_entry(mapping, index, entry); in clear_shadow_entry() 81 pgoff_t index = indices[i]; in truncate_exceptional_pvec_entries() local 89 dax_delete_mapping_entry(mapping, index); in truncate_exceptional_pvec_entries() 93 __clear_shadow_entry(mapping, index, page); in truncate_exceptional_pvec_entries() 106 pgoff_t index, void *entry) in invalidate_exceptional_entry() argument 111 clear_shadow_entry(mapping, index, entry); in invalidate_exceptional_entry() 120 pgoff_t index, void *entry) in invalidate_exceptional_entry2() argument [all …]
|
D | shmem.c | 140 static int shmem_swapin_page(struct inode *inode, pgoff_t index, 144 static int shmem_getpage_gfp(struct inode *inode, pgoff_t index, 149 int shmem_getpage(struct inode *inode, pgoff_t index, in shmem_getpage() argument 152 return shmem_getpage_gfp(inode, index, pagep, sgp, in shmem_getpage() 413 pgoff_t index, void *expected, void *replacement) in shmem_replace_entry() argument 415 XA_STATE(xas, &mapping->i_pages, index); in shmem_replace_entry() 435 pgoff_t index, swp_entry_t swap) in shmem_confirm_swap() argument 437 return xa_load(&mapping->i_pages, index) == swp_to_radix_entry(swap); in shmem_confirm_swap() 478 struct inode *inode, pgoff_t index) in shmem_is_huge() argument 494 index = round_up(index + 1, HPAGE_PMD_NR); in shmem_is_huge() [all …]
|
D | filemap.c | 130 XA_STATE(xas, &mapping->i_pages, page->index); in page_cache_delete() 137 xas_set_order(&xas, page->index, compound_order(page)); in page_cache_delete() 293 XA_STATE(xas, &mapping->i_pages, pvec->pages[0]->index); in page_cache_delete_batch() 314 VM_BUG_ON_PAGE(page->index > pvec->pages[i]->index, in page_cache_delete_batch() 321 if (page->index == xas.xa_index) in page_cache_delete_batch() 330 if (page->index + compound_nr(page) - 1 == xas.xa_index) in page_cache_delete_batch() 519 pgoff_t index = start_byte >> PAGE_SHIFT; in __filemap_fdatawait_range() local 528 while (index <= end) { in __filemap_fdatawait_range() 531 nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, in __filemap_fdatawait_range() 843 pgoff_t offset = old->index; in replace_page_cache_page() [all …]
|
D | khugepaged.c | 1686 pgoff_t index, end = start + HPAGE_PMD_NR; in collapse_file() local 1727 new_page->index = start; in collapse_file() 1737 for (index = start; index < end; index++) { in collapse_file() 1740 VM_BUG_ON(index != xas.xa_index); in collapse_file() 1748 if (index == start) { in collapse_file() 1753 xas_set(&xas, index); in collapse_file() 1767 if (shmem_getpage(mapping->host, index, &page, in collapse_file() 1783 file, index, in collapse_file() 1784 end - index); in collapse_file() 1787 page = find_lock_page(mapping, index); in collapse_file() [all …]
|
D | page_pinner.c | 50 unsigned long index; member 99 pp_buffer.index = 0; in init_page_pinner() 147 idx = pp_buf->index++; in add_record() 148 pp_buf->index %= pp_buf_size; in add_record() 344 tmp = pp_buffer.index - 1 - i + pp_buf_size; in read_buffer() 393 pp_buffer.index = 0; in buffer_size_set()
|
D | page_ext.c | 127 static inline struct page_ext *get_entry(void *base, unsigned long index) in get_entry() argument 129 return base + page_ext_size * index; in get_entry() 185 unsigned long index; in lookup_page_ext() local 198 index = pfn - round_down(node_start_pfn(page_to_nid(page)), in lookup_page_ext() 200 return get_entry(base, index); in lookup_page_ext()
|
D | zsmalloc.c | 212 unsigned int index; member 606 if (class->index != i) in zs_stats_size_show() 872 return page->index; in obj_to_head() 1387 zspage->first_page->index = handle; in obj_malloc() 1451 set_zspage_mapping(zspage, class->index, newfg); in zs_malloc() 1606 int index = *obj_idx; in find_alloced_obj() local 1611 offset += class->size * index; in find_alloced_obj() 1623 index++; in find_alloced_obj() 1628 *obj_idx = index; in find_alloced_obj() 1734 set_zspage_mapping(zspage, class->index, fullness); in putback_zspage() [all …]
|
D | cleancache.c | 195 ret = cleancache_ops->get_page(pool_id, key, page->index, page); in __cleancache_get_page() 229 cleancache_ops->put_page(pool_id, key, page->index, page); in __cleancache_put_page() 257 key, page->index); in __cleancache_invalidate_page()
|
D | percpu.c | 260 page->index = (unsigned long)pcpu; in pcpu_set_page_chunk() 266 return (struct pcpu_chunk *)page->index; in pcpu_get_page_chunk() 290 static unsigned long *pcpu_index_alloc_map(struct pcpu_chunk *chunk, int index) in pcpu_index_alloc_map() argument 293 (index * PCPU_BITMAP_BLOCK_BITS / BITS_PER_LONG); in pcpu_index_alloc_map() 306 static unsigned long pcpu_block_off_to_off(int index, int off) in pcpu_block_off_to_off() argument 308 return index * PCPU_BITMAP_BLOCK_BITS + off; in pcpu_block_off_to_off() 778 static void pcpu_block_refresh_hint(struct pcpu_chunk *chunk, int index) in pcpu_block_refresh_hint() argument 780 struct pcpu_block_md *block = chunk->md_blocks + index; in pcpu_block_refresh_hint() 781 unsigned long *alloc_map = pcpu_index_alloc_map(chunk, index); in pcpu_block_refresh_hint() 1165 unsigned long index, end, i, area_off, area_bits; in pcpu_find_zero_area() local [all …]
|
D | vmstat.c | 2146 int index; in unusable_show_print() local 2154 index = unusable_free_index(order, &info); in unusable_show_print() 2155 seq_printf(m, "%d.%03d ", index / 1000, index % 1000); in unusable_show_print() 2196 int index; in extfrag_show_print() local 2206 index = __fragmentation_index(order, &info); in extfrag_show_print() 2207 seq_printf(m, "%d.%03d ", index / 1000, index % 1000); in extfrag_show_print()
|
D | page-writeback.c | 2186 pgoff_t index; in write_cache_pages() local 2194 index = mapping->writeback_index; /* prev offset */ in write_cache_pages() 2197 index = wbc->range_start >> PAGE_SHIFT; in write_cache_pages() 2203 tag_pages_for_writeback(mapping, index, end); in write_cache_pages() 2208 done_index = index; in write_cache_pages() 2209 while (!done && (index <= end)) { in write_cache_pages() 2212 nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end, in write_cache_pages() 2220 done_index = page->index; in write_cache_pages() 2274 done_index = page->index + 1; in write_cache_pages()
|
D | madvise.c | 201 unsigned long index; in swapin_walk_pmd_entry() local 206 for (index = start; index != end; index += PAGE_SIZE) { in swapin_walk_pmd_entry() 213 pte = *(orig_pte + ((index - start) / PAGE_SIZE)); in swapin_walk_pmd_entry() 223 vma, index, false); in swapin_walk_pmd_entry()
|
D | mincore.c | 51 static unsigned char mincore_page(struct address_space *mapping, pgoff_t index) in mincore_page() argument 62 page = find_get_incore_page(mapping, index); in mincore_page()
|
D | slab_common.c | 746 unsigned int index; in kmalloc_slab() local 752 index = size_index[size_index_elem(size)]; in kmalloc_slab() 756 index = fls(size - 1); in kmalloc_slab() 759 return kmalloc_caches[kmalloc_type(flags)][index]; in kmalloc_slab()
|
D | swap_state.c | 393 struct page *find_get_incore_page(struct address_space *mapping, pgoff_t index) in find_get_incore_page() argument 397 struct page *page = pagecache_get_page(mapping, index, in find_get_incore_page() 403 return find_subpage(page, index); in find_get_incore_page()
|
D | internal.h | 57 struct file *file, pgoff_t index, unsigned long nr_to_read) in force_page_cache_readahead() argument 59 DEFINE_READAHEAD(ractl, file, &file->f_ra, mapping, index); in force_page_cache_readahead()
|
D | huge_memory.c | 2376 page_tail->index = head->index + tail; in __split_huge_page_tail() 2437 if (head[i].index >= end) { in __split_huge_page() 2444 __xa_store(&head->mapping->i_pages, head[i].index, in __split_huge_page() 3024 pgoff_t index; in split_huge_pages_in_file() local 3041 for (index = off_start; index < off_end; index += nr_pages) { in split_huge_pages_in_file() 3042 struct page *fpage = pagecache_get_page(mapping, index, in split_huge_pages_in_file()
|
D | migrate.c | 194 new = page - pvmw.page->index + in remove_migration_pte() 396 newpage->index = page->index; in migrate_page_move_mapping() 422 newpage->index = page->index; in migrate_page_move_mapping() 527 newpage->index = page->index; in migrate_huge_page_move_mapping()
|
D | gup.c | 334 unsigned long index; in unpin_user_pages_dirty_lock() local 343 for_each_compound_head(index, pages, npages, head, ntails) { in unpin_user_pages_dirty_lock() 395 unsigned long index; in unpin_user_page_range_dirty_lock() local 399 for_each_compound_range(index, &page, npages, head, ntails) { in unpin_user_page_range_dirty_lock() 418 unsigned long index; in unpin_user_pages() local 430 for_each_compound_head(index, pages, npages, head, ntails) in unpin_user_pages()
|
D | swapfile.c | 610 cluster_set_null(&percpu_cluster->index); in scan_swap_map_ssd_cluster_conflict() 627 if (cluster_is_null(&cluster->index)) { in scan_swap_map_try_ssd_cluster() 629 cluster->index = si->free_clusters.head; in scan_swap_map_try_ssd_cluster() 630 cluster->next = cluster_next(&cluster->index) * in scan_swap_map_try_ssd_cluster() 652 (cluster_next(&cluster->index) + 1) * SWAPFILE_CLUSTER); in scan_swap_map_try_ssd_cluster() 663 cluster_set_null(&cluster->index); in scan_swap_map_try_ssd_cluster() 3267 cluster_set_null(&cluster->index); in SYSCALL_DEFINE2()
|
D | memfd.c | 50 xas_set(xas, page->index + cache_count); in memfd_tag_pins()
|
D | swap.c | 1178 struct address_space *mapping, pgoff_t *index, pgoff_t end, in pagevec_lookup_range_tag() argument 1181 pvec->nr = find_get_pages_range_tag(mapping, index, end, tag, in pagevec_lookup_range_tag()
|
/mm/kfence/ |
D | kfence.h | 101 long index; in addr_to_metadata() local 113 index = (addr - (unsigned long)__kfence_pool) / (PAGE_SIZE * 2) - 1; in addr_to_metadata() 114 if (index < 0 || index >= CONFIG_KFENCE_NUM_OBJECTS) in addr_to_metadata() 117 return &kfence_metadata[index]; in addr_to_metadata()
|
/mm/damon/ |
D | Kconfig | 12 See https://damonitor.github.io/doc/html/latest-damon/index.html for
|