Home
last modified time | relevance | path

Searched refs:page (Results 1 – 11 of 11) sorted by relevance

/lib/
Dtest_hmm.c48 #define BACKING_PAGE(page) (is_device_private_page((page)) ? \ argument
49 (page)->zone_device_data : (page))
121 struct page *free_pages;
196 static struct dmirror_chunk *dmirror_page_to_chunk(struct page *page) in dmirror_page_to_chunk() argument
198 return container_of(page->pgmap, struct dmirror_chunk, pagemap); in dmirror_page_to_chunk()
201 static struct dmirror_device *dmirror_page_to_device(struct page *page) in dmirror_page_to_device() argument
204 return dmirror_page_to_chunk(page)->mdevice; in dmirror_page_to_device()
215 struct page *page; in dmirror_do_fault() local
225 page = hmm_pfn_to_page(*pfns); in dmirror_do_fault()
226 WARN_ON(!page); in dmirror_do_fault()
[all …]
Diov_iter.c227 pipe->bufs[idx].page, in sanity()
237 static struct page *push_anon(struct pipe_inode_info *pipe, unsigned size) in push_anon()
239 struct page *page = alloc_page(GFP_USER); in push_anon() local
240 if (page) { in push_anon()
244 .page = page, in push_anon()
249 return page; in push_anon()
252 static void push_page(struct pipe_inode_info *pipe, struct page *page, in push_page() argument
258 .page = page, in push_page()
262 get_page(page); in push_page()
273 static struct page *append_pipe(struct iov_iter *i, size_t size, in append_pipe()
[all …]
Dtest_free_pages.c20 struct page *page = virt_to_page((void *)addr); in test_free_pages() local
23 get_page(page); in test_free_pages()
25 put_page(page); in test_free_pages()
Dbuildid.c130 struct page *page; in build_id_parse() local
138 page = find_get_page(vma->vm_file->f_mapping, 0); in build_id_parse()
139 if (!page) in build_id_parse()
143 page_addr = kmap_atomic(page); in build_id_parse()
160 put_page(page); in build_id_parse()
Dscatterlist.c443 struct page **pages, unsigned int n_pages, unsigned int offset, in sg_alloc_append_table_from_pages()
565 int sg_alloc_table_from_pages_segment(struct sg_table *sgt, struct page **pages, in sg_alloc_table_from_pages_segment()
603 struct page *page; in sgl_alloc_order() local
627 page = alloc_pages(gfp, order); in sgl_alloc_order()
628 if (!page) { in sgl_alloc_order()
633 sg_set_page(sg, page, elem_len, 0); in sgl_alloc_order()
675 struct page *page; in sgl_free_n_order() local
681 page = sg_page(sg); in sgl_free_n_order()
682 if (page) in sgl_free_n_order()
683 __free_pages(page, order); in sgl_free_n_order()
[all …]
Dtest_meminit.c65 struct page *page; in do_alloc_pages_order() local
69 page = alloc_pages(GFP_KERNEL, order); in do_alloc_pages_order()
70 if (!page) in do_alloc_pages_order()
72 buf = page_address(page); in do_alloc_pages_order()
74 __free_pages(page, order); in do_alloc_pages_order()
76 page = alloc_pages(GFP_KERNEL, order); in do_alloc_pages_order()
77 if (!page) in do_alloc_pages_order()
79 buf = page_address(page); in do_alloc_pages_order()
83 __free_pages(page, order); in do_alloc_pages_order()
Dtest_lockup.c306 struct page *page; in test_alloc_pages() local
310 page = alloc_pages(alloc_pages_gfp, alloc_pages_order); in test_alloc_pages()
311 if (!page) { in test_alloc_pages()
315 list_add(&page->lru, pages); in test_alloc_pages()
321 struct page *page, *next; in test_free_pages() local
323 list_for_each_entry_safe(page, next, pages, lru) in test_free_pages()
324 __free_pages(page, alloc_pages_order); in test_free_pages()
Dstackdepot.c426 struct page *page = NULL; in __stack_depot_save() local
474 page = alloc_pages(alloc_flags, STACK_ALLOC_ORDER); in __stack_depot_save()
475 if (page) in __stack_depot_save()
476 prealloc = page_address(page); in __stack_depot_save()
Dkfifo.c301 struct page *page; in setup_sgl_buf() local
310 page = virt_to_page(buf); in setup_sgl_buf()
315 struct page *npage; in setup_sgl_buf()
320 if (page_to_phys(page) != page_to_phys(npage) - l) { in setup_sgl_buf()
321 sg_set_page(sgl, page, l - off, off); in setup_sgl_buf()
325 page = npage; in setup_sgl_buf()
330 sg_set_page(sgl, page, len, off); in setup_sgl_buf()
Dtest_bpf.c14363 struct page *page; in generate_test_data() local
14387 page = alloc_page(GFP_KERNEL); in generate_test_data()
14389 if (!page) in generate_test_data()
14392 ptr = kmap(page); in generate_test_data()
14396 kunmap(page); in generate_test_data()
14397 skb_add_rx_frag(skb, 0, page, 0, MAX_DATA, MAX_DATA); in generate_test_data()
14403 __free_page(page); in generate_test_data()
14608 struct page *page[2]; in build_test_skb() local
14612 page[i] = alloc_page(GFP_KERNEL); in build_test_skb()
14613 if (!page[i]) { in build_test_skb()
[all …]
DKconfig.debug855 bool "Debug page-flags operations"
858 Enables extra validation on page flags operations.
863 bool "Debug arch page table for semantics compliance"
869 architecture page table helper functions on various platforms in
885 Enable some costly sanity checks in virtual to page code. This can
2714 tristate "Test heap/page initialization"
2716 Test if the kernel is zero-initializing heap and page allocations.
2738 freeing a block of pages and a speculative page reference.