/lib/ |
D | test_hmm.c | 48 #define BACKING_PAGE(page) (is_device_private_page((page)) ? \ argument 49 (page)->zone_device_data : (page)) 121 struct page *free_pages; 196 static struct dmirror_chunk *dmirror_page_to_chunk(struct page *page) in dmirror_page_to_chunk() argument 198 return container_of(page->pgmap, struct dmirror_chunk, pagemap); in dmirror_page_to_chunk() 201 static struct dmirror_device *dmirror_page_to_device(struct page *page) in dmirror_page_to_device() argument 204 return dmirror_page_to_chunk(page)->mdevice; in dmirror_page_to_device() 215 struct page *page; in dmirror_do_fault() local 225 page = hmm_pfn_to_page(*pfns); in dmirror_do_fault() 226 WARN_ON(!page); in dmirror_do_fault() [all …]
|
D | iov_iter.c | 227 pipe->bufs[idx].page, in sanity() 237 static struct page *push_anon(struct pipe_inode_info *pipe, unsigned size) in push_anon() 239 struct page *page = alloc_page(GFP_USER); in push_anon() local 240 if (page) { in push_anon() 244 .page = page, in push_anon() 249 return page; in push_anon() 252 static void push_page(struct pipe_inode_info *pipe, struct page *page, in push_page() argument 258 .page = page, in push_page() 262 get_page(page); in push_page() 273 static struct page *append_pipe(struct iov_iter *i, size_t size, in append_pipe() [all …]
|
D | test_free_pages.c | 20 struct page *page = virt_to_page((void *)addr); in test_free_pages() local 23 get_page(page); in test_free_pages() 25 put_page(page); in test_free_pages()
|
D | buildid.c | 130 struct page *page; in build_id_parse() local 138 page = find_get_page(vma->vm_file->f_mapping, 0); in build_id_parse() 139 if (!page) in build_id_parse() 143 page_addr = kmap_atomic(page); in build_id_parse() 160 put_page(page); in build_id_parse()
|
D | scatterlist.c | 443 struct page **pages, unsigned int n_pages, unsigned int offset, in sg_alloc_append_table_from_pages() 565 int sg_alloc_table_from_pages_segment(struct sg_table *sgt, struct page **pages, in sg_alloc_table_from_pages_segment() 603 struct page *page; in sgl_alloc_order() local 627 page = alloc_pages(gfp, order); in sgl_alloc_order() 628 if (!page) { in sgl_alloc_order() 633 sg_set_page(sg, page, elem_len, 0); in sgl_alloc_order() 675 struct page *page; in sgl_free_n_order() local 681 page = sg_page(sg); in sgl_free_n_order() 682 if (page) in sgl_free_n_order() 683 __free_pages(page, order); in sgl_free_n_order() [all …]
|
D | test_meminit.c | 65 struct page *page; in do_alloc_pages_order() local 69 page = alloc_pages(GFP_KERNEL, order); in do_alloc_pages_order() 70 if (!page) in do_alloc_pages_order() 72 buf = page_address(page); in do_alloc_pages_order() 74 __free_pages(page, order); in do_alloc_pages_order() 76 page = alloc_pages(GFP_KERNEL, order); in do_alloc_pages_order() 77 if (!page) in do_alloc_pages_order() 79 buf = page_address(page); in do_alloc_pages_order() 83 __free_pages(page, order); in do_alloc_pages_order()
|
D | test_lockup.c | 306 struct page *page; in test_alloc_pages() local 310 page = alloc_pages(alloc_pages_gfp, alloc_pages_order); in test_alloc_pages() 311 if (!page) { in test_alloc_pages() 315 list_add(&page->lru, pages); in test_alloc_pages() 321 struct page *page, *next; in test_free_pages() local 323 list_for_each_entry_safe(page, next, pages, lru) in test_free_pages() 324 __free_pages(page, alloc_pages_order); in test_free_pages()
|
D | stackdepot.c | 426 struct page *page = NULL; in __stack_depot_save() local 474 page = alloc_pages(alloc_flags, STACK_ALLOC_ORDER); in __stack_depot_save() 475 if (page) in __stack_depot_save() 476 prealloc = page_address(page); in __stack_depot_save()
|
D | kfifo.c | 301 struct page *page; in setup_sgl_buf() local 310 page = virt_to_page(buf); in setup_sgl_buf() 315 struct page *npage; in setup_sgl_buf() 320 if (page_to_phys(page) != page_to_phys(npage) - l) { in setup_sgl_buf() 321 sg_set_page(sgl, page, l - off, off); in setup_sgl_buf() 325 page = npage; in setup_sgl_buf() 330 sg_set_page(sgl, page, len, off); in setup_sgl_buf()
|
D | test_bpf.c | 14363 struct page *page; in generate_test_data() local 14387 page = alloc_page(GFP_KERNEL); in generate_test_data() 14389 if (!page) in generate_test_data() 14392 ptr = kmap(page); in generate_test_data() 14396 kunmap(page); in generate_test_data() 14397 skb_add_rx_frag(skb, 0, page, 0, MAX_DATA, MAX_DATA); in generate_test_data() 14403 __free_page(page); in generate_test_data() 14608 struct page *page[2]; in build_test_skb() local 14612 page[i] = alloc_page(GFP_KERNEL); in build_test_skb() 14613 if (!page[i]) { in build_test_skb() [all …]
|
D | Kconfig.debug | 855 bool "Debug page-flags operations" 858 Enables extra validation on page flags operations. 863 bool "Debug arch page table for semantics compliance" 869 architecture page table helper functions on various platforms in 885 Enable some costly sanity checks in virtual to page code. This can 2714 tristate "Test heap/page initialization" 2716 Test if the kernel is zero-initializing heap and page allocations. 2738 freeing a block of pages and a speculative page reference.
|