Searched refs:page (Results 1 – 5 of 5) sorted by relevance
/lib/ |
D | scatterlist.c | 365 struct page **pages, unsigned int n_pages, in sg_alloc_table_from_pages() 555 miter->page = sg_page_iter_page(&miter->piter); in sg_miter_next() 559 miter->addr = kmap_atomic(miter->page) + miter->__offset; in sg_miter_next() 561 miter->addr = kmap(miter->page) + miter->__offset; in sg_miter_next() 591 !PageSlab(miter->page)) in sg_miter_stop() 592 flush_kernel_dcache_page(miter->page); in sg_miter_stop() 598 kunmap(miter->page); in sg_miter_stop() 600 miter->page = NULL; in sg_miter_stop()
|
D | kfifo.c | 314 struct page *page; in setup_sgl_buf() local 323 page = virt_to_page(buf); in setup_sgl_buf() 328 struct page *npage; in setup_sgl_buf() 333 if (page_to_phys(page) != page_to_phys(npage) - l) { in setup_sgl_buf() 334 sg_set_page(sgl, page, l - off, off); in setup_sgl_buf() 338 page = npage; in setup_sgl_buf() 343 sg_set_page(sgl, page, len, off); in setup_sgl_buf()
|
D | dma-debug.c | 560 void debug_dma_assert_idle(struct page *page) in debug_dma_assert_idle() argument 569 if (!page) in debug_dma_assert_idle() 572 cln = (phys_addr_t) page_to_pfn(page) << CACHELINE_PER_PAGE_SHIFT; in debug_dma_assert_idle() 1240 void debug_dma_map_page(struct device *dev, struct page *page, size_t offset, in debug_dma_map_page() argument 1258 entry->pfn = page_to_pfn(page); in debug_dma_map_page() 1268 if (!PageHighMem(page)) { in debug_dma_map_page() 1269 void *addr = page_address(page) + offset; in debug_dma_map_page()
|
D | swiotlb.c | 730 dma_addr_t swiotlb_map_page(struct device *dev, struct page *page, in swiotlb_map_page() argument 735 phys_addr_t map, phys = page_to_phys(page) + offset; in swiotlb_map_page()
|
D | Kconfig.debug | 550 Enable some costly sanity checks in virtual to page code. This can 1683 This also attempts to catch cases where a page owned by DMA is 1685 example, this enables cow_user_page() to check that the source page is
|