Home
last modified time | relevance | path

Searched refs:page (Results 1 – 5 of 5) sorted by relevance

/lib/
Dscatterlist.c365 struct page **pages, unsigned int n_pages, in sg_alloc_table_from_pages()
555 miter->page = sg_page_iter_page(&miter->piter); in sg_miter_next()
559 miter->addr = kmap_atomic(miter->page) + miter->__offset; in sg_miter_next()
561 miter->addr = kmap(miter->page) + miter->__offset; in sg_miter_next()
591 !PageSlab(miter->page)) in sg_miter_stop()
592 flush_kernel_dcache_page(miter->page); in sg_miter_stop()
598 kunmap(miter->page); in sg_miter_stop()
600 miter->page = NULL; in sg_miter_stop()
Dkfifo.c314 struct page *page; in setup_sgl_buf() local
323 page = virt_to_page(buf); in setup_sgl_buf()
328 struct page *npage; in setup_sgl_buf()
333 if (page_to_phys(page) != page_to_phys(npage) - l) { in setup_sgl_buf()
334 sg_set_page(sgl, page, l - off, off); in setup_sgl_buf()
338 page = npage; in setup_sgl_buf()
343 sg_set_page(sgl, page, len, off); in setup_sgl_buf()
Ddma-debug.c560 void debug_dma_assert_idle(struct page *page) in debug_dma_assert_idle() argument
569 if (!page) in debug_dma_assert_idle()
572 cln = (phys_addr_t) page_to_pfn(page) << CACHELINE_PER_PAGE_SHIFT; in debug_dma_assert_idle()
1240 void debug_dma_map_page(struct device *dev, struct page *page, size_t offset, in debug_dma_map_page() argument
1258 entry->pfn = page_to_pfn(page); in debug_dma_map_page()
1268 if (!PageHighMem(page)) { in debug_dma_map_page()
1269 void *addr = page_address(page) + offset; in debug_dma_map_page()
Dswiotlb.c730 dma_addr_t swiotlb_map_page(struct device *dev, struct page *page, in swiotlb_map_page() argument
735 phys_addr_t map, phys = page_to_phys(page) + offset; in swiotlb_map_page()
DKconfig.debug550 Enable some costly sanity checks in virtual to page code. This can
1683 This also attempts to catch cases where a page owned by DMA is
1685 example, this enables cow_user_page() to check that the source page is