/drivers/iommu/iommufd/ |
D | pages.c | 163 static void iopt_pages_add_npinned(struct iopt_pages *pages, size_t npages) in iopt_pages_add_npinned() 172 static void iopt_pages_sub_npinned(struct iopt_pages *pages, size_t npages) in iopt_pages_sub_npinned() 181 static void iopt_pages_err_unpin(struct iopt_pages *pages, in iopt_pages_err_unpin() 249 static struct iopt_area *iopt_pages_find_domain_area(struct iopt_pages *pages, in iopt_pages_find_domain_area() 576 unsigned long last_index, struct page **pages) in pages_to_xarray() 615 static void batch_from_pages(struct pfn_batch *batch, struct page **pages, in batch_from_pages() 625 static void batch_unpin(struct pfn_batch *batch, struct iopt_pages *pages, in batch_unpin() 709 struct iopt_pages *pages) in pfn_reader_user_init() 722 struct iopt_pages *pages) in pfn_reader_user_destroy() 737 struct iopt_pages *pages, in pfn_reader_user_pin() [all …]
|
D | io_pagetable.c | 23 struct iopt_pages *pages; member 190 struct iopt_pages *pages, unsigned long iova, in iopt_insert_area() 487 struct iopt_pages *pages; in iopt_unmap_iova_range() local 715 struct iopt_pages *pages = area->pages; in iopt_unfill_domain() local 734 struct iopt_pages *pages = area->pages; in iopt_unfill_domain() local 768 struct iopt_pages *pages = area->pages; in iopt_fill_domain() local 793 struct iopt_pages *pages = area->pages; in iopt_fill_domain() local 1003 struct iopt_pages *pages = area->pages; in iopt_area_split() local
|
/drivers/xen/ |
D | xlate_mmu.c | 48 static void xen_for_each_gfn(struct page **pages, unsigned nr_gfn, in xen_for_each_gfn() 71 struct page **pages; member 148 struct page **pages) in xen_xlate_remap_gfn_array() 184 int nr, struct page **pages) in xen_xlate_unmap_gfn_range() 217 struct page **pages; in xen_xlate_map_ballooned_pages() local 267 struct page **pages; member
|
D | privcmd.c | 95 static void free_page_list(struct list_head *pages) in free_page_list() 344 struct page **pages = vma->vm_private_data; in mmap_batch_fn() local 426 struct page **pages; in alloc_empty_pages() local 587 struct page *pages[], unsigned int nr_pages, unsigned int *pinned) in lock_pages() 618 static void unlock_pages(struct page *pages[], unsigned int nr_pages) in unlock_pages() 629 struct page **pages = NULL; in privcmd_ioctl_dm_op() local 777 struct page **pages; in privcmd_ioctl_mmap_resource() local 1173 struct page **pages = vma->vm_private_data; in privcmd_close() local
|
D | mem-reservation.c | 30 struct page **pages, in __xenmem_reservation_va_mapping_update() 59 struct page **pages) in __xenmem_reservation_va_mapping_reset()
|
D | unpopulated-alloc.c | 158 int xen_alloc_unpopulated_pages(unsigned int nr_pages, struct page **pages) in xen_alloc_unpopulated_pages() 214 void xen_free_unpopulated_pages(unsigned int nr_pages, struct page **pages) in xen_free_unpopulated_pages()
|
D | balloon.c | 600 int xen_alloc_ballooned_pages(unsigned int nr_pages, struct page **pages) in xen_alloc_ballooned_pages() 653 void xen_free_ballooned_pages(unsigned int nr_pages, struct page **pages) in xen_free_ballooned_pages() 676 unsigned long start_pfn, pages; in balloon_add_regions() local
|
/drivers/hwtracing/coresight/ |
D | coresight-tmc-etr.c | 44 void **pages; member 191 enum dma_data_direction dir, void **pages) in tmc_pages_alloc() 287 static int tmc_alloc_data_pages(struct tmc_sg_table *sg_table, void **pages) in tmc_alloc_data_pages() 321 void **pages) in tmc_alloc_sg_table() 563 unsigned long size, void **pages) in tmc_init_etr_sg_table() 598 void **pages) in tmc_etr_alloc_flat_buf() 697 void **pages) in tmc_etr_alloc_sg_buf() 814 void **pages) in tmc_etr_mode_alloc_buf() 843 int node, void **pages) in tmc_alloc_etr_buf() 1262 int nr_pages, void **pages, bool snapshot) in alloc_etr_buf() [all …]
|
/drivers/media/common/videobuf2/ |
D | frame_vector.c | 82 struct page **pages; in put_vaddr_frames() local 114 struct page **pages; in frame_vector_to_pages() local 140 struct page **pages; in frame_vector_to_pfns() local
|
/drivers/gpu/drm/i915/gem/selftests/ |
D | huge_gem_object.c | 12 struct sg_table *pages) in huge_free_pages() 34 struct sg_table *pages; in huge_get_pages() local 86 struct sg_table *pages) in huge_put_pages()
|
/drivers/block/xen-blkback/ |
D | blkback.c | 246 struct page *pages[BLKIF_MAX_SEGMENTS_PER_REQUEST]; in free_persistent_gnts() local 293 struct page *pages[BLKIF_MAX_SEGMENTS_PER_REQUEST]; in xen_blkbk_unmap_purged_grants() local 647 struct grant_page **pages, in xen_blkbk_unmap_prepare() 707 struct grant_page **pages = req->segments; in xen_blkbk_unmap_and_respond() local 732 struct grant_page *pages[], in xen_blkbk_unmap() 757 struct grant_page *pages[], in xen_blkbk_map() 918 struct grant_page **pages = pending_req->indirect_pages; in xen_blkbk_parse_indirect() local 1304 struct grant_page **pages = pending_req->segments; in dispatch_rw_block_io() local
|
/drivers/staging/media/ipu3/ |
D | ipu3-dmamap.c | 20 static void imgu_dmamap_free_buffer(struct page **pages, in imgu_dmamap_free_buffer() 36 struct page **pages; in imgu_dmamap_alloc_buffer() local 100 struct page **pages; in imgu_dmamap_alloc() local
|
/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_gart.c | 291 int pages) in amdgpu_gart_unbind() 342 int pages, dma_addr_t *dma_addr, uint64_t flags, in amdgpu_gart_map() 378 int pages, dma_addr_t *dma_addr, in amdgpu_gart_bind()
|
/drivers/gpu/drm/i915/gem/ |
D | i915_gem_pages.c | 19 struct sg_table *pages) in __i915_gem_object_set_pages() 211 struct sg_table *pages; in __i915_gem_object_unset_pages() local 240 struct sg_table *pages; in __i915_gem_object_put_pages() local 274 struct page *stack[32], **pages = stack, *page; in i915_gem_object_map_page() local
|
D | i915_gem_phys.c | 101 struct sg_table *pages) in i915_gem_object_put_pages_phys() 197 struct sg_table *pages; in i915_gem_object_shmem_to_phys() local
|
/drivers/gpu/drm/i915/selftests/ |
D | mock_region.c | 16 struct sg_table *pages) in mock_region_put_pages() 25 struct sg_table *pages; in mock_region_get_pages() local
|
/drivers/usb/storage/ |
D | alauda.c | 732 unsigned int page, unsigned int pages, unsigned char *data) in alauda_read_block_raw() 758 unsigned int page, unsigned int pages, unsigned char *data) in alauda_read_block() 811 unsigned int page, unsigned int pages, in alauda_write_lba() 955 unsigned int pages; in alauda_read_data() local 1056 unsigned int pages = min(sectors, blocksize - page); in alauda_write_data() local 1174 unsigned int page, pages; in alauda_transport() local 1191 unsigned int page, pages; in alauda_transport() local
|
/drivers/gpu/drm/i915/ |
D | i915_gem_gtt.c | 29 struct sg_table *pages) in i915_gem_gtt_prepare_pages() 57 struct sg_table *pages) in i915_gem_gtt_finish_pages()
|
/drivers/infiniband/hw/hfi1/ |
D | user_exp_rcv.c | 139 struct page **pages; in unpin_rcv_pages() local 164 struct page **pages = NULL; in pin_rcv_pages() local 569 struct page **pages = tidbuf->pages; in find_phys_blocks() local 733 struct page **pages = tbuf->pages + pageidx; in set_rcvarray_entry() local
|
D | pin_system.c | 19 struct page **pages; member 65 static void unpin_vector_pages(struct mm_struct *mm, struct page **pages, in unpin_vector_pages() 120 struct page **pages; in pin_system_pages() local
|
/drivers/iommu/ |
D | io-pgtable-arm.c | 26 static dma_addr_t __arm_lpae_dma_addr(void *pages) in __arm_lpae_dma_addr() 37 void *pages; in __arm_lpae_alloc_pages() local 68 void __arm_lpae_free_pages(void *pages, size_t size, struct io_pgtable_cfg *cfg) in __arm_lpae_free_pages()
|
D | dma-iommu.c | 97 unsigned long pages; member 177 unsigned long pfn, unsigned long pages, in queue_iova() 780 static void __iommu_dma_free_pages(struct page **pages, int count) in __iommu_dma_free_pages() 790 struct page **pages; in __iommu_dma_alloc_pages() local 854 struct page **pages; in __iommu_dma_alloc_noncontiguous() local 923 struct page **pages; in iommu_dma_alloc_remap() local 1439 struct page *page = NULL, **pages = NULL; in __iommu_dma_free() local 1562 struct page **pages = dma_common_find_pages(cpu_addr); in iommu_dma_mmap() local 1584 struct page **pages = dma_common_find_pages(cpu_addr); in iommu_dma_get_sgtable() local
|
/drivers/gpu/drm/radeon/ |
D | radeon_gart.c | 242 int pages) in radeon_gart_unbind() 286 int pages, struct page **pagelist, dma_addr_t *dma_addr, in radeon_gart_bind()
|
/drivers/gpu/drm/etnaviv/ |
D | etnaviv_gem.c | 168 struct page **pages; in etnaviv_gem_fault() local 247 struct page **pages; in etnaviv_gem_mapping_get() local 345 struct page **pages; in etnaviv_gem_vmap_impl() local 665 struct page **pages = pvec + pinned; in etnaviv_gem_userptr_get_pages() local
|
/drivers/gpu/drm/ |
D | drm_vma_manager.c | 142 unsigned long pages) in drm_vma_offset_lookup_locked() 202 struct drm_vma_offset_node *node, unsigned long pages) in drm_vma_offset_add()
|