/drivers/gpu/drm/i915/ |
D | i915_vma_resource.c | 53 struct i915_vma_resource *vma_res = in i915_vma_resource_alloc() local 56 return vma_res ? vma_res : ERR_PTR(-ENOMEM); in i915_vma_resource_alloc() 63 void i915_vma_resource_free(struct i915_vma_resource *vma_res) in i915_vma_resource_free() argument 65 if (vma_res) in i915_vma_resource_free() 66 kmem_cache_free(slab_vma_resources, vma_res); in i915_vma_resource_free() 81 struct i915_vma_resource *vma_res = in unbind_fence_free_rcu() local 82 container_of(head, typeof(*vma_res), unbind_fence.rcu); in unbind_fence_free_rcu() 84 i915_vma_resource_free(vma_res); in unbind_fence_free_rcu() 89 struct i915_vma_resource *vma_res = in unbind_fence_release() local 90 container_of(fence, typeof(*vma_res), unbind_fence); in unbind_fence_release() [all …]
|
D | i915_vma_resource.h | 127 bool i915_vma_resource_hold(struct i915_vma_resource *vma_res, 130 void i915_vma_resource_unhold(struct i915_vma_resource *vma_res, 135 void i915_vma_resource_free(struct i915_vma_resource *vma_res); 137 struct dma_fence *i915_vma_resource_unbind(struct i915_vma_resource *vma_res, 140 void __i915_vma_resource_init(struct i915_vma_resource *vma_res); 149 *i915_vma_resource_get(struct i915_vma_resource *vma_res) in i915_vma_resource_get() argument 151 dma_fence_get(&vma_res->unbind_fence); in i915_vma_resource_get() 152 return vma_res; in i915_vma_resource_get() 159 static inline void i915_vma_resource_put(struct i915_vma_resource *vma_res) in i915_vma_resource_put() argument 161 dma_fence_put(&vma_res->unbind_fence); in i915_vma_resource_put() [all …]
|
D | i915_vma.c | 312 struct i915_vma_resource *vma_res; member 322 struct i915_vma_resource *vma_res = vw->vma_res; in __vma_bind() local 333 vma_res->ops->bind_vma(vma_res->vm, &vw->stash, in __vma_bind() 334 vma_res, vw->cache_level, vw->flags); in __vma_bind() 345 if (vw->vma_res) in __vma_release() 346 i915_vma_resource_put(vw->vma_res); in __vma_release() 411 i915_vma_resource_init_from_vma(struct i915_vma_resource *vma_res, in i915_vma_resource_init_from_vma() argument 416 i915_vma_resource_init(vma_res, vma->vm, vma->pages, &vma->page_sizes, in i915_vma_resource_init_from_vma() 440 struct i915_vma_resource *vma_res) in i915_vma_bind() argument 453 i915_vma_resource_free(vma_res); in i915_vma_bind() [all …]
|
D | i915_gpu_error.c | 1075 const struct i915_vma_resource *vma_res, in i915_vma_coredump_create() argument 1088 if (!vma_res || !vma_res->bi.pages || !compress) in i915_vma_coredump_create() 1104 dst->gtt_offset = vma_res->start; in i915_vma_coredump_create() 1105 dst->gtt_size = vma_res->node_size; in i915_vma_coredump_create() 1106 dst->gtt_page_sizes = vma_res->page_sizes_gtt; in i915_vma_coredump_create() 1114 for_each_sgt_daddr(dma, iter, vma_res->bi.pages) { in i915_vma_coredump_create() 1136 } else if (vma_res->bi.lmem) { in i915_vma_coredump_create() 1137 struct intel_memory_region *mem = vma_res->mr; in i915_vma_coredump_create() 1140 for_each_sgt_daddr(dma, iter, vma_res->bi.pages) { in i915_vma_coredump_create() 1160 for_each_sgt_page(page, iter, vma_res->bi.pages) { in i915_vma_coredump_create() [all …]
|
D | i915_vma.h | 209 struct i915_vma_resource *vma_res); 434 void i915_vma_resource_init_from_vma(struct i915_vma_resource *vma_res,
|
D | i915_request.h | 55 struct i915_vma_resource *vma_res; member
|
D | i915_request.c | 337 i915_vma_resource_put(capture->vma_res); in i915_request_free_capture_list()
|
/drivers/gpu/drm/i915/gt/ |
D | intel_ppgtt.c | 183 struct i915_vma_resource *vma_res, in ppgtt_bind_vma() argument 189 if (!vma_res->allocated) { in ppgtt_bind_vma() 190 vm->allocate_va_range(vm, stash, vma_res->start, in ppgtt_bind_vma() 191 vma_res->vma_size); in ppgtt_bind_vma() 192 vma_res->allocated = true; in ppgtt_bind_vma() 197 if (vma_res->bi.readonly) in ppgtt_bind_vma() 199 if (vma_res->bi.lmem) in ppgtt_bind_vma() 202 vm->insert_entries(vm, vma_res, cache_level, pte_flags); in ppgtt_bind_vma() 207 struct i915_vma_resource *vma_res) in ppgtt_unbind_vma() argument 209 if (!vma_res->allocated) in ppgtt_unbind_vma() [all …]
|
D | intel_ggtt.c | 273 struct i915_vma_resource *vma_res, in gen8_ggtt_insert_entries() argument 290 gte += vma_res->start / I915_GTT_PAGE_SIZE; in gen8_ggtt_insert_entries() 291 end = gte + vma_res->node_size / I915_GTT_PAGE_SIZE; in gen8_ggtt_insert_entries() 293 for_each_sgt_daddr(addr, iter, vma_res->bi.pages) in gen8_ggtt_insert_entries() 330 struct i915_vma_resource *vma_res, in gen6_ggtt_insert_entries() argument 341 gte += vma_res->start / I915_GTT_PAGE_SIZE; in gen6_ggtt_insert_entries() 342 end = gte + vma_res->node_size / I915_GTT_PAGE_SIZE; in gen6_ggtt_insert_entries() 344 for_each_sgt_daddr(addr, iter, vma_res->bi.pages) in gen6_ggtt_insert_entries() 427 struct i915_vma_resource *vma_res; member 436 gen8_ggtt_insert_entries(arg->vm, arg->vma_res, arg->level, arg->flags); in bxt_vtd_ggtt_insert_entries__cb() [all …]
|
D | gen8_ppgtt.c | 471 struct i915_vma_resource *vma_res, in xehpsdv_ppgtt_insert_huge() argument 478 u64 start = vma_res->start; in xehpsdv_ppgtt_insert_huge() 496 if (vma_res->bi.page_sizes.sg & I915_GTT_PAGE_SIZE_2M && in xehpsdv_ppgtt_insert_huge() 554 vma_res->page_sizes_gtt |= page_size; in xehpsdv_ppgtt_insert_huge() 559 struct i915_vma_resource *vma_res, in gen8_ppgtt_insert_huge() argument 566 u64 start = vma_res->start; in gen8_ppgtt_insert_huge() 581 if (vma_res->bi.page_sizes.sg & I915_GTT_PAGE_SIZE_2M && in gen8_ppgtt_insert_huge() 598 vma_res->bi.page_sizes.sg & I915_GTT_PAGE_SIZE_64K && in gen8_ppgtt_insert_huge() 648 !iter->sg && IS_ALIGNED(vma_res->start + in gen8_ppgtt_insert_huge() 649 vma_res->node_size, in gen8_ppgtt_insert_huge() [all …]
|
D | intel_gtt.h | 217 struct i915_vma_resource *vma_res, 225 struct i915_vma_resource *vma_res); 306 struct i915_vma_resource *vma_res, 315 struct i915_vma_resource *vma_res, 572 struct i915_vma_resource *vma_res, 576 struct i915_vma_resource *vma_res); 662 struct i915_vma_resource *vma_res, 666 struct i915_vma_resource *vma_res); 689 } sgt_dma(struct i915_vma_resource *vma_res) { in sgt_dma() argument 690 struct scatterlist *sg = vma_res->bi.pages->sgl; in sgt_dma()
|
D | gen6_ppgtt.c | 111 struct i915_vma_resource *vma_res, in gen6_ppgtt_insert_entries() argument 117 unsigned int first_entry = vma_res->start / I915_GTT_PAGE_SIZE; in gen6_ppgtt_insert_entries() 121 struct sgt_dma iter = sgt_dma(vma_res); in gen6_ppgtt_insert_entries() 147 vma_res->page_sizes_gtt = I915_GTT_PAGE_SIZE; in gen6_ppgtt_insert_entries() 280 struct i915_vma_resource *vma_res, in pd_vma_bind() argument 285 struct gen6_ppgtt *ppgtt = vma_res->private; in pd_vma_bind() 286 u32 ggtt_offset = vma_res->start / I915_GTT_PAGE_SIZE; in pd_vma_bind() 295 struct i915_vma_resource *vma_res) in pd_vma_unbind() argument 297 struct gen6_ppgtt *ppgtt = vma_res->private; in pd_vma_unbind()
|
D | intel_ggtt_gmch.c | 32 struct i915_vma_resource *vma_res, in gmch_ggtt_insert_entries() argument 39 intel_gmch_gtt_insert_sg_entries(vma_res->bi.pages, vma_res->start >> PAGE_SHIFT, in gmch_ggtt_insert_entries()
|
D | intel_engine_cs.c | 2048 struct i915_vma_resource *vma_res = rq->batch_res; in print_request_ring() local 2055 vma_res ? upper_32_bits(vma_res->start) : ~0u, in print_request_ring() 2056 vma_res ? lower_32_bits(vma_res->start) : ~0u); in print_request_ring()
|
/drivers/gpu/drm/i915/display/ |
D | intel_dpt.c | 54 struct i915_vma_resource *vma_res, in dpt_insert_entries() argument 70 i = vma_res->start / I915_GTT_PAGE_SIZE; in dpt_insert_entries() 71 for_each_sgt_daddr(addr, sgt_iter, vma_res->bi.pages) in dpt_insert_entries() 82 struct i915_vma_resource *vma_res, in dpt_bind_vma() argument 88 if (vma_res->bound_flags) in dpt_bind_vma() 93 if (vm->has_read_only && vma_res->bi.readonly) in dpt_bind_vma() 95 if (vma_res->bi.lmem) in dpt_bind_vma() 98 vm->insert_entries(vm, vma_res, cache_level, pte_flags); in dpt_bind_vma() 100 vma_res->page_sizes_gtt = I915_GTT_PAGE_SIZE; in dpt_bind_vma() 107 vma_res->bound_flags = I915_VMA_GLOBAL_BIND | I915_VMA_LOCAL_BIND; in dpt_bind_vma() [all …]
|
/drivers/gpu/drm/i915/selftests/ |
D | mock_gtt.c | 36 struct i915_vma_resource *vma_res, in mock_insert_entries() argument 43 struct i915_vma_resource *vma_res, in mock_bind_ppgtt() argument 48 vma_res->bound_flags |= flags; in mock_bind_ppgtt() 52 struct i915_vma_resource *vma_res) in mock_unbind_ppgtt() argument 96 struct i915_vma_resource *vma_res, in mock_bind_ggtt() argument 103 struct i915_vma_resource *vma_res) in mock_unbind_ggtt() argument
|
D | i915_gem_gtt.c | 1505 struct i915_vma_resource *vma_res; in reserve_gtt_with_resource() local 1509 vma_res = i915_vma_resource_alloc(); in reserve_gtt_with_resource() 1510 if (IS_ERR(vma_res)) in reserve_gtt_with_resource() 1511 return PTR_ERR(vma_res); in reserve_gtt_with_resource() 1519 i915_vma_resource_init_from_vma(vma_res, vma); in reserve_gtt_with_resource() 1520 vma->resource = vma_res; in reserve_gtt_with_resource() 1522 kfree(vma_res); in reserve_gtt_with_resource() 1686 struct i915_vma_resource *vma_res; in insert_gtt_with_resource() local 1690 vma_res = i915_vma_resource_alloc(); in insert_gtt_with_resource() 1691 if (IS_ERR(vma_res)) in insert_gtt_with_resource() [all …]
|
/drivers/gpu/drm/i915/gem/ |
D | i915_gem_execbuffer.c | 2015 capture->vma_res = i915_vma_resource_get(vma->resource); in eb_capture_stage()
|