Home
last modified time | relevance | path

Searched refs:vma_res (Results 1 – 18 of 18) sorted by relevance

/drivers/gpu/drm/i915/
Di915_vma_resource.c53 struct i915_vma_resource *vma_res = in i915_vma_resource_alloc() local
56 return vma_res ? vma_res : ERR_PTR(-ENOMEM); in i915_vma_resource_alloc()
63 void i915_vma_resource_free(struct i915_vma_resource *vma_res) in i915_vma_resource_free() argument
65 if (vma_res) in i915_vma_resource_free()
66 kmem_cache_free(slab_vma_resources, vma_res); in i915_vma_resource_free()
81 struct i915_vma_resource *vma_res = in unbind_fence_free_rcu() local
82 container_of(head, typeof(*vma_res), unbind_fence.rcu); in unbind_fence_free_rcu()
84 i915_vma_resource_free(vma_res); in unbind_fence_free_rcu()
89 struct i915_vma_resource *vma_res = in unbind_fence_release() local
90 container_of(fence, typeof(*vma_res), unbind_fence); in unbind_fence_release()
[all …]
Di915_vma_resource.h127 bool i915_vma_resource_hold(struct i915_vma_resource *vma_res,
130 void i915_vma_resource_unhold(struct i915_vma_resource *vma_res,
135 void i915_vma_resource_free(struct i915_vma_resource *vma_res);
137 struct dma_fence *i915_vma_resource_unbind(struct i915_vma_resource *vma_res,
140 void __i915_vma_resource_init(struct i915_vma_resource *vma_res);
149 *i915_vma_resource_get(struct i915_vma_resource *vma_res) in i915_vma_resource_get() argument
151 dma_fence_get(&vma_res->unbind_fence); in i915_vma_resource_get()
152 return vma_res; in i915_vma_resource_get()
159 static inline void i915_vma_resource_put(struct i915_vma_resource *vma_res) in i915_vma_resource_put() argument
161 dma_fence_put(&vma_res->unbind_fence); in i915_vma_resource_put()
[all …]
Di915_vma.c312 struct i915_vma_resource *vma_res; member
322 struct i915_vma_resource *vma_res = vw->vma_res; in __vma_bind() local
333 vma_res->ops->bind_vma(vma_res->vm, &vw->stash, in __vma_bind()
334 vma_res, vw->cache_level, vw->flags); in __vma_bind()
345 if (vw->vma_res) in __vma_release()
346 i915_vma_resource_put(vw->vma_res); in __vma_release()
411 i915_vma_resource_init_from_vma(struct i915_vma_resource *vma_res, in i915_vma_resource_init_from_vma() argument
416 i915_vma_resource_init(vma_res, vma->vm, vma->pages, &vma->page_sizes, in i915_vma_resource_init_from_vma()
440 struct i915_vma_resource *vma_res) in i915_vma_bind() argument
453 i915_vma_resource_free(vma_res); in i915_vma_bind()
[all …]
Di915_gpu_error.c1075 const struct i915_vma_resource *vma_res, in i915_vma_coredump_create() argument
1088 if (!vma_res || !vma_res->bi.pages || !compress) in i915_vma_coredump_create()
1104 dst->gtt_offset = vma_res->start; in i915_vma_coredump_create()
1105 dst->gtt_size = vma_res->node_size; in i915_vma_coredump_create()
1106 dst->gtt_page_sizes = vma_res->page_sizes_gtt; in i915_vma_coredump_create()
1114 for_each_sgt_daddr(dma, iter, vma_res->bi.pages) { in i915_vma_coredump_create()
1136 } else if (vma_res->bi.lmem) { in i915_vma_coredump_create()
1137 struct intel_memory_region *mem = vma_res->mr; in i915_vma_coredump_create()
1140 for_each_sgt_daddr(dma, iter, vma_res->bi.pages) { in i915_vma_coredump_create()
1160 for_each_sgt_page(page, iter, vma_res->bi.pages) { in i915_vma_coredump_create()
[all …]
Di915_vma.h209 struct i915_vma_resource *vma_res);
434 void i915_vma_resource_init_from_vma(struct i915_vma_resource *vma_res,
Di915_request.h55 struct i915_vma_resource *vma_res; member
Di915_request.c337 i915_vma_resource_put(capture->vma_res); in i915_request_free_capture_list()
/drivers/gpu/drm/i915/gt/
Dintel_ppgtt.c183 struct i915_vma_resource *vma_res, in ppgtt_bind_vma() argument
189 if (!vma_res->allocated) { in ppgtt_bind_vma()
190 vm->allocate_va_range(vm, stash, vma_res->start, in ppgtt_bind_vma()
191 vma_res->vma_size); in ppgtt_bind_vma()
192 vma_res->allocated = true; in ppgtt_bind_vma()
197 if (vma_res->bi.readonly) in ppgtt_bind_vma()
199 if (vma_res->bi.lmem) in ppgtt_bind_vma()
202 vm->insert_entries(vm, vma_res, cache_level, pte_flags); in ppgtt_bind_vma()
207 struct i915_vma_resource *vma_res) in ppgtt_unbind_vma() argument
209 if (!vma_res->allocated) in ppgtt_unbind_vma()
[all …]
Dintel_ggtt.c273 struct i915_vma_resource *vma_res, in gen8_ggtt_insert_entries() argument
290 gte += vma_res->start / I915_GTT_PAGE_SIZE; in gen8_ggtt_insert_entries()
291 end = gte + vma_res->node_size / I915_GTT_PAGE_SIZE; in gen8_ggtt_insert_entries()
293 for_each_sgt_daddr(addr, iter, vma_res->bi.pages) in gen8_ggtt_insert_entries()
330 struct i915_vma_resource *vma_res, in gen6_ggtt_insert_entries() argument
341 gte += vma_res->start / I915_GTT_PAGE_SIZE; in gen6_ggtt_insert_entries()
342 end = gte + vma_res->node_size / I915_GTT_PAGE_SIZE; in gen6_ggtt_insert_entries()
344 for_each_sgt_daddr(addr, iter, vma_res->bi.pages) in gen6_ggtt_insert_entries()
427 struct i915_vma_resource *vma_res; member
436 gen8_ggtt_insert_entries(arg->vm, arg->vma_res, arg->level, arg->flags); in bxt_vtd_ggtt_insert_entries__cb()
[all …]
Dgen8_ppgtt.c471 struct i915_vma_resource *vma_res, in xehpsdv_ppgtt_insert_huge() argument
478 u64 start = vma_res->start; in xehpsdv_ppgtt_insert_huge()
496 if (vma_res->bi.page_sizes.sg & I915_GTT_PAGE_SIZE_2M && in xehpsdv_ppgtt_insert_huge()
554 vma_res->page_sizes_gtt |= page_size; in xehpsdv_ppgtt_insert_huge()
559 struct i915_vma_resource *vma_res, in gen8_ppgtt_insert_huge() argument
566 u64 start = vma_res->start; in gen8_ppgtt_insert_huge()
581 if (vma_res->bi.page_sizes.sg & I915_GTT_PAGE_SIZE_2M && in gen8_ppgtt_insert_huge()
598 vma_res->bi.page_sizes.sg & I915_GTT_PAGE_SIZE_64K && in gen8_ppgtt_insert_huge()
648 !iter->sg && IS_ALIGNED(vma_res->start + in gen8_ppgtt_insert_huge()
649 vma_res->node_size, in gen8_ppgtt_insert_huge()
[all …]
Dintel_gtt.h217 struct i915_vma_resource *vma_res,
225 struct i915_vma_resource *vma_res);
306 struct i915_vma_resource *vma_res,
315 struct i915_vma_resource *vma_res,
572 struct i915_vma_resource *vma_res,
576 struct i915_vma_resource *vma_res);
662 struct i915_vma_resource *vma_res,
666 struct i915_vma_resource *vma_res);
689 } sgt_dma(struct i915_vma_resource *vma_res) { in sgt_dma() argument
690 struct scatterlist *sg = vma_res->bi.pages->sgl; in sgt_dma()
Dgen6_ppgtt.c111 struct i915_vma_resource *vma_res, in gen6_ppgtt_insert_entries() argument
117 unsigned int first_entry = vma_res->start / I915_GTT_PAGE_SIZE; in gen6_ppgtt_insert_entries()
121 struct sgt_dma iter = sgt_dma(vma_res); in gen6_ppgtt_insert_entries()
147 vma_res->page_sizes_gtt = I915_GTT_PAGE_SIZE; in gen6_ppgtt_insert_entries()
280 struct i915_vma_resource *vma_res, in pd_vma_bind() argument
285 struct gen6_ppgtt *ppgtt = vma_res->private; in pd_vma_bind()
286 u32 ggtt_offset = vma_res->start / I915_GTT_PAGE_SIZE; in pd_vma_bind()
295 struct i915_vma_resource *vma_res) in pd_vma_unbind() argument
297 struct gen6_ppgtt *ppgtt = vma_res->private; in pd_vma_unbind()
Dintel_ggtt_gmch.c32 struct i915_vma_resource *vma_res, in gmch_ggtt_insert_entries() argument
39 intel_gmch_gtt_insert_sg_entries(vma_res->bi.pages, vma_res->start >> PAGE_SHIFT, in gmch_ggtt_insert_entries()
Dintel_engine_cs.c2048 struct i915_vma_resource *vma_res = rq->batch_res; in print_request_ring() local
2055 vma_res ? upper_32_bits(vma_res->start) : ~0u, in print_request_ring()
2056 vma_res ? lower_32_bits(vma_res->start) : ~0u); in print_request_ring()
/drivers/gpu/drm/i915/display/
Dintel_dpt.c54 struct i915_vma_resource *vma_res, in dpt_insert_entries() argument
70 i = vma_res->start / I915_GTT_PAGE_SIZE; in dpt_insert_entries()
71 for_each_sgt_daddr(addr, sgt_iter, vma_res->bi.pages) in dpt_insert_entries()
82 struct i915_vma_resource *vma_res, in dpt_bind_vma() argument
88 if (vma_res->bound_flags) in dpt_bind_vma()
93 if (vm->has_read_only && vma_res->bi.readonly) in dpt_bind_vma()
95 if (vma_res->bi.lmem) in dpt_bind_vma()
98 vm->insert_entries(vm, vma_res, cache_level, pte_flags); in dpt_bind_vma()
100 vma_res->page_sizes_gtt = I915_GTT_PAGE_SIZE; in dpt_bind_vma()
107 vma_res->bound_flags = I915_VMA_GLOBAL_BIND | I915_VMA_LOCAL_BIND; in dpt_bind_vma()
[all …]
/drivers/gpu/drm/i915/selftests/
Dmock_gtt.c36 struct i915_vma_resource *vma_res, in mock_insert_entries() argument
43 struct i915_vma_resource *vma_res, in mock_bind_ppgtt() argument
48 vma_res->bound_flags |= flags; in mock_bind_ppgtt()
52 struct i915_vma_resource *vma_res) in mock_unbind_ppgtt() argument
96 struct i915_vma_resource *vma_res, in mock_bind_ggtt() argument
103 struct i915_vma_resource *vma_res) in mock_unbind_ggtt() argument
Di915_gem_gtt.c1505 struct i915_vma_resource *vma_res; in reserve_gtt_with_resource() local
1509 vma_res = i915_vma_resource_alloc(); in reserve_gtt_with_resource()
1510 if (IS_ERR(vma_res)) in reserve_gtt_with_resource()
1511 return PTR_ERR(vma_res); in reserve_gtt_with_resource()
1519 i915_vma_resource_init_from_vma(vma_res, vma); in reserve_gtt_with_resource()
1520 vma->resource = vma_res; in reserve_gtt_with_resource()
1522 kfree(vma_res); in reserve_gtt_with_resource()
1686 struct i915_vma_resource *vma_res; in insert_gtt_with_resource() local
1690 vma_res = i915_vma_resource_alloc(); in insert_gtt_with_resource()
1691 if (IS_ERR(vma_res)) in insert_gtt_with_resource()
[all …]
/drivers/gpu/drm/i915/gem/
Di915_gem_execbuffer.c2015 capture->vma_res = i915_vma_resource_get(vma->resource); in eb_capture_stage()