Home
last modified time | relevance | path

Searched refs:wa_ctx (Results 1 – 9 of 9) sorted by relevance

/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gvt/
Dscheduler.c394 static void release_shadow_wa_ctx(struct intel_shadow_wa_ctx *wa_ctx) in release_shadow_wa_ctx() argument
396 if (!wa_ctx->indirect_ctx.obj) in release_shadow_wa_ctx()
399 i915_gem_object_unpin_map(wa_ctx->indirect_ctx.obj); in release_shadow_wa_ctx()
400 i915_gem_object_put(wa_ctx->indirect_ctx.obj); in release_shadow_wa_ctx()
402 wa_ctx->indirect_ctx.obj = NULL; in release_shadow_wa_ctx()
403 wa_ctx->indirect_ctx.shadow_va = NULL; in release_shadow_wa_ctx()
486 workload->wa_ctx.indirect_ctx.size) { in intel_gvt_scan_and_shadow_workload()
487 ret = intel_gvt_scan_and_shadow_wa_ctx(&workload->wa_ctx); in intel_gvt_scan_and_shadow_workload()
496 release_shadow_wa_ctx(&workload->wa_ctx); in intel_gvt_scan_and_shadow_workload()
559 static void update_wa_ctx_2_shadow_ctx(struct intel_shadow_wa_ctx *wa_ctx) in update_wa_ctx_2_shadow_ctx() argument
[all …]
Dcmd_parser.c2836 static int scan_wa_ctx(struct intel_shadow_wa_ctx *wa_ctx) in scan_wa_ctx() argument
2842 struct intel_vgpu_workload *workload = container_of(wa_ctx, in scan_wa_ctx()
2844 wa_ctx); in scan_wa_ctx()
2847 if (WARN_ON(!IS_ALIGNED(wa_ctx->indirect_ctx.guest_gma, in scan_wa_ctx()
2851 ring_tail = wa_ctx->indirect_ctx.size + 3 * sizeof(u32); in scan_wa_ctx()
2852 ring_size = round_up(wa_ctx->indirect_ctx.size + CACHELINE_BYTES, in scan_wa_ctx()
2854 gma_head = wa_ctx->indirect_ctx.guest_gma; in scan_wa_ctx()
2855 gma_tail = wa_ctx->indirect_ctx.guest_gma + ring_tail; in scan_wa_ctx()
2856 gma_bottom = wa_ctx->indirect_ctx.guest_gma + ring_size; in scan_wa_ctx()
2862 s.ring_start = wa_ctx->indirect_ctx.guest_gma; in scan_wa_ctx()
[all …]
Dcmd_parser.h51 int intel_gvt_scan_and_shadow_wa_ctx(struct intel_shadow_wa_ctx *wa_ctx);
Dscheduler.h114 struct intel_shadow_wa_ctx wa_ctx; member
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/
Dselftest_ring_submission.c223 engine->wa_ctx.vma = bb; in __live_ctx_switch_wa()
238 intel_context_put(engine->wa_ctx.vma->private); in __live_ctx_switch_wa()
239 i915_vma_unpin_and_release(&engine->wa_ctx.vma, I915_VMA_RELEASE_MAP); in __live_ctx_switch_wa()
272 saved_wa = fetch_and_zero(&engine->wa_ctx.vma); in live_ctx_switch_wa()
280 engine->wa_ctx.vma = saved_wa; in live_ctx_switch_wa()
Dintel_ring_submission.c866 engine->wa_ctx.vma->node.start, 0, in clear_residuals()
888 if (engine->wa_ctx.vma && ce != engine->kernel_context) { in switch_context()
889 if (engine->wa_ctx.vma->private != ce && in switch_context()
895 residuals = &engine->wa_ctx.vma->private; in switch_context()
1032 if (engine->wa_ctx.vma) { in ring_release()
1033 intel_context_put(engine->wa_ctx.vma->private); in ring_release()
1034 i915_vma_unpin_and_release(&engine->wa_ctx.vma, 0); in ring_release()
1228 engine->wa_ctx.vma = vma; in gen7_ctx_switch_bb_init()
Dintel_lrc.c3485 GEM_BUG_ON(engine->wa_ctx.indirect_ctx.size); in __execlists_update_reg_state()
3928 engine->wa_ctx.vma = vma; in lrc_setup_wa_ctx()
3938 i915_vma_unpin_and_release(&engine->wa_ctx.vma, 0); in lrc_destroy_wa_ctx()
3941 memset(&engine->wa_ctx, 0, sizeof(engine->wa_ctx)); in lrc_destroy_wa_ctx()
3948 struct i915_ctx_workarounds *wa_ctx = &engine->wa_ctx; in intel_init_workaround_bb() local
3949 struct i915_wa_ctx_bb *wa_bb[2] = { &wa_ctx->indirect_ctx, in intel_init_workaround_bb()
3950 &wa_ctx->per_ctx }; in intel_init_workaround_bb()
3987 batch = i915_gem_object_pin_map(wa_ctx->vma->obj, I915_MAP_WB); in intel_init_workaround_bb()
4008 __i915_gem_object_flush_map(wa_ctx->vma->obj, 0, batch_ptr - batch); in intel_init_workaround_bb()
4009 __i915_gem_object_release_map(wa_ctx->vma->obj); in intel_init_workaround_bb()
[all …]
Dintel_engine_types.h405 struct i915_ctx_workarounds wa_ctx; member
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/
Di915_gpu_error.c1427 engine->wa_ctx.vma, in intel_engine_coredump_add_vma()