/drivers/gpu/drm/i915/ |
D | i915_deps.c | 36 static void i915_deps_reset_fences(struct i915_deps *deps) in i915_deps_reset_fences() argument 38 if (deps->fences != &deps->single) in i915_deps_reset_fences() 39 kfree(deps->fences); in i915_deps_reset_fences() 40 deps->num_deps = 0; in i915_deps_reset_fences() 41 deps->fences_size = 1; in i915_deps_reset_fences() 42 deps->fences = &deps->single; in i915_deps_reset_fences() 50 void i915_deps_init(struct i915_deps *deps, gfp_t gfp) in i915_deps_init() argument 52 deps->fences = NULL; in i915_deps_init() 53 deps->gfp = gfp; in i915_deps_init() 54 i915_deps_reset_fences(deps); in i915_deps_init() [all …]
|
D | i915_deps.h | 32 void i915_deps_init(struct i915_deps *deps, gfp_t gfp); 34 void i915_deps_fini(struct i915_deps *deps); 36 int i915_deps_add_dependency(struct i915_deps *deps, 40 int i915_deps_add_resv(struct i915_deps *deps, struct dma_resv *resv, 43 int i915_deps_sync(const struct i915_deps *deps,
|
D | i915_request.h | 422 int i915_request_await_deps(struct i915_request *rq, const struct i915_deps *deps);
|
D | i915_request.c | 1569 int i915_request_await_deps(struct i915_request *rq, const struct i915_deps *deps) in i915_request_await_deps() argument 1573 for (i = 0; i < deps->num_deps; ++i) { in i915_request_await_deps() 1574 err = i915_request_await_dma_fence(rq, deps->fences[i]); in i915_request_await_deps()
|
/drivers/gpu/drm/i915/gt/ |
D | intel_migrate.h | 26 const struct i915_deps *deps, 36 const struct i915_deps *deps, 48 const struct i915_deps *deps, 56 const struct i915_deps *deps,
|
D | intel_migrate.c | 679 const struct i915_deps *deps, in intel_context_migrate_copy() argument 759 if (deps) { in intel_context_migrate_copy() 760 err = i915_request_await_deps(rq, deps); in intel_context_migrate_copy() 770 deps = NULL; in intel_context_migrate_copy() 985 const struct i915_deps *deps, in intel_context_migrate_clear() argument 1016 if (deps) { in intel_context_migrate_clear() 1017 err = i915_request_await_deps(rq, deps); in intel_context_migrate_clear() 1027 deps = NULL; in intel_context_migrate_clear() 1080 const struct i915_deps *deps, in intel_migrate_copy() argument 1105 err = intel_context_migrate_copy(ce, deps, in intel_migrate_copy() [all …]
|
D | selftest_migrate.c | 138 const struct i915_deps *deps, in intel_context_copy_ccs() argument 169 if (deps) { in intel_context_copy_ccs() 170 err = i915_request_await_deps(rq, deps); in intel_context_copy_ccs() 180 deps = NULL; in intel_context_copy_ccs() 224 const struct i915_deps *deps, in intel_migrate_ccs_copy() argument 246 err = intel_context_copy_ccs(ce, deps, sg, pat_index, in intel_migrate_ccs_copy()
|
/drivers/gpu/drm/i915/gem/ |
D | i915_gem_ttm_move.c | 192 const struct i915_deps *deps) in i915_ttm_accel_move() argument 216 ret = intel_context_migrate_clear(to_gt(i915)->migrate.context, deps, in i915_ttm_accel_move() 231 deps, src_rsgt->table.sgl, in i915_ttm_accel_move() 639 struct i915_deps deps; in i915_ttm_move() local 641 i915_deps_init(&deps, GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN); in i915_ttm_move() 642 ret = i915_deps_add_resv(&deps, bo->base.resv, ctx); in i915_ttm_move() 649 dst_rsgt, true, &deps); in i915_ttm_move() 650 i915_deps_fini(&deps); in i915_ttm_move() 715 struct i915_deps deps; in i915_gem_obj_copy_ttm() local 724 i915_deps_init(&deps, GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN); in i915_gem_obj_copy_ttm() [all …]
|
/drivers/gpu/drm/i915/gem/selftests/ |
D | i915_gem_migrate.c | 184 struct i915_deps *deps, in __igt_lmem_pages_migrate() argument 221 err = intel_migrate_clear(>->migrate, &ww, deps, in __igt_lmem_pages_migrate() 395 struct i915_deps deps; in igt_async_migrate() local 418 i915_deps_init(&deps, GFP_KERNEL); in igt_async_migrate() 419 err = i915_deps_add_dependency(&deps, &rq->fence, &ctx); in igt_async_migrate() 425 err = __igt_lmem_pages_migrate(gt, &ppgtt->vm, &deps, &spin, in igt_async_migrate() 427 i915_deps_fini(&deps); in igt_async_migrate()
|
/drivers/net/ethernet/ibm/emac/ |
D | core.c | 2358 struct emac_depentry *deps) in emac_check_deps() argument 2365 if (deps[i].phandle == 0) { in emac_check_deps() 2373 deps[i].phandle = 0; in emac_check_deps() 2377 if (deps[i].node == NULL) in emac_check_deps() 2378 deps[i].node = of_node_get(np); in emac_check_deps() 2380 if (deps[i].node == NULL) in emac_check_deps() 2381 deps[i].node = of_find_node_by_phandle(deps[i].phandle); in emac_check_deps() 2382 if (deps[i].node == NULL) in emac_check_deps() 2384 if (deps[i].ofdev == NULL) in emac_check_deps() 2385 deps[i].ofdev = of_find_device_by_node(deps[i].node); in emac_check_deps() [all …]
|
/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_cs.c | 384 struct drm_amdgpu_cs_chunk_dep *deps = chunk->kdata; in amdgpu_cs_p2_dependencies() local 397 ctx = amdgpu_ctx_get(fpriv, deps[i].ctx_id); in amdgpu_cs_p2_dependencies() 401 r = amdgpu_ctx_get_entity(ctx, deps[i].ip_type, in amdgpu_cs_p2_dependencies() 402 deps[i].ip_instance, in amdgpu_cs_p2_dependencies() 403 deps[i].ring, &entity); in amdgpu_cs_p2_dependencies() 409 fence = amdgpu_ctx_get_fence(ctx, entity, deps[i].handle); in amdgpu_cs_p2_dependencies() 456 struct drm_amdgpu_cs_chunk_sem *deps = chunk->kdata; in amdgpu_cs_p2_syncobj_in() local 463 r = amdgpu_syncobj_lookup_and_add(p, deps[i].handle, 0, 0); in amdgpu_cs_p2_syncobj_in() 494 struct drm_amdgpu_cs_chunk_sem *deps = chunk->kdata; in amdgpu_cs_p2_syncobj_out() local 514 drm_syncobj_find(p->filp, deps[i].handle); in amdgpu_cs_p2_syncobj_out()
|
/drivers/firmware/efi/libstub/ |
D | Makefile | 76 libfdt-deps := fdt_rw.c fdt_ro.c fdt_wip.c fdt.c \ 80 $(patsubst %.c,lib-%.o,$(libfdt-deps))
|
/drivers/md/ |
D | dm-ioctl.c | 1649 struct dm_target_deps *deps; in retrieve_deps() local 1653 deps = get_result_buffer(param, param_size, &len); in retrieve_deps() 1664 needed = struct_size(deps, dev, count); in retrieve_deps() 1673 deps->count = count; in retrieve_deps() 1676 deps->dev[count++] = huge_encode_dev(dd->dm_dev->bdev->bd_dev); in retrieve_deps()
|