/drivers/gpu/drm/i915/gem/ |
D | i915_gem_wait.c | 40 struct dma_fence *excl; in i915_gem_object_wait_reservation() local 48 ret = dma_resv_get_fences(resv, &excl, &count, &shared); in i915_gem_object_wait_reservation() 76 excl = dma_resv_get_excl_unlocked(resv); in i915_gem_object_wait_reservation() 79 if (excl && timeout >= 0) in i915_gem_object_wait_reservation() 80 timeout = i915_gem_object_wait_fence(excl, flags, timeout); in i915_gem_object_wait_reservation() 82 dma_fence_put(excl); in i915_gem_object_wait_reservation() 154 struct dma_fence *excl; in i915_gem_object_wait_priority() local 161 ret = dma_resv_get_fences(obj->base.resv, &excl, &count, in i915_gem_object_wait_priority() 173 excl = dma_resv_get_excl_unlocked(obj->base.resv); in i915_gem_object_wait_priority() 176 if (excl) { in i915_gem_object_wait_priority() [all …]
|
/drivers/gpu/drm/i915/ |
D | i915_sw_fence.c | 575 struct dma_fence *excl; in i915_sw_fence_await_reservation() local 585 ret = dma_resv_get_fences(resv, &excl, &count, &shared); in i915_sw_fence_await_reservation() 609 excl = dma_resv_get_excl_unlocked(resv); in i915_sw_fence_await_reservation() 612 if (ret >= 0 && excl && excl->ops != exclude) { in i915_sw_fence_await_reservation() 614 excl, in i915_sw_fence_await_reservation() 623 dma_fence_put(excl); in i915_sw_fence_await_reservation()
|
D | i915_active.h | 184 return rcu_access_pointer(ref->excl.fence); in i915_active_has_exclusive() 241 fence = i915_active_fence_get(&active->excl); in __i915_request_await_exclusive()
|
D | i915_active.c | 138 GEM_BUG_ON(rcu_access_pointer(ref->excl.fence)); in __active_retire() 229 active_retire(container_of(cb, struct i915_active, excl.cb)); in excl_retire() 358 __i915_active_fence_init(&ref->excl, NULL, excl_retire); in __i915_active_init() 507 return __i915_active_set_fence(ref, &ref->excl, f); in i915_active_set_exclusive() 610 enable_signaling(&ref->excl); in flush_lazy_signals() 723 rcu_access_pointer(ref->excl.fence)) { in await_active() 724 err = __await_active(&ref->excl, fn, arg); in await_active()
|
D | i915_active_types.h | 36 struct i915_active_fence excl; member
|
D | i915_request.c | 1505 struct dma_fence *excl; in i915_request_await_object() local 1512 ret = dma_resv_get_fences(obj->base.resv, &excl, &count, in i915_request_await_object() 1529 excl = dma_resv_get_excl_unlocked(obj->base.resv); in i915_request_await_object() 1532 if (excl) { in i915_request_await_object() 1534 ret = i915_request_await_dma_fence(to, excl); in i915_request_await_object() 1536 dma_fence_put(excl); in i915_request_await_object()
|
D | i915_vma.c | 345 if (rcu_access_pointer(vma->active.excl.fence)) { in i915_vma_wait_for_bind() 349 fence = dma_fence_get_rcu_safe(&vma->active.excl.fence); in i915_vma_wait_for_bind()
|
/drivers/gpu/drm/etnaviv/ |
D | etnaviv_sched.c | 42 if (bo->excl) { in etnaviv_sched_dependency() 43 fence = bo->excl; in etnaviv_sched_dependency() 44 bo->excl = NULL; in etnaviv_sched_dependency()
|
D | etnaviv_gem.h | 83 struct dma_fence *excl; member
|
D | etnaviv_gem_submit.c | 192 ret = dma_resv_get_fences(robj, &bo->excl, in submit_fence_sync() 198 bo->excl = dma_resv_get_excl_unlocked(robj); in submit_fence_sync()
|
/drivers/tty/ |
D | vcc.c | 157 static struct vcc_port *vcc_get(unsigned long index, bool excl) in vcc_get() argument 171 if (!excl) { in vcc_get() 207 static void vcc_put(struct vcc_port *port, bool excl) in vcc_put() argument 217 if (WARN_ON((excl && !port->excl_locked) || in vcc_put() 218 (!excl && port->excl_locked))) in vcc_put() 223 if (excl) in vcc_put()
|
D | tty_io.c | 2739 int excl = test_bit(TTY_EXCLUSIVE, &tty->flags); in tty_ioctl() local 2741 return put_user(excl, (int __user *)p); in tty_ioctl()
|
/drivers/dma-buf/ |
D | dma-resv.c | 117 struct dma_fence *excl; in dma_resv_fini() local 123 excl = rcu_dereference_protected(obj->fence_excl, 1); in dma_resv_fini() 124 if (excl) in dma_resv_fini() 125 dma_fence_put(excl); in dma_resv_fini()
|
/drivers/fpga/ |
D | dfl.h | 291 bool excl) in dfl_feature_dev_use_begin() argument 296 if (excl) { in dfl_feature_dev_use_begin()
|
/drivers/gpu/drm/i915/selftests/ |
D | i915_active.c | 338 active_flush(ref, &ref->excl); in i915_active_unlock_wait()
|
/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_display.c | 86 if (amdgpu_display_flip_handle_fence(work, &work->excl)) in amdgpu_display_flip_work_func() 206 r = dma_resv_get_fences(new_abo->tbo.base.resv, &work->excl, in amdgpu_display_crtc_page_flip_target() 256 dma_fence_put(work->excl); in amdgpu_display_crtc_page_flip_target()
|
D | amdgpu_vm.c | 2098 struct dma_fence *excl, **shared; in amdgpu_vm_prt_fini() local 2102 r = dma_resv_get_fences(resv, &excl, &shared_count, &shared); in amdgpu_vm_prt_fini() 2114 amdgpu_vm_add_prt_cb(adev, excl); in amdgpu_vm_prt_fini()
|
D | amdgpu.h | 460 struct dma_fence *excl; member
|
/drivers/scsi/sym53c8xx_2/ |
D | sym_glue.c | 59 module_param_named(excl, excl_string, charp, 0); 73 MODULE_PARM_DESC(excl, "List ioport addresses here to prevent controllers from being attached");
|
/drivers/pci/ |
D | pci.c | 3976 const char *res_name, int excl) in __pci_request_selected_regions() argument 3982 if (__pci_request_region(pdev, i, res_name, excl)) in __pci_request_selected_regions()
|