Home
last modified time | relevance | path

Searched refs:excl (Results 1 – 20 of 20) sorted by relevance

/drivers/gpu/drm/i915/gem/
Di915_gem_wait.c40 struct dma_fence *excl; in i915_gem_object_wait_reservation() local
48 ret = dma_resv_get_fences(resv, &excl, &count, &shared); in i915_gem_object_wait_reservation()
76 excl = dma_resv_get_excl_unlocked(resv); in i915_gem_object_wait_reservation()
79 if (excl && timeout >= 0) in i915_gem_object_wait_reservation()
80 timeout = i915_gem_object_wait_fence(excl, flags, timeout); in i915_gem_object_wait_reservation()
82 dma_fence_put(excl); in i915_gem_object_wait_reservation()
154 struct dma_fence *excl; in i915_gem_object_wait_priority() local
161 ret = dma_resv_get_fences(obj->base.resv, &excl, &count, in i915_gem_object_wait_priority()
173 excl = dma_resv_get_excl_unlocked(obj->base.resv); in i915_gem_object_wait_priority()
176 if (excl) { in i915_gem_object_wait_priority()
[all …]
/drivers/gpu/drm/i915/
Di915_sw_fence.c575 struct dma_fence *excl; in i915_sw_fence_await_reservation() local
585 ret = dma_resv_get_fences(resv, &excl, &count, &shared); in i915_sw_fence_await_reservation()
609 excl = dma_resv_get_excl_unlocked(resv); in i915_sw_fence_await_reservation()
612 if (ret >= 0 && excl && excl->ops != exclude) { in i915_sw_fence_await_reservation()
614 excl, in i915_sw_fence_await_reservation()
623 dma_fence_put(excl); in i915_sw_fence_await_reservation()
Di915_active.h184 return rcu_access_pointer(ref->excl.fence); in i915_active_has_exclusive()
241 fence = i915_active_fence_get(&active->excl); in __i915_request_await_exclusive()
Di915_active.c138 GEM_BUG_ON(rcu_access_pointer(ref->excl.fence)); in __active_retire()
229 active_retire(container_of(cb, struct i915_active, excl.cb)); in excl_retire()
358 __i915_active_fence_init(&ref->excl, NULL, excl_retire); in __i915_active_init()
507 return __i915_active_set_fence(ref, &ref->excl, f); in i915_active_set_exclusive()
610 enable_signaling(&ref->excl); in flush_lazy_signals()
723 rcu_access_pointer(ref->excl.fence)) { in await_active()
724 err = __await_active(&ref->excl, fn, arg); in await_active()
Di915_active_types.h36 struct i915_active_fence excl; member
Di915_request.c1505 struct dma_fence *excl; in i915_request_await_object() local
1512 ret = dma_resv_get_fences(obj->base.resv, &excl, &count, in i915_request_await_object()
1529 excl = dma_resv_get_excl_unlocked(obj->base.resv); in i915_request_await_object()
1532 if (excl) { in i915_request_await_object()
1534 ret = i915_request_await_dma_fence(to, excl); in i915_request_await_object()
1536 dma_fence_put(excl); in i915_request_await_object()
Di915_vma.c345 if (rcu_access_pointer(vma->active.excl.fence)) { in i915_vma_wait_for_bind()
349 fence = dma_fence_get_rcu_safe(&vma->active.excl.fence); in i915_vma_wait_for_bind()
/drivers/gpu/drm/etnaviv/
Detnaviv_sched.c42 if (bo->excl) { in etnaviv_sched_dependency()
43 fence = bo->excl; in etnaviv_sched_dependency()
44 bo->excl = NULL; in etnaviv_sched_dependency()
Detnaviv_gem.h83 struct dma_fence *excl; member
Detnaviv_gem_submit.c192 ret = dma_resv_get_fences(robj, &bo->excl, in submit_fence_sync()
198 bo->excl = dma_resv_get_excl_unlocked(robj); in submit_fence_sync()
/drivers/tty/
Dvcc.c157 static struct vcc_port *vcc_get(unsigned long index, bool excl) in vcc_get() argument
171 if (!excl) { in vcc_get()
207 static void vcc_put(struct vcc_port *port, bool excl) in vcc_put() argument
217 if (WARN_ON((excl && !port->excl_locked) || in vcc_put()
218 (!excl && port->excl_locked))) in vcc_put()
223 if (excl) in vcc_put()
Dtty_io.c2739 int excl = test_bit(TTY_EXCLUSIVE, &tty->flags); in tty_ioctl() local
2741 return put_user(excl, (int __user *)p); in tty_ioctl()
/drivers/dma-buf/
Ddma-resv.c117 struct dma_fence *excl; in dma_resv_fini() local
123 excl = rcu_dereference_protected(obj->fence_excl, 1); in dma_resv_fini()
124 if (excl) in dma_resv_fini()
125 dma_fence_put(excl); in dma_resv_fini()
/drivers/fpga/
Ddfl.h291 bool excl) in dfl_feature_dev_use_begin() argument
296 if (excl) { in dfl_feature_dev_use_begin()
/drivers/gpu/drm/i915/selftests/
Di915_active.c338 active_flush(ref, &ref->excl); in i915_active_unlock_wait()
/drivers/gpu/drm/amd/amdgpu/
Damdgpu_display.c86 if (amdgpu_display_flip_handle_fence(work, &work->excl)) in amdgpu_display_flip_work_func()
206 r = dma_resv_get_fences(new_abo->tbo.base.resv, &work->excl, in amdgpu_display_crtc_page_flip_target()
256 dma_fence_put(work->excl); in amdgpu_display_crtc_page_flip_target()
Damdgpu_vm.c2098 struct dma_fence *excl, **shared; in amdgpu_vm_prt_fini() local
2102 r = dma_resv_get_fences(resv, &excl, &shared_count, &shared); in amdgpu_vm_prt_fini()
2114 amdgpu_vm_add_prt_cb(adev, excl); in amdgpu_vm_prt_fini()
Damdgpu.h460 struct dma_fence *excl; member
/drivers/scsi/sym53c8xx_2/
Dsym_glue.c59 module_param_named(excl, excl_string, charp, 0);
73 MODULE_PARM_DESC(excl, "List ioport addresses here to prevent controllers from being attached");
/drivers/pci/
Dpci.c3976 const char *res_name, int excl) in __pci_request_selected_regions() argument
3982 if (__pci_request_region(pdev, i, res_name, excl)) in __pci_request_selected_regions()