Home
last modified time | relevance | path

Searched refs:bos (Results 1 – 25 of 38) sorted by relevance

12

/drivers/gpu/drm/etnaviv/
Detnaviv_gem_submit.c35 size_t sz = size_vstruct(nr_bos, sizeof(submit->bos[0]), sizeof(*submit)); in submit_create()
74 submit->bos[i].flags = bo->flags; in submit_lookup_objects()
81 submit->bos[i].va = bo->presumed; in submit_lookup_objects()
101 submit->bos[i].obj = to_etnaviv_bo(obj); in submit_lookup_objects()
113 if (submit->bos[i].flags & BO_LOCKED) { in submit_unlock_object()
114 struct drm_gem_object *obj = &submit->bos[i].obj->base; in submit_unlock_object()
117 submit->bos[i].flags &= ~BO_LOCKED; in submit_unlock_object()
128 struct drm_gem_object *obj = &submit->bos[i].obj->base; in submit_lock_objects()
135 if (!(submit->bos[i].flags & BO_LOCKED)) { in submit_lock_objects()
142 submit->bos[i].flags |= BO_LOCKED; in submit_lock_objects()
[all …]
Detnaviv_dump.c141 obj = submit->bos[i].obj; in etnaviv_core_dump()
201 obj = submit->bos[i].obj; in etnaviv_core_dump()
202 vram = submit->bos[i].mapping; in etnaviv_core_dump()
/drivers/gpu/drm/msm/
Dmsm_gem_submit.c39 sz = struct_size(submit, bos, nr_bos) + in submit_create()
61 submit->cmd = (void *)&submit->bos[nr_bos]; in submit_create()
112 u64_to_user_ptr(args->bos + (i * sizeof(submit_bo))); in submit_lookup_objects()
117 submit->bos[i].flags = 0; in submit_lookup_objects()
136 submit->bos[i].handle = submit_bo.handle; in submit_lookup_objects()
137 submit->bos[i].flags = submit_bo.flags; in submit_lookup_objects()
139 submit->bos[i].iova = submit_bo.presumed; in submit_lookup_objects()
150 obj = idr_find(&file->object_idr, submit->bos[i].handle); in submit_lookup_objects()
152 DRM_ERROR("invalid handle %u at index %u\n", submit->bos[i].handle, i); in submit_lookup_objects()
159 submit->bos[i].obj = to_msm_bo(obj); in submit_lookup_objects()
[all …]
Dmsm_fb.c28 const struct drm_mode_fb_cmd2 *mode_cmd, struct drm_gem_object **bos);
132 struct drm_gem_object *bos[4] = {0}; in msm_framebuffer_create() local
137 bos[i] = drm_gem_object_lookup(file, mode_cmd->handles[i]); in msm_framebuffer_create()
138 if (!bos[i]) { in msm_framebuffer_create()
144 fb = msm_framebuffer_init(dev, mode_cmd, bos); in msm_framebuffer_create()
154 drm_gem_object_put(bos[i]); in msm_framebuffer_create()
159 const struct drm_mode_fb_cmd2 *mode_cmd, struct drm_gem_object **bos) in msm_framebuffer_init() argument
208 if (bos[i]->size < min_size) { in msm_framebuffer_init()
213 msm_fb->base.obj[i] = bos[i]; in msm_framebuffer_init()
Dmsm_gpu.c232 struct msm_gpu_state_bo *state_bo = &state->bos[state->nr_bos]; in msm_gpu_crashstate_get_bo()
296 state->bos = kcalloc(nr, in msm_gpu_crashstate_capture()
299 for (i = 0; state->bos && i < submit->nr_bos; i++) { in msm_gpu_crashstate_capture()
301 msm_gpu_crashstate_get_bo(state, submit->bos[i].obj, in msm_gpu_crashstate_capture()
302 submit->bos[i].iova, submit->bos[i].flags); in msm_gpu_crashstate_capture()
306 for (i = 0; state->bos && i < submit->nr_cmds; i++) { in msm_gpu_crashstate_capture()
310 msm_gpu_crashstate_get_bo(state, submit->bos[idx].obj, in msm_gpu_crashstate_capture()
311 submit->bos[idx].iova, submit->bos[idx].flags); in msm_gpu_crashstate_capture()
Dmsm_rd.c309 struct msm_gem_object *obj = submit->bos[idx].obj; in snapshot_buf()
314 offset = iova - submit->bos[idx].iova; in snapshot_buf()
316 iova = submit->bos[idx].iova; in snapshot_buf()
331 if (!(submit->bos[idx].flags & MSM_SUBMIT_BO_READ)) in snapshot_buf()
Dmsm_gem.h354 } bos[]; member
383 return rd_full || (submit->bos[idx].flags & MSM_SUBMIT_BO_DUMP); in should_dump()
/drivers/gpu/drm/qxl/
Dqxl_release.c100 INIT_LIST_HEAD(&release->bos); in qxl_release_alloc()
122 while (!list_empty(&release->bos)) { in qxl_release_free_list()
126 entry = container_of(release->bos.next, in qxl_release_free_list()
150 WARN_ON(list_empty(&release->bos)); in qxl_release_free()
175 list_for_each_entry(entry, &release->bos, tv.head) { in qxl_release_list_add()
187 list_add_tail(&entry->tv.head, &release->bos); in qxl_release_list_add()
221 if (list_is_singular(&release->bos)) in qxl_release_reserve_list()
224 ret = ttm_eu_reserve_buffers(&release->ticket, &release->bos, in qxl_release_reserve_list()
229 list_for_each_entry(entry, &release->bos, tv.head) { in qxl_release_reserve_list()
234 ttm_eu_backoff_reservation(&release->ticket, &release->bos); in qxl_release_reserve_list()
[all …]
/drivers/gpu/drm/lima/
Dlima_gem.c306 struct lima_bo **bos = submit->lbos; in lima_gem_submit() local
318 obj = drm_gem_object_lookup(file, submit->bos[i].handle); in lima_gem_submit()
335 bos[i] = bo; in lima_gem_submit()
338 err = drm_gem_lock_reservations((struct drm_gem_object **)bos, in lima_gem_submit()
345 bos, submit->nr_bos, vm); in lima_gem_submit()
355 submit->task, bos[i], in lima_gem_submit()
356 submit->bos[i].flags & LIMA_SUBMIT_BO_WRITE, in lima_gem_submit()
366 if (submit->bos[i].flags & LIMA_SUBMIT_BO_WRITE) in lima_gem_submit()
367 dma_resv_add_excl_fence(lima_bo_resv(bos[i]), fence); in lima_gem_submit()
369 dma_resv_add_shared_fence(lima_bo_resv(bos[i]), fence); in lima_gem_submit()
[all …]
Dlima_sched.c114 struct lima_bo **bos, int num_bos, in lima_sched_task_init() argument
119 task->bos = kmemdup(bos, sizeof(*bos) * num_bos, GFP_KERNEL); in lima_sched_task_init()
120 if (!task->bos) in lima_sched_task_init()
124 drm_gem_object_get(&bos[i]->base.base); in lima_sched_task_init()
128 kfree(task->bos); in lima_sched_task_init()
153 if (task->bos) { in lima_sched_task_fini()
155 drm_gem_object_put(&task->bos[i]->base.base); in lima_sched_task_fini()
156 kfree(task->bos); in lima_sched_task_fini()
325 struct lima_bo *bo = task->bos[i]; in lima_sched_build_error_task_list()
370 struct lima_bo *bo = task->bos[i]; in lima_sched_build_error_task_list()
[all …]
Dlima_drv.c106 struct drm_lima_gem_submit_bo *bos; in lima_ioctl_gem_submit() local
124 bos = kvcalloc(args->nr_bos, sizeof(*submit.bos) + sizeof(*submit.lbos), GFP_KERNEL); in lima_ioctl_gem_submit()
125 if (!bos) in lima_ioctl_gem_submit()
128 size = args->nr_bos * sizeof(*submit.bos); in lima_ioctl_gem_submit()
129 if (copy_from_user(bos, u64_to_user_ptr(args->bos), size)) { in lima_ioctl_gem_submit()
157 submit.bos = bos; in lima_ioctl_gem_submit()
158 submit.lbos = (void *)bos + size; in lima_ioctl_gem_submit()
174 kvfree(bos); in lima_ioctl_gem_submit()
Dlima_sched.h29 struct lima_bo **bos; member
92 struct lima_bo **bos, int num_bos,
Dlima_drv.h32 struct drm_lima_gem_submit_bo *bos; member
/drivers/usb/core/
Dconfig.c961 if (dev->bos) { in usb_release_bos_descriptor()
962 kfree(dev->bos->desc); in usb_release_bos_descriptor()
963 kfree(dev->bos); in usb_release_bos_descriptor()
964 dev->bos = NULL; in usb_release_bos_descriptor()
981 struct usb_bos_descriptor *bos; in usb_get_bos_descriptor() local
989 bos = kzalloc(sizeof(struct usb_bos_descriptor), GFP_KERNEL); in usb_get_bos_descriptor()
990 if (!bos) in usb_get_bos_descriptor()
994 ret = usb_get_descriptor(dev, USB_DT_BOS, 0, bos, USB_DT_BOS_SIZE); in usb_get_bos_descriptor()
995 if (ret < USB_DT_BOS_SIZE || bos->bLength < USB_DT_BOS_SIZE) { in usb_get_bos_descriptor()
999 kfree(bos); in usb_get_bos_descriptor()
[all …]
Dhub.h152 hdev->bos && hdev->bos->ssp_cap); in hub_is_superspeedplus()
Dhub.c155 if (!udev->bos) in usb_device_supports_lpm()
162 if (udev->bos->ext_cap && in usb_device_supports_lpm()
164 le32_to_cpu(udev->bos->ext_cap->bmAttributes))) in usb_device_supports_lpm()
174 if (!udev->bos->ss_cap) { in usb_device_supports_lpm()
179 if (udev->bos->ss_cap->bU1devExitLat == 0 && in usb_device_supports_lpm()
180 udev->bos->ss_cap->bU2DevExitLat == 0) { in usb_device_supports_lpm()
335 if (!udev->bos) in usb_set_lpm_parameters()
345 udev_u1_del = udev->bos->ss_cap->bU1devExitLat; in usb_set_lpm_parameters()
346 udev_u2_del = le16_to_cpu(udev->bos->ss_cap->bU2DevExitLat); in usb_set_lpm_parameters()
347 hub_u1_del = udev->parent->bos->ss_cap->bU1devExitLat; in usb_set_lpm_parameters()
[all …]
/drivers/gpu/drm/omapdrm/
Domap_fb.c309 struct drm_gem_object *bos[4]; in omap_framebuffer_create() local
314 bos[i] = drm_gem_object_lookup(file, mode_cmd->handles[i]); in omap_framebuffer_create()
315 if (!bos[i]) { in omap_framebuffer_create()
321 fb = omap_framebuffer_init(dev, mode_cmd, bos); in omap_framebuffer_create()
329 drm_gem_object_put(bos[i]); in omap_framebuffer_create()
335 const struct drm_mode_fb_cmd2 *mode_cmd, struct drm_gem_object **bos) in omap_framebuffer_init() argument
397 if (size > omap_gem_mmap_size(bos[i]) - mode_cmd->offsets[i]) { in omap_framebuffer_init()
400 bos[i]->size - mode_cmd->offsets[i], size); in omap_framebuffer_init()
405 fb->obj[i] = bos[i]; in omap_framebuffer_init()
Domap_fb.h25 const struct drm_mode_fb_cmd2 *mode_cmd, struct drm_gem_object **bos);
/drivers/gpu/drm/panfrost/
Dpanfrost_job.c243 static int panfrost_acquire_object_fences(struct drm_gem_object **bos, in panfrost_acquire_object_fences() argument
251 ret = drm_gem_fence_array_add_implicit(deps, bos[i], true); in panfrost_acquire_object_fences()
259 static void panfrost_attach_object_fences(struct drm_gem_object **bos, in panfrost_attach_object_fences() argument
266 dma_resv_add_excl_fence(bos[i]->resv, fence); in panfrost_attach_object_fences()
278 ret = drm_gem_lock_reservations(job->bos, job->bo_count, in panfrost_job_push()
293 ret = panfrost_acquire_object_fences(job->bos, job->bo_count, in panfrost_job_push()
306 panfrost_attach_object_fences(job->bos, job->bo_count, in panfrost_job_push()
310 drm_gem_unlock_reservations(job->bos, job->bo_count, &acquire_ctx); in panfrost_job_push()
342 if (job->bos) { in panfrost_job_cleanup()
344 drm_gem_object_put(job->bos[i]); in panfrost_job_cleanup()
[all …]
Dpanfrost_job.h34 struct drm_gem_object **bos; member
/drivers/gpu/drm/msm/adreno/
Dadreno_gpu.c574 for (i = 0; state->bos && i < state->nr_bos; i++) in adreno_gpu_state_destroy()
575 kvfree(state->bos[i].data); in adreno_gpu_state_destroy()
577 kfree(state->bos); in adreno_gpu_state_destroy()
720 if (state->bos) { in adreno_show()
725 state->bos[i].iova); in adreno_show()
726 drm_printf(p, " size: %zd\n", state->bos[i].size); in adreno_show()
728 adreno_show_object(p, &state->bos[i].data, in adreno_show()
729 state->bos[i].size, &state->bos[i].encoded); in adreno_show()
/drivers/usb/gadget/
Dcomposite.c693 struct usb_bos_descriptor *bos = cdev->req->buf; in bos_desc() local
696 bos->bLength = USB_DT_BOS_SIZE; in bos_desc()
697 bos->bDescriptorType = USB_DT_BOS; in bos_desc()
699 bos->wTotalLength = cpu_to_le16(USB_DT_BOS_SIZE); in bos_desc()
700 bos->bNumDeviceCaps = 0; in bos_desc()
729 usb_ext = cdev->req->buf + le16_to_cpu(bos->wTotalLength); in bos_desc()
730 bos->bNumDeviceCaps++; in bos_desc()
731 le16_add_cpu(&bos->wTotalLength, USB_DT_USB_EXT_CAP_SIZE); in bos_desc()
745 ss_cap = cdev->req->buf + le16_to_cpu(bos->wTotalLength); in bos_desc()
746 bos->bNumDeviceCaps++; in bos_desc()
[all …]
/drivers/usb/host/
Dxhci-hub.c38 struct usb_bos_descriptor *bos; in xhci_create_usb3x_bos_desc() local
52 bos = (struct usb_bos_descriptor *)buf; in xhci_create_usb3x_bos_desc()
53 bos->bLength = USB_DT_BOS_SIZE; in xhci_create_usb3x_bos_desc()
54 bos->bDescriptorType = USB_DT_BOS; in xhci_create_usb3x_bos_desc()
55 bos->wTotalLength = cpu_to_le16(USB_DT_BOS_SIZE + in xhci_create_usb3x_bos_desc()
57 bos->bNumDeviceCaps = 1; in xhci_create_usb3x_bos_desc()
91 bos->bNumDeviceCaps++; in xhci_create_usb3x_bos_desc()
92 bos->wTotalLength = cpu_to_le16(USB_DT_BOS_SIZE + in xhci_create_usb3x_bos_desc()
121 if (wLength < le16_to_cpu(bos->wTotalLength)) in xhci_create_usb3x_bos_desc()
125 return le16_to_cpu(bos->wTotalLength); in xhci_create_usb3x_bos_desc()
[all …]
/drivers/gpu/drm/msm/disp/dpu1/
Ddpu_formats.c941 struct drm_gem_object **bos) in dpu_format_check_modified_format() argument
949 if (!msm_fmt || !cmd || !bos) { in dpu_format_check_modified_format()
965 if (!bos[i]) { in dpu_format_check_modified_format()
969 if ((i == 0) || (bos[i] != bos[0])) in dpu_format_check_modified_format()
970 bos_total_size += bos[i]->size; in dpu_format_check_modified_format()
Ddpu_formats.h49 struct drm_gem_object **bos);

12