1 /*
2 * Copyright © 2014 Advanced Micro Devices, Inc.
3 * All Rights Reserved.
4 *
5 * Permission is hereby granted, free of charge, to any person obtaining a
6 * copy of this software and associated documentation files (the "Software"),
7 * to deal in the Software without restriction, including without limitation
8 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
9 * and/or sell copies of the Software, and to permit persons to whom the
10 * Software is furnished to do so, subject to the following conditions:
11 *
12 * The above copyright notice and this permission notice shall be included in
13 * all copies or substantial portions of the Software.
14 *
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
18 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
19 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
20 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
21 * OTHER DEALINGS IN THE SOFTWARE.
22 *
23 */
24
25 #include <stdlib.h>
26 #include <stdio.h>
27 #include <stdint.h>
28 #include <string.h>
29 #include <errno.h>
30 #include <fcntl.h>
31 #include <unistd.h>
32 #include <sys/ioctl.h>
33 #include <sys/mman.h>
34 #include <sys/time.h>
35
36 #include "libdrm_macros.h"
37 #include "xf86drm.h"
38 #include "amdgpu_drm.h"
39 #include "amdgpu_internal.h"
40 #include "util_math.h"
41
amdgpu_bo_create(amdgpu_device_handle dev,uint64_t size,uint32_t handle,amdgpu_bo_handle * buf_handle)42 static int amdgpu_bo_create(amdgpu_device_handle dev,
43 uint64_t size,
44 uint32_t handle,
45 amdgpu_bo_handle *buf_handle)
46 {
47 struct amdgpu_bo *bo;
48 int r;
49
50 bo = calloc(1, sizeof(struct amdgpu_bo));
51 if (!bo)
52 return -ENOMEM;
53
54 r = handle_table_insert(&dev->bo_handles, handle, bo);
55 if (r) {
56 free(bo);
57 return r;
58 }
59
60 atomic_set(&bo->refcount, 1);
61 bo->dev = dev;
62 bo->alloc_size = size;
63 bo->handle = handle;
64 pthread_mutex_init(&bo->cpu_access_mutex, NULL);
65
66 *buf_handle = bo;
67 return 0;
68 }
69
amdgpu_bo_alloc(amdgpu_device_handle dev,struct amdgpu_bo_alloc_request * alloc_buffer,amdgpu_bo_handle * buf_handle)70 drm_public int amdgpu_bo_alloc(amdgpu_device_handle dev,
71 struct amdgpu_bo_alloc_request *alloc_buffer,
72 amdgpu_bo_handle *buf_handle)
73 {
74 union drm_amdgpu_gem_create args;
75 int r;
76
77 memset(&args, 0, sizeof(args));
78 args.in.bo_size = alloc_buffer->alloc_size;
79 args.in.alignment = alloc_buffer->phys_alignment;
80
81 /* Set the placement. */
82 args.in.domains = alloc_buffer->preferred_heap;
83 args.in.domain_flags = alloc_buffer->flags;
84
85 /* Allocate the buffer with the preferred heap. */
86 r = drmCommandWriteRead(dev->fd, DRM_AMDGPU_GEM_CREATE,
87 &args, sizeof(args));
88 if (r)
89 goto out;
90
91 pthread_mutex_lock(&dev->bo_table_mutex);
92 r = amdgpu_bo_create(dev, alloc_buffer->alloc_size, args.out.handle,
93 buf_handle);
94 pthread_mutex_unlock(&dev->bo_table_mutex);
95 if (r) {
96 drmCloseBufferHandle(dev->fd, args.out.handle);
97 }
98
99 out:
100 return r;
101 }
102
amdgpu_bo_set_metadata(amdgpu_bo_handle bo,struct amdgpu_bo_metadata * info)103 drm_public int amdgpu_bo_set_metadata(amdgpu_bo_handle bo,
104 struct amdgpu_bo_metadata *info)
105 {
106 struct drm_amdgpu_gem_metadata args = {};
107
108 args.handle = bo->handle;
109 args.op = AMDGPU_GEM_METADATA_OP_SET_METADATA;
110 args.data.flags = info->flags;
111 args.data.tiling_info = info->tiling_info;
112
113 if (info->size_metadata > sizeof(args.data.data))
114 return -EINVAL;
115
116 if (info->size_metadata) {
117 args.data.data_size_bytes = info->size_metadata;
118 memcpy(args.data.data, info->umd_metadata, info->size_metadata);
119 }
120
121 return drmCommandWriteRead(bo->dev->fd,
122 DRM_AMDGPU_GEM_METADATA,
123 &args, sizeof(args));
124 }
125
amdgpu_bo_query_info(amdgpu_bo_handle bo,struct amdgpu_bo_info * info)126 drm_public int amdgpu_bo_query_info(amdgpu_bo_handle bo,
127 struct amdgpu_bo_info *info)
128 {
129 struct drm_amdgpu_gem_metadata metadata = {};
130 struct drm_amdgpu_gem_create_in bo_info = {};
131 struct drm_amdgpu_gem_op gem_op = {};
132 int r;
133
134 /* Validate the BO passed in */
135 if (!bo->handle)
136 return -EINVAL;
137
138 /* Query metadata. */
139 metadata.handle = bo->handle;
140 metadata.op = AMDGPU_GEM_METADATA_OP_GET_METADATA;
141
142 r = drmCommandWriteRead(bo->dev->fd, DRM_AMDGPU_GEM_METADATA,
143 &metadata, sizeof(metadata));
144 if (r)
145 return r;
146
147 if (metadata.data.data_size_bytes >
148 sizeof(info->metadata.umd_metadata))
149 return -EINVAL;
150
151 /* Query buffer info. */
152 gem_op.handle = bo->handle;
153 gem_op.op = AMDGPU_GEM_OP_GET_GEM_CREATE_INFO;
154 gem_op.value = (uintptr_t)&bo_info;
155
156 r = drmCommandWriteRead(bo->dev->fd, DRM_AMDGPU_GEM_OP,
157 &gem_op, sizeof(gem_op));
158 if (r)
159 return r;
160
161 memset(info, 0, sizeof(*info));
162 info->alloc_size = bo_info.bo_size;
163 info->phys_alignment = bo_info.alignment;
164 info->preferred_heap = bo_info.domains;
165 info->alloc_flags = bo_info.domain_flags;
166 info->metadata.flags = metadata.data.flags;
167 info->metadata.tiling_info = metadata.data.tiling_info;
168
169 info->metadata.size_metadata = metadata.data.data_size_bytes;
170 if (metadata.data.data_size_bytes > 0)
171 memcpy(info->metadata.umd_metadata, metadata.data.data,
172 metadata.data.data_size_bytes);
173
174 return 0;
175 }
176
amdgpu_bo_export_flink(amdgpu_bo_handle bo)177 static int amdgpu_bo_export_flink(amdgpu_bo_handle bo)
178 {
179 struct drm_gem_flink flink;
180 int fd, dma_fd;
181 uint32_t handle;
182 int r;
183
184 fd = bo->dev->fd;
185 handle = bo->handle;
186 if (bo->flink_name)
187 return 0;
188
189
190 if (bo->dev->flink_fd != bo->dev->fd) {
191 r = drmPrimeHandleToFD(bo->dev->fd, bo->handle, DRM_CLOEXEC,
192 &dma_fd);
193 if (!r) {
194 r = drmPrimeFDToHandle(bo->dev->flink_fd, dma_fd, &handle);
195 close(dma_fd);
196 }
197 if (r)
198 return r;
199 fd = bo->dev->flink_fd;
200 }
201 memset(&flink, 0, sizeof(flink));
202 flink.handle = handle;
203
204 r = drmIoctl(fd, DRM_IOCTL_GEM_FLINK, &flink);
205 if (r)
206 return r;
207
208 bo->flink_name = flink.name;
209
210 if (bo->dev->flink_fd != bo->dev->fd)
211 drmCloseBufferHandle(bo->dev->flink_fd, handle);
212
213 pthread_mutex_lock(&bo->dev->bo_table_mutex);
214 r = handle_table_insert(&bo->dev->bo_flink_names, bo->flink_name, bo);
215 pthread_mutex_unlock(&bo->dev->bo_table_mutex);
216
217 return r;
218 }
219
amdgpu_bo_export(amdgpu_bo_handle bo,enum amdgpu_bo_handle_type type,uint32_t * shared_handle)220 drm_public int amdgpu_bo_export(amdgpu_bo_handle bo,
221 enum amdgpu_bo_handle_type type,
222 uint32_t *shared_handle)
223 {
224 int r;
225
226 switch (type) {
227 case amdgpu_bo_handle_type_gem_flink_name:
228 r = amdgpu_bo_export_flink(bo);
229 if (r)
230 return r;
231
232 *shared_handle = bo->flink_name;
233 return 0;
234
235 case amdgpu_bo_handle_type_kms:
236 case amdgpu_bo_handle_type_kms_noimport:
237 *shared_handle = bo->handle;
238 return 0;
239
240 case amdgpu_bo_handle_type_dma_buf_fd:
241 return drmPrimeHandleToFD(bo->dev->fd, bo->handle,
242 DRM_CLOEXEC | DRM_RDWR,
243 (int*)shared_handle);
244 }
245 return -EINVAL;
246 }
247
amdgpu_bo_import(amdgpu_device_handle dev,enum amdgpu_bo_handle_type type,uint32_t shared_handle,struct amdgpu_bo_import_result * output)248 drm_public int amdgpu_bo_import(amdgpu_device_handle dev,
249 enum amdgpu_bo_handle_type type,
250 uint32_t shared_handle,
251 struct amdgpu_bo_import_result *output)
252 {
253 struct drm_gem_open open_arg = {};
254 struct amdgpu_bo *bo = NULL;
255 uint32_t handle = 0, flink_name = 0;
256 uint64_t alloc_size = 0;
257 int r = 0;
258 int dma_fd;
259 uint64_t dma_buf_size = 0;
260
261 /* We must maintain a list of pairs <handle, bo>, so that we always
262 * return the same amdgpu_bo instance for the same handle. */
263 pthread_mutex_lock(&dev->bo_table_mutex);
264
265 /* Convert a DMA buf handle to a KMS handle now. */
266 if (type == amdgpu_bo_handle_type_dma_buf_fd) {
267 off_t size;
268
269 /* Get a KMS handle. */
270 r = drmPrimeFDToHandle(dev->fd, shared_handle, &handle);
271 if (r)
272 goto unlock;
273
274 /* Query the buffer size. */
275 size = lseek(shared_handle, 0, SEEK_END);
276 if (size == (off_t)-1) {
277 r = -errno;
278 goto free_bo_handle;
279 }
280 lseek(shared_handle, 0, SEEK_SET);
281
282 dma_buf_size = size;
283 shared_handle = handle;
284 }
285
286 /* If we have already created a buffer with this handle, find it. */
287 switch (type) {
288 case amdgpu_bo_handle_type_gem_flink_name:
289 bo = handle_table_lookup(&dev->bo_flink_names, shared_handle);
290 break;
291
292 case amdgpu_bo_handle_type_dma_buf_fd:
293 bo = handle_table_lookup(&dev->bo_handles, shared_handle);
294 break;
295
296 case amdgpu_bo_handle_type_kms:
297 case amdgpu_bo_handle_type_kms_noimport:
298 /* Importing a KMS handle in not allowed. */
299 r = -EPERM;
300 goto unlock;
301
302 default:
303 r = -EINVAL;
304 goto unlock;
305 }
306
307 if (bo) {
308 /* The buffer already exists, just bump the refcount. */
309 atomic_inc(&bo->refcount);
310 pthread_mutex_unlock(&dev->bo_table_mutex);
311
312 output->buf_handle = bo;
313 output->alloc_size = bo->alloc_size;
314 return 0;
315 }
316
317 /* Open the handle. */
318 switch (type) {
319 case amdgpu_bo_handle_type_gem_flink_name:
320 open_arg.name = shared_handle;
321 r = drmIoctl(dev->flink_fd, DRM_IOCTL_GEM_OPEN, &open_arg);
322 if (r)
323 goto unlock;
324
325 flink_name = shared_handle;
326 handle = open_arg.handle;
327 alloc_size = open_arg.size;
328 if (dev->flink_fd != dev->fd) {
329 r = drmPrimeHandleToFD(dev->flink_fd, handle,
330 DRM_CLOEXEC, &dma_fd);
331 if (r)
332 goto free_bo_handle;
333 r = drmPrimeFDToHandle(dev->fd, dma_fd, &handle);
334 close(dma_fd);
335 if (r)
336 goto free_bo_handle;
337 r = drmCloseBufferHandle(dev->flink_fd,
338 open_arg.handle);
339 if (r)
340 goto free_bo_handle;
341 }
342 open_arg.handle = 0;
343 break;
344
345 case amdgpu_bo_handle_type_dma_buf_fd:
346 handle = shared_handle;
347 alloc_size = dma_buf_size;
348 break;
349
350 case amdgpu_bo_handle_type_kms:
351 case amdgpu_bo_handle_type_kms_noimport:
352 assert(0); /* unreachable */
353 }
354
355 /* Initialize it. */
356 r = amdgpu_bo_create(dev, alloc_size, handle, &bo);
357 if (r)
358 goto free_bo_handle;
359
360 if (flink_name) {
361 bo->flink_name = flink_name;
362 r = handle_table_insert(&dev->bo_flink_names, flink_name,
363 bo);
364 if (r)
365 goto free_bo_handle;
366
367 }
368
369 output->buf_handle = bo;
370 output->alloc_size = bo->alloc_size;
371 pthread_mutex_unlock(&dev->bo_table_mutex);
372 return 0;
373
374 free_bo_handle:
375 if (flink_name && open_arg.handle)
376 drmCloseBufferHandle(dev->flink_fd, open_arg.handle);
377
378 if (bo)
379 amdgpu_bo_free(bo);
380 else
381 drmCloseBufferHandle(dev->fd, handle);
382 unlock:
383 pthread_mutex_unlock(&dev->bo_table_mutex);
384 return r;
385 }
386
amdgpu_bo_free(amdgpu_bo_handle buf_handle)387 drm_public int amdgpu_bo_free(amdgpu_bo_handle buf_handle)
388 {
389 struct amdgpu_device *dev;
390 struct amdgpu_bo *bo = buf_handle;
391
392 assert(bo != NULL);
393 dev = bo->dev;
394 pthread_mutex_lock(&dev->bo_table_mutex);
395
396 if (update_references(&bo->refcount, NULL)) {
397 /* Remove the buffer from the hash tables. */
398 handle_table_remove(&dev->bo_handles, bo->handle);
399
400 if (bo->flink_name)
401 handle_table_remove(&dev->bo_flink_names,
402 bo->flink_name);
403
404 /* Release CPU access. */
405 if (bo->cpu_map_count > 0) {
406 bo->cpu_map_count = 1;
407 amdgpu_bo_cpu_unmap(bo);
408 }
409
410 drmCloseBufferHandle(dev->fd, bo->handle);
411 pthread_mutex_destroy(&bo->cpu_access_mutex);
412 free(bo);
413 }
414
415 pthread_mutex_unlock(&dev->bo_table_mutex);
416
417 return 0;
418 }
419
amdgpu_bo_inc_ref(amdgpu_bo_handle bo)420 drm_public void amdgpu_bo_inc_ref(amdgpu_bo_handle bo)
421 {
422 atomic_inc(&bo->refcount);
423 }
424
amdgpu_bo_cpu_map(amdgpu_bo_handle bo,void ** cpu)425 drm_public int amdgpu_bo_cpu_map(amdgpu_bo_handle bo, void **cpu)
426 {
427 union drm_amdgpu_gem_mmap args;
428 void *ptr;
429 int r;
430
431 pthread_mutex_lock(&bo->cpu_access_mutex);
432
433 if (bo->cpu_ptr) {
434 /* already mapped */
435 assert(bo->cpu_map_count > 0);
436 bo->cpu_map_count++;
437 *cpu = bo->cpu_ptr;
438 pthread_mutex_unlock(&bo->cpu_access_mutex);
439 return 0;
440 }
441
442 assert(bo->cpu_map_count == 0);
443
444 memset(&args, 0, sizeof(args));
445
446 /* Query the buffer address (args.addr_ptr).
447 * The kernel driver ignores the offset and size parameters. */
448 args.in.handle = bo->handle;
449
450 r = drmCommandWriteRead(bo->dev->fd, DRM_AMDGPU_GEM_MMAP, &args,
451 sizeof(args));
452 if (r) {
453 pthread_mutex_unlock(&bo->cpu_access_mutex);
454 return r;
455 }
456
457 /* Map the buffer. */
458 ptr = drm_mmap(NULL, bo->alloc_size, PROT_READ | PROT_WRITE, MAP_SHARED,
459 bo->dev->fd, args.out.addr_ptr);
460 if (ptr == MAP_FAILED) {
461 pthread_mutex_unlock(&bo->cpu_access_mutex);
462 return -errno;
463 }
464
465 bo->cpu_ptr = ptr;
466 bo->cpu_map_count = 1;
467 pthread_mutex_unlock(&bo->cpu_access_mutex);
468
469 *cpu = ptr;
470 return 0;
471 }
472
amdgpu_bo_cpu_unmap(amdgpu_bo_handle bo)473 drm_public int amdgpu_bo_cpu_unmap(amdgpu_bo_handle bo)
474 {
475 int r;
476
477 pthread_mutex_lock(&bo->cpu_access_mutex);
478 assert(bo->cpu_map_count >= 0);
479
480 if (bo->cpu_map_count == 0) {
481 /* not mapped */
482 pthread_mutex_unlock(&bo->cpu_access_mutex);
483 return -EINVAL;
484 }
485
486 bo->cpu_map_count--;
487 if (bo->cpu_map_count > 0) {
488 /* mapped multiple times */
489 pthread_mutex_unlock(&bo->cpu_access_mutex);
490 return 0;
491 }
492
493 r = drm_munmap(bo->cpu_ptr, bo->alloc_size) == 0 ? 0 : -errno;
494 bo->cpu_ptr = NULL;
495 pthread_mutex_unlock(&bo->cpu_access_mutex);
496 return r;
497 }
498
amdgpu_query_buffer_size_alignment(amdgpu_device_handle dev,struct amdgpu_buffer_size_alignments * info)499 drm_public int amdgpu_query_buffer_size_alignment(amdgpu_device_handle dev,
500 struct amdgpu_buffer_size_alignments *info)
501 {
502 info->size_local = dev->dev_info.pte_fragment_size;
503 info->size_remote = dev->dev_info.gart_page_size;
504 return 0;
505 }
506
amdgpu_bo_wait_for_idle(amdgpu_bo_handle bo,uint64_t timeout_ns,bool * busy)507 drm_public int amdgpu_bo_wait_for_idle(amdgpu_bo_handle bo,
508 uint64_t timeout_ns,
509 bool *busy)
510 {
511 union drm_amdgpu_gem_wait_idle args;
512 int r;
513
514 memset(&args, 0, sizeof(args));
515 args.in.handle = bo->handle;
516 args.in.timeout = amdgpu_cs_calculate_timeout(timeout_ns);
517
518 r = drmCommandWriteRead(bo->dev->fd, DRM_AMDGPU_GEM_WAIT_IDLE,
519 &args, sizeof(args));
520
521 if (r == 0) {
522 *busy = args.out.status;
523 return 0;
524 } else {
525 fprintf(stderr, "amdgpu: GEM_WAIT_IDLE failed with %i\n", r);
526 return r;
527 }
528 }
529
amdgpu_find_bo_by_cpu_mapping(amdgpu_device_handle dev,void * cpu,uint64_t size,amdgpu_bo_handle * buf_handle,uint64_t * offset_in_bo)530 drm_public int amdgpu_find_bo_by_cpu_mapping(amdgpu_device_handle dev,
531 void *cpu,
532 uint64_t size,
533 amdgpu_bo_handle *buf_handle,
534 uint64_t *offset_in_bo)
535 {
536 struct amdgpu_bo *bo;
537 uint32_t i;
538 int r = 0;
539
540 if (cpu == NULL || size == 0)
541 return -EINVAL;
542
543 /*
544 * Workaround for a buggy application which tries to import previously
545 * exposed CPU pointers. If we find a real world use case we should
546 * improve that by asking the kernel for the right handle.
547 */
548 pthread_mutex_lock(&dev->bo_table_mutex);
549 for (i = 0; i < dev->bo_handles.max_key; i++) {
550 bo = handle_table_lookup(&dev->bo_handles, i);
551 if (!bo || !bo->cpu_ptr || size > bo->alloc_size)
552 continue;
553 if (cpu >= bo->cpu_ptr &&
554 cpu < (void*)((uintptr_t)bo->cpu_ptr + bo->alloc_size))
555 break;
556 }
557
558 if (i < dev->bo_handles.max_key) {
559 atomic_inc(&bo->refcount);
560 *buf_handle = bo;
561 *offset_in_bo = (uintptr_t)cpu - (uintptr_t)bo->cpu_ptr;
562 } else {
563 *buf_handle = NULL;
564 *offset_in_bo = 0;
565 r = -ENXIO;
566 }
567 pthread_mutex_unlock(&dev->bo_table_mutex);
568
569 return r;
570 }
571
amdgpu_create_bo_from_user_mem(amdgpu_device_handle dev,void * cpu,uint64_t size,amdgpu_bo_handle * buf_handle)572 drm_public int amdgpu_create_bo_from_user_mem(amdgpu_device_handle dev,
573 void *cpu,
574 uint64_t size,
575 amdgpu_bo_handle *buf_handle)
576 {
577 int r;
578 struct drm_amdgpu_gem_userptr args;
579
580 args.addr = (uintptr_t)cpu;
581 args.flags = AMDGPU_GEM_USERPTR_ANONONLY | AMDGPU_GEM_USERPTR_REGISTER |
582 AMDGPU_GEM_USERPTR_VALIDATE;
583 args.size = size;
584 r = drmCommandWriteRead(dev->fd, DRM_AMDGPU_GEM_USERPTR,
585 &args, sizeof(args));
586 if (r)
587 goto out;
588
589 pthread_mutex_lock(&dev->bo_table_mutex);
590 r = amdgpu_bo_create(dev, size, args.handle, buf_handle);
591 pthread_mutex_unlock(&dev->bo_table_mutex);
592 if (r) {
593 drmCloseBufferHandle(dev->fd, args.handle);
594 }
595
596 out:
597 return r;
598 }
599
amdgpu_bo_list_create_raw(amdgpu_device_handle dev,uint32_t number_of_buffers,struct drm_amdgpu_bo_list_entry * buffers,uint32_t * result)600 drm_public int amdgpu_bo_list_create_raw(amdgpu_device_handle dev,
601 uint32_t number_of_buffers,
602 struct drm_amdgpu_bo_list_entry *buffers,
603 uint32_t *result)
604 {
605 union drm_amdgpu_bo_list args;
606 int r;
607
608 memset(&args, 0, sizeof(args));
609 args.in.operation = AMDGPU_BO_LIST_OP_CREATE;
610 args.in.bo_number = number_of_buffers;
611 args.in.bo_info_size = sizeof(struct drm_amdgpu_bo_list_entry);
612 args.in.bo_info_ptr = (uint64_t)(uintptr_t)buffers;
613
614 r = drmCommandWriteRead(dev->fd, DRM_AMDGPU_BO_LIST,
615 &args, sizeof(args));
616 if (!r)
617 *result = args.out.list_handle;
618 return r;
619 }
620
amdgpu_bo_list_destroy_raw(amdgpu_device_handle dev,uint32_t bo_list)621 drm_public int amdgpu_bo_list_destroy_raw(amdgpu_device_handle dev,
622 uint32_t bo_list)
623 {
624 union drm_amdgpu_bo_list args;
625
626 memset(&args, 0, sizeof(args));
627 args.in.operation = AMDGPU_BO_LIST_OP_DESTROY;
628 args.in.list_handle = bo_list;
629
630 return drmCommandWriteRead(dev->fd, DRM_AMDGPU_BO_LIST,
631 &args, sizeof(args));
632 }
633
amdgpu_bo_list_create(amdgpu_device_handle dev,uint32_t number_of_resources,amdgpu_bo_handle * resources,uint8_t * resource_prios,amdgpu_bo_list_handle * result)634 drm_public int amdgpu_bo_list_create(amdgpu_device_handle dev,
635 uint32_t number_of_resources,
636 amdgpu_bo_handle *resources,
637 uint8_t *resource_prios,
638 amdgpu_bo_list_handle *result)
639 {
640 struct drm_amdgpu_bo_list_entry *list;
641 union drm_amdgpu_bo_list args;
642 unsigned i;
643 int r;
644
645 if (!number_of_resources)
646 return -EINVAL;
647
648 /* overflow check for multiplication */
649 if (number_of_resources > UINT32_MAX / sizeof(struct drm_amdgpu_bo_list_entry))
650 return -EINVAL;
651
652 list = malloc(number_of_resources * sizeof(struct drm_amdgpu_bo_list_entry));
653 if (!list)
654 return -ENOMEM;
655
656 *result = malloc(sizeof(struct amdgpu_bo_list));
657 if (!*result) {
658 free(list);
659 return -ENOMEM;
660 }
661
662 memset(&args, 0, sizeof(args));
663 args.in.operation = AMDGPU_BO_LIST_OP_CREATE;
664 args.in.bo_number = number_of_resources;
665 args.in.bo_info_size = sizeof(struct drm_amdgpu_bo_list_entry);
666 args.in.bo_info_ptr = (uint64_t)(uintptr_t)list;
667
668 for (i = 0; i < number_of_resources; i++) {
669 list[i].bo_handle = resources[i]->handle;
670 if (resource_prios)
671 list[i].bo_priority = resource_prios[i];
672 else
673 list[i].bo_priority = 0;
674 }
675
676 r = drmCommandWriteRead(dev->fd, DRM_AMDGPU_BO_LIST,
677 &args, sizeof(args));
678 free(list);
679 if (r) {
680 free(*result);
681 return r;
682 }
683
684 (*result)->dev = dev;
685 (*result)->handle = args.out.list_handle;
686 return 0;
687 }
688
amdgpu_bo_list_destroy(amdgpu_bo_list_handle list)689 drm_public int amdgpu_bo_list_destroy(amdgpu_bo_list_handle list)
690 {
691 union drm_amdgpu_bo_list args;
692 int r;
693
694 memset(&args, 0, sizeof(args));
695 args.in.operation = AMDGPU_BO_LIST_OP_DESTROY;
696 args.in.list_handle = list->handle;
697
698 r = drmCommandWriteRead(list->dev->fd, DRM_AMDGPU_BO_LIST,
699 &args, sizeof(args));
700
701 if (!r)
702 free(list);
703
704 return r;
705 }
706
amdgpu_bo_list_update(amdgpu_bo_list_handle handle,uint32_t number_of_resources,amdgpu_bo_handle * resources,uint8_t * resource_prios)707 drm_public int amdgpu_bo_list_update(amdgpu_bo_list_handle handle,
708 uint32_t number_of_resources,
709 amdgpu_bo_handle *resources,
710 uint8_t *resource_prios)
711 {
712 struct drm_amdgpu_bo_list_entry *list;
713 union drm_amdgpu_bo_list args;
714 unsigned i;
715 int r;
716
717 if (!number_of_resources)
718 return -EINVAL;
719
720 /* overflow check for multiplication */
721 if (number_of_resources > UINT32_MAX / sizeof(struct drm_amdgpu_bo_list_entry))
722 return -EINVAL;
723
724 list = malloc(number_of_resources * sizeof(struct drm_amdgpu_bo_list_entry));
725 if (!list)
726 return -ENOMEM;
727
728 args.in.operation = AMDGPU_BO_LIST_OP_UPDATE;
729 args.in.list_handle = handle->handle;
730 args.in.bo_number = number_of_resources;
731 args.in.bo_info_size = sizeof(struct drm_amdgpu_bo_list_entry);
732 args.in.bo_info_ptr = (uintptr_t)list;
733
734 for (i = 0; i < number_of_resources; i++) {
735 list[i].bo_handle = resources[i]->handle;
736 if (resource_prios)
737 list[i].bo_priority = resource_prios[i];
738 else
739 list[i].bo_priority = 0;
740 }
741
742 r = drmCommandWriteRead(handle->dev->fd, DRM_AMDGPU_BO_LIST,
743 &args, sizeof(args));
744 free(list);
745 return r;
746 }
747
amdgpu_bo_va_op(amdgpu_bo_handle bo,uint64_t offset,uint64_t size,uint64_t addr,uint64_t flags,uint32_t ops)748 drm_public int amdgpu_bo_va_op(amdgpu_bo_handle bo,
749 uint64_t offset,
750 uint64_t size,
751 uint64_t addr,
752 uint64_t flags,
753 uint32_t ops)
754 {
755 amdgpu_device_handle dev = bo->dev;
756
757 size = ALIGN(size, getpagesize());
758
759 return amdgpu_bo_va_op_raw(dev, bo, offset, size, addr,
760 AMDGPU_VM_PAGE_READABLE |
761 AMDGPU_VM_PAGE_WRITEABLE |
762 AMDGPU_VM_PAGE_EXECUTABLE, ops);
763 }
764
amdgpu_bo_va_op_raw(amdgpu_device_handle dev,amdgpu_bo_handle bo,uint64_t offset,uint64_t size,uint64_t addr,uint64_t flags,uint32_t ops)765 drm_public int amdgpu_bo_va_op_raw(amdgpu_device_handle dev,
766 amdgpu_bo_handle bo,
767 uint64_t offset,
768 uint64_t size,
769 uint64_t addr,
770 uint64_t flags,
771 uint32_t ops)
772 {
773 struct drm_amdgpu_gem_va va;
774 int r;
775
776 if (ops != AMDGPU_VA_OP_MAP && ops != AMDGPU_VA_OP_UNMAP &&
777 ops != AMDGPU_VA_OP_REPLACE && ops != AMDGPU_VA_OP_CLEAR)
778 return -EINVAL;
779
780 memset(&va, 0, sizeof(va));
781 va.handle = bo ? bo->handle : 0;
782 va.operation = ops;
783 va.flags = flags;
784 va.va_address = addr;
785 va.offset_in_bo = offset;
786 va.map_size = size;
787
788 r = drmCommandWriteRead(dev->fd, DRM_AMDGPU_GEM_VA, &va, sizeof(va));
789
790 return r;
791 }
792