• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * Copyright © 2008 Jérôme Glisse
3  * All Rights Reserved.
4  *
5  * Permission is hereby granted, free of charge, to any person obtaining
6  * a copy of this software and associated documentation files (the
7  * "Software"), to deal in the Software without restriction, including
8  * without limitation the rights to use, copy, modify, merge, publish,
9  * distribute, sub license, and/or sell copies of the Software, and to
10  * permit persons to whom the Software is furnished to do so, subject to
11  * the following conditions:
12  *
13  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
14  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
15  * OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
16  * NON-INFRINGEMENT. IN NO EVENT SHALL THE COPYRIGHT HOLDERS, AUTHORS
17  * AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
18  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
20  * USE OR OTHER DEALINGS IN THE SOFTWARE.
21  *
22  * The above copyright notice and this permission notice (including the
23  * next paragraph) shall be included in all copies or substantial portions
24  * of the Software.
25  */
26 /*
27  * Authors:
28  *      Aapo Tahkola <aet@rasterburn.org>
29  *      Nicolai Haehnle <prefect_@gmx.net>
30  *      Jérôme Glisse <glisse@freedesktop.org>
31  */
32 #include <assert.h>
33 #include <errno.h>
34 #include <stdlib.h>
35 #include <string.h>
36 #include <pthread.h>
37 #include <sys/ioctl.h>
38 #include "radeon_cs.h"
39 #include "radeon_cs_int.h"
40 #include "radeon_bo_int.h"
41 #include "radeon_cs_gem.h"
42 #include "radeon_bo_gem.h"
43 #include "drm.h"
44 #include "libdrm_macros.h"
45 #include "xf86drm.h"
46 #include "xf86atomic.h"
47 #include "radeon_drm.h"
48 
49 /* Add LIBDRM_RADEON_BOF_FILES to libdrm_radeon_la_SOURCES when building with BOF_DUMP */
50 #define CS_BOF_DUMP 0
51 #if CS_BOF_DUMP
52 #include "bof.h"
53 #endif
54 
55 struct radeon_cs_manager_gem {
56     struct radeon_cs_manager    base;
57     uint32_t                    device_id;
58     unsigned                    nbof;
59 };
60 
61 #pragma pack(1)
62 struct cs_reloc_gem {
63     uint32_t    handle;
64     uint32_t    read_domain;
65     uint32_t    write_domain;
66     uint32_t    flags;
67 };
68 
69 #pragma pack()
70 #define RELOC_SIZE (sizeof(struct cs_reloc_gem) / sizeof(uint32_t))
71 
72 struct cs_gem {
73     struct radeon_cs_int        base;
74     struct drm_radeon_cs        cs;
75     struct drm_radeon_cs_chunk  chunks[2];
76     unsigned                    nrelocs;
77     uint32_t                    *relocs;
78     struct radeon_bo_int        **relocs_bo;
79 };
80 
81 static pthread_mutex_t id_mutex = PTHREAD_MUTEX_INITIALIZER;
82 static uint32_t cs_id_source = 0;
83 
84 /**
85  * result is undefined if called with ~0
86  */
get_first_zero(const uint32_t n)87 static uint32_t get_first_zero(const uint32_t n)
88 {
89     /* __builtin_ctz returns number of trailing zeros. */
90     return 1 << __builtin_ctz(~n);
91 }
92 
93 /**
94  * Returns a free id for cs.
95  * If there is no free id we return zero
96  **/
generate_id(void)97 static uint32_t generate_id(void)
98 {
99     uint32_t r = 0;
100     pthread_mutex_lock( &id_mutex );
101     /* check for free ids */
102     if (cs_id_source != ~r) {
103         /* find first zero bit */
104         r = get_first_zero(cs_id_source);
105 
106         /* set id as reserved */
107         cs_id_source |= r;
108     }
109     pthread_mutex_unlock( &id_mutex );
110     return r;
111 }
112 
113 /**
114  * Free the id for later reuse
115  **/
free_id(uint32_t id)116 static void free_id(uint32_t id)
117 {
118     pthread_mutex_lock( &id_mutex );
119 
120     cs_id_source &= ~id;
121 
122     pthread_mutex_unlock( &id_mutex );
123 }
124 
cs_gem_create(struct radeon_cs_manager * csm,uint32_t ndw)125 static struct radeon_cs_int *cs_gem_create(struct radeon_cs_manager *csm,
126                                        uint32_t ndw)
127 {
128     struct cs_gem *csg;
129 
130     /* max cmd buffer size is 64Kb */
131     if (ndw > (64 * 1024 / 4)) {
132         return NULL;
133     }
134     csg = (struct cs_gem*)calloc(1, sizeof(struct cs_gem));
135     if (csg == NULL) {
136         return NULL;
137     }
138     csg->base.csm = csm;
139     csg->base.ndw = 64 * 1024 / 4;
140     csg->base.packets = (uint32_t*)calloc(1, 64 * 1024);
141     if (csg->base.packets == NULL) {
142         free(csg);
143         return NULL;
144     }
145     csg->base.relocs_total_size = 0;
146     csg->base.crelocs = 0;
147     csg->base.id = generate_id();
148     csg->nrelocs = 4096 / (4 * 4) ;
149     csg->relocs_bo = (struct radeon_bo_int**)calloc(1,
150                                                 csg->nrelocs*sizeof(void*));
151     if (csg->relocs_bo == NULL) {
152         free(csg->base.packets);
153         free(csg);
154         return NULL;
155     }
156     csg->base.relocs = csg->relocs = (uint32_t*)calloc(1, 4096);
157     if (csg->relocs == NULL) {
158         free(csg->relocs_bo);
159         free(csg->base.packets);
160         free(csg);
161         return NULL;
162     }
163     csg->chunks[0].chunk_id = RADEON_CHUNK_ID_IB;
164     csg->chunks[0].length_dw = 0;
165     csg->chunks[0].chunk_data = (uint64_t)(uintptr_t)csg->base.packets;
166     csg->chunks[1].chunk_id = RADEON_CHUNK_ID_RELOCS;
167     csg->chunks[1].length_dw = 0;
168     csg->chunks[1].chunk_data = (uint64_t)(uintptr_t)csg->relocs;
169     return (struct radeon_cs_int*)csg;
170 }
171 
cs_gem_write_reloc(struct radeon_cs_int * cs,struct radeon_bo * bo,uint32_t read_domain,uint32_t write_domain,uint32_t flags)172 static int cs_gem_write_reloc(struct radeon_cs_int *cs,
173                               struct radeon_bo *bo,
174                               uint32_t read_domain,
175                               uint32_t write_domain,
176                               uint32_t flags)
177 {
178     struct radeon_bo_int *boi = (struct radeon_bo_int *)bo;
179     struct cs_gem *csg = (struct cs_gem*)cs;
180     struct cs_reloc_gem *reloc;
181     uint32_t idx;
182     unsigned i;
183 
184     assert(boi->space_accounted);
185 
186     /* check domains */
187     if ((read_domain && write_domain) || (!read_domain && !write_domain)) {
188         /* in one CS a bo can only be in read or write domain but not
189          * in read & write domain at the same time
190          */
191         return -EINVAL;
192     }
193     if (read_domain == RADEON_GEM_DOMAIN_CPU) {
194         return -EINVAL;
195     }
196     if (write_domain == RADEON_GEM_DOMAIN_CPU) {
197         return -EINVAL;
198     }
199     /* use bit field hash function to determine
200        if this bo is for sure not in this cs.*/
201     if ((atomic_read((atomic_t *)radeon_gem_get_reloc_in_cs(bo)) & cs->id)) {
202         /* check if bo is already referenced.
203          * Scanning from end to begin reduces cycles with mesa because
204          * it often relocates same shared dma bo again. */
205         for(i = cs->crelocs; i != 0;) {
206             --i;
207             idx = i * RELOC_SIZE;
208             reloc = (struct cs_reloc_gem*)&csg->relocs[idx];
209             if (reloc->handle == bo->handle) {
210                 /* Check domains must be in read or write. As we check already
211                  * checked that in argument one of the read or write domain was
212                  * set we only need to check that if previous reloc as the read
213                  * domain set then the read_domain should also be set for this
214                  * new relocation.
215                  */
216                 /* the DDX expects to read and write from same pixmap */
217                 if (write_domain && (reloc->read_domain & write_domain)) {
218                     reloc->read_domain = 0;
219                     reloc->write_domain = write_domain;
220                 } else if (read_domain & reloc->write_domain) {
221                     reloc->read_domain = 0;
222                 } else {
223                     if (write_domain != reloc->write_domain)
224                         return -EINVAL;
225                     if (read_domain != reloc->read_domain)
226                         return -EINVAL;
227                 }
228 
229                 reloc->read_domain |= read_domain;
230                 reloc->write_domain |= write_domain;
231                 /* update flags */
232                 reloc->flags |= (flags & reloc->flags);
233                 /* write relocation packet */
234                 radeon_cs_write_dword((struct radeon_cs *)cs, 0xc0001000);
235                 radeon_cs_write_dword((struct radeon_cs *)cs, idx);
236                 return 0;
237             }
238         }
239     }
240     /* new relocation */
241     if (csg->base.crelocs >= csg->nrelocs) {
242         /* allocate more memory (TODO: should use a slab allocator maybe) */
243         uint32_t *tmp, size;
244         size = ((csg->nrelocs + 1) * sizeof(struct radeon_bo*));
245         tmp = (uint32_t*)realloc(csg->relocs_bo, size);
246         if (tmp == NULL) {
247             return -ENOMEM;
248         }
249         csg->relocs_bo = (struct radeon_bo_int **)tmp;
250         size = ((csg->nrelocs + 1) * RELOC_SIZE * 4);
251         tmp = (uint32_t*)realloc(csg->relocs, size);
252         if (tmp == NULL) {
253             return -ENOMEM;
254         }
255         cs->relocs = csg->relocs = tmp;
256         csg->nrelocs += 1;
257         csg->chunks[1].chunk_data = (uint64_t)(uintptr_t)csg->relocs;
258     }
259     csg->relocs_bo[csg->base.crelocs] = boi;
260     idx = (csg->base.crelocs++) * RELOC_SIZE;
261     reloc = (struct cs_reloc_gem*)&csg->relocs[idx];
262     reloc->handle = bo->handle;
263     reloc->read_domain = read_domain;
264     reloc->write_domain = write_domain;
265     reloc->flags = flags;
266     csg->chunks[1].length_dw += RELOC_SIZE;
267     radeon_bo_ref(bo);
268     /* bo might be referenced from another context so have to use atomic operations */
269     atomic_add((atomic_t *)radeon_gem_get_reloc_in_cs(bo), cs->id);
270     cs->relocs_total_size += boi->size;
271     radeon_cs_write_dword((struct radeon_cs *)cs, 0xc0001000);
272     radeon_cs_write_dword((struct radeon_cs *)cs, idx);
273     return 0;
274 }
275 
cs_gem_begin(struct radeon_cs_int * cs,uint32_t ndw,const char * file,const char * func,int line)276 static int cs_gem_begin(struct radeon_cs_int *cs,
277                         uint32_t ndw,
278                         const char *file,
279                         const char *func,
280                         int line)
281 {
282 
283     if (cs->section_ndw) {
284         fprintf(stderr, "CS already in a section(%s,%s,%d)\n",
285                 cs->section_file, cs->section_func, cs->section_line);
286         fprintf(stderr, "CS can't start section(%s,%s,%d)\n",
287                 file, func, line);
288         return -EPIPE;
289     }
290     cs->section_ndw = ndw;
291     cs->section_cdw = 0;
292     cs->section_file = file;
293     cs->section_func = func;
294     cs->section_line = line;
295 
296     if (cs->cdw + ndw > cs->ndw) {
297         uint32_t tmp, *ptr;
298 
299         /* round up the required size to a multiple of 1024 */
300         tmp = (cs->cdw + ndw + 0x3FF) & (~0x3FF);
301         ptr = (uint32_t*)realloc(cs->packets, 4 * tmp);
302         if (ptr == NULL) {
303             return -ENOMEM;
304         }
305         cs->packets = ptr;
306         cs->ndw = tmp;
307     }
308     return 0;
309 }
310 
cs_gem_end(struct radeon_cs_int * cs,const char * file,const char * func,int line)311 static int cs_gem_end(struct radeon_cs_int *cs,
312                       const char *file,
313                       const char *func,
314                       int line)
315 
316 {
317     if (!cs->section_ndw) {
318         fprintf(stderr, "CS no section to end at (%s,%s,%d)\n",
319                 file, func, line);
320         return -EPIPE;
321     }
322     if (cs->section_ndw != cs->section_cdw) {
323         fprintf(stderr, "CS section size mismatch start at (%s,%s,%d) %d vs %d\n",
324                 cs->section_file, cs->section_func, cs->section_line, cs->section_ndw, cs->section_cdw);
325         fprintf(stderr, "CS section end at (%s,%s,%d)\n",
326                 file, func, line);
327 
328         /* We must reset the section even when there is error. */
329         cs->section_ndw = 0;
330         return -EPIPE;
331     }
332     cs->section_ndw = 0;
333     return 0;
334 }
335 
336 #if CS_BOF_DUMP
cs_gem_dump_bof(struct radeon_cs_int * cs)337 static void cs_gem_dump_bof(struct radeon_cs_int *cs)
338 {
339     struct cs_gem *csg = (struct cs_gem*)cs;
340     struct radeon_cs_manager_gem *csm;
341     bof_t *bcs, *blob, *array, *bo, *size, *handle, *device_id, *root;
342     char tmp[256];
343     unsigned i;
344 
345     csm = (struct radeon_cs_manager_gem *)cs->csm;
346     root = device_id = bcs = blob = array = bo = size = handle = NULL;
347     root = bof_object();
348     if (root == NULL)
349         goto out_err;
350     device_id = bof_int32(csm->device_id);
351     if (device_id == NULL)
352         return;
353     if (bof_object_set(root, "device_id", device_id))
354         goto out_err;
355     bof_decref(device_id);
356     device_id = NULL;
357     /* dump relocs */
358     blob = bof_blob(csg->nrelocs * 16, csg->relocs);
359     if (blob == NULL)
360         goto out_err;
361     if (bof_object_set(root, "reloc", blob))
362         goto out_err;
363     bof_decref(blob);
364     blob = NULL;
365     /* dump cs */
366     blob = bof_blob(cs->cdw * 4, cs->packets);
367     if (blob == NULL)
368         goto out_err;
369     if (bof_object_set(root, "pm4", blob))
370         goto out_err;
371     bof_decref(blob);
372     blob = NULL;
373     /* dump bo */
374     array = bof_array();
375     if (array == NULL)
376         goto out_err;
377     for (i = 0; i < csg->base.crelocs; i++) {
378         bo = bof_object();
379         if (bo == NULL)
380             goto out_err;
381         size = bof_int32(csg->relocs_bo[i]->size);
382         if (size == NULL)
383             goto out_err;
384         if (bof_object_set(bo, "size", size))
385             goto out_err;
386         bof_decref(size);
387         size = NULL;
388         handle = bof_int32(csg->relocs_bo[i]->handle);
389         if (handle == NULL)
390             goto out_err;
391         if (bof_object_set(bo, "handle", handle))
392             goto out_err;
393         bof_decref(handle);
394         handle = NULL;
395         radeon_bo_map((struct radeon_bo*)csg->relocs_bo[i], 0);
396         blob = bof_blob(csg->relocs_bo[i]->size, csg->relocs_bo[i]->ptr);
397         radeon_bo_unmap((struct radeon_bo*)csg->relocs_bo[i]);
398         if (blob == NULL)
399             goto out_err;
400         if (bof_object_set(bo, "data", blob))
401             goto out_err;
402         bof_decref(blob);
403         blob = NULL;
404         if (bof_array_append(array, bo))
405             goto out_err;
406         bof_decref(bo);
407         bo = NULL;
408     }
409     if (bof_object_set(root, "bo", array))
410         goto out_err;
411     sprintf(tmp, "d-0x%04X-%08d.bof", csm->device_id, csm->nbof++);
412     bof_dump_file(root, tmp);
413 out_err:
414     bof_decref(blob);
415     bof_decref(array);
416     bof_decref(bo);
417     bof_decref(size);
418     bof_decref(handle);
419     bof_decref(device_id);
420     bof_decref(root);
421 }
422 #endif
423 
cs_gem_emit(struct radeon_cs_int * cs)424 static int cs_gem_emit(struct radeon_cs_int *cs)
425 {
426     struct cs_gem *csg = (struct cs_gem*)cs;
427     uint64_t chunk_array[2];
428     unsigned i;
429     int r;
430 
431     while (cs->cdw & 7)
432 	radeon_cs_write_dword((struct radeon_cs *)cs, 0x80000000);
433 
434 #if CS_BOF_DUMP
435     cs_gem_dump_bof(cs);
436 #endif
437     csg->chunks[0].length_dw = cs->cdw;
438 
439     chunk_array[0] = (uint64_t)(uintptr_t)&csg->chunks[0];
440     chunk_array[1] = (uint64_t)(uintptr_t)&csg->chunks[1];
441 
442     csg->cs.num_chunks = 2;
443     csg->cs.chunks = (uint64_t)(uintptr_t)chunk_array;
444 
445     r = drmCommandWriteRead(cs->csm->fd, DRM_RADEON_CS,
446                             &csg->cs, sizeof(struct drm_radeon_cs));
447     for (i = 0; i < csg->base.crelocs; i++) {
448         csg->relocs_bo[i]->space_accounted = 0;
449         /* bo might be referenced from another context so have to use atomic operations */
450         atomic_dec((atomic_t *)radeon_gem_get_reloc_in_cs((struct radeon_bo*)csg->relocs_bo[i]), cs->id);
451         radeon_bo_unref((struct radeon_bo *)csg->relocs_bo[i]);
452         csg->relocs_bo[i] = NULL;
453     }
454 
455     cs->csm->read_used = 0;
456     cs->csm->vram_write_used = 0;
457     cs->csm->gart_write_used = 0;
458     return r;
459 }
460 
cs_gem_destroy(struct radeon_cs_int * cs)461 static int cs_gem_destroy(struct radeon_cs_int *cs)
462 {
463     struct cs_gem *csg = (struct cs_gem*)cs;
464 
465     free_id(cs->id);
466     free(csg->relocs_bo);
467     free(cs->relocs);
468     free(cs->packets);
469     free(cs);
470     return 0;
471 }
472 
cs_gem_erase(struct radeon_cs_int * cs)473 static int cs_gem_erase(struct radeon_cs_int *cs)
474 {
475     struct cs_gem *csg = (struct cs_gem*)cs;
476     unsigned i;
477 
478     if (csg->relocs_bo) {
479         for (i = 0; i < csg->base.crelocs; i++) {
480             if (csg->relocs_bo[i]) {
481                 /* bo might be referenced from another context so have to use atomic operations */
482                 atomic_dec((atomic_t *)radeon_gem_get_reloc_in_cs((struct radeon_bo*)csg->relocs_bo[i]), cs->id);
483                 radeon_bo_unref((struct radeon_bo *)csg->relocs_bo[i]);
484                 csg->relocs_bo[i] = NULL;
485             }
486         }
487     }
488     cs->relocs_total_size = 0;
489     cs->cdw = 0;
490     cs->section_ndw = 0;
491     cs->crelocs = 0;
492     csg->chunks[0].length_dw = 0;
493     csg->chunks[1].length_dw = 0;
494     return 0;
495 }
496 
cs_gem_need_flush(struct radeon_cs_int * cs)497 static int cs_gem_need_flush(struct radeon_cs_int *cs)
498 {
499     return 0; //(cs->relocs_total_size > (32*1024*1024));
500 }
501 
cs_gem_print(struct radeon_cs_int * cs,FILE * file)502 static void cs_gem_print(struct radeon_cs_int *cs, FILE *file)
503 {
504     struct radeon_cs_manager_gem *csm;
505     unsigned int i;
506 
507     csm = (struct radeon_cs_manager_gem *)cs->csm;
508     fprintf(file, "VENDORID:DEVICEID 0x%04X:0x%04X\n", 0x1002, csm->device_id);
509     for (i = 0; i < cs->cdw; i++) {
510         fprintf(file, "0x%08X\n", cs->packets[i]);
511     }
512 }
513 
514 static const struct radeon_cs_funcs radeon_cs_gem_funcs = {
515     .cs_create = cs_gem_create,
516     .cs_write_reloc = cs_gem_write_reloc,
517     .cs_begin = cs_gem_begin,
518     .cs_end = cs_gem_end,
519     .cs_emit = cs_gem_emit,
520     .cs_destroy = cs_gem_destroy,
521     .cs_erase = cs_gem_erase,
522     .cs_need_flush = cs_gem_need_flush,
523     .cs_print = cs_gem_print,
524 };
525 
radeon_get_device_id(int fd,uint32_t * device_id)526 static int radeon_get_device_id(int fd, uint32_t *device_id)
527 {
528     struct drm_radeon_info info = {};
529     int r;
530 
531     *device_id = 0;
532     info.request = RADEON_INFO_DEVICE_ID;
533     info.value = (uintptr_t)device_id;
534     r = drmCommandWriteRead(fd, DRM_RADEON_INFO, &info,
535                             sizeof(struct drm_radeon_info));
536     return r;
537 }
538 
radeon_cs_manager_gem_ctor(int fd)539 struct radeon_cs_manager *radeon_cs_manager_gem_ctor(int fd)
540 {
541     struct radeon_cs_manager_gem *csm;
542 
543     csm = calloc(1, sizeof(struct radeon_cs_manager_gem));
544     if (csm == NULL) {
545         return NULL;
546     }
547     csm->base.funcs = &radeon_cs_gem_funcs;
548     csm->base.fd = fd;
549     radeon_get_device_id(fd, &csm->device_id);
550     return &csm->base;
551 }
552 
radeon_cs_manager_gem_dtor(struct radeon_cs_manager * csm)553 void radeon_cs_manager_gem_dtor(struct radeon_cs_manager *csm)
554 {
555     free(csm);
556 }
557