• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /* -*- mode: C; c-file-style: "k&r"; tab-width 4; indent-tabs-mode: t; -*- */
2 
3 /*
4  * Copyright (C) 2011 Texas Instruments, Inc
5  *
6  * Permission is hereby granted, free of charge, to any person obtaining a
7  * copy of this software and associated documentation files (the "Software"),
8  * to deal in the Software without restriction, including without limitation
9  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
10  * and/or sell copies of the Software, and to permit persons to whom the
11  * Software is furnished to do so, subject to the following conditions:
12  *
13  * The above copyright notice and this permission notice (including the next
14  * paragraph) shall be included in all copies or substantial portions of the
15  * Software.
16  *
17  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
18  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
19  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
20  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
21  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
22  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
23  * SOFTWARE.
24  *
25  * Authors:
26  *    Rob Clark <rob@ti.com>
27  */
28 
29 #include <stdlib.h>
30 #include <linux/stddef.h>
31 #include <linux/types.h>
32 #include <errno.h>
33 #include <sys/mman.h>
34 #include <fcntl.h>
35 #include <unistd.h>
36 #include <pthread.h>
37 
38 #include <libdrm_macros.h>
39 #include <xf86drm.h>
40 #include <xf86atomic.h>
41 
42 #include "omap_drm.h"
43 #include "omap_drmif.h"
44 
45 #define __round_mask(x, y) ((__typeof__(x))((y)-1))
46 #define round_up(x, y) ((((x)-1) | __round_mask(x, y))+1)
47 #define PAGE_SIZE 4096
48 
49 static pthread_mutex_t table_lock = PTHREAD_MUTEX_INITIALIZER;
50 static void * dev_table;
51 
52 struct omap_device {
53 	int fd;
54 	atomic_t refcnt;
55 
56 	/* The handle_table is used to track GEM bo handles associated w/
57 	 * this fd.  This is needed, in particular, when importing
58 	 * dmabuf's because we don't want multiple 'struct omap_bo's
59 	 * floating around with the same handle.  Otherwise, when the
60 	 * first one is omap_bo_del()'d the handle becomes no longer
61 	 * valid, and the remaining 'struct omap_bo's are left pointing
62 	 * to an invalid handle (and possible a GEM bo that is already
63 	 * free'd).
64 	 */
65 	void *handle_table;
66 };
67 
68 /* a GEM buffer object allocated from the DRM device */
69 struct omap_bo {
70 	struct omap_device	*dev;
71 	void		*map;		/* userspace mmap'ing (if there is one) */
72 	uint32_t	size;
73 	uint32_t	handle;
74 	uint32_t	name;		/* flink global handle (DRI2 name) */
75 	uint64_t	offset;		/* offset to mmap() */
76 	int		fd;		/* dmabuf handle */
77 	atomic_t	refcnt;
78 };
79 
omap_device_new_impl(int fd)80 static struct omap_device * omap_device_new_impl(int fd)
81 {
82 	struct omap_device *dev = calloc(sizeof(*dev), 1);
83 	if (!dev)
84 		return NULL;
85 	dev->fd = fd;
86 	atomic_set(&dev->refcnt, 1);
87 	dev->handle_table = drmHashCreate();
88 	return dev;
89 }
90 
omap_device_new(int fd)91 struct omap_device * omap_device_new(int fd)
92 {
93 	struct omap_device *dev = NULL;
94 
95 	pthread_mutex_lock(&table_lock);
96 
97 	if (!dev_table)
98 		dev_table = drmHashCreate();
99 
100 	if (drmHashLookup(dev_table, fd, (void **)&dev)) {
101 		/* not found, create new device */
102 		dev = omap_device_new_impl(fd);
103 		drmHashInsert(dev_table, fd, dev);
104 	} else {
105 		/* found, just incr refcnt */
106 		dev = omap_device_ref(dev);
107 	}
108 
109 	pthread_mutex_unlock(&table_lock);
110 
111 	return dev;
112 }
113 
omap_device_ref(struct omap_device * dev)114 struct omap_device * omap_device_ref(struct omap_device *dev)
115 {
116 	atomic_inc(&dev->refcnt);
117 	return dev;
118 }
119 
omap_device_del(struct omap_device * dev)120 void omap_device_del(struct omap_device *dev)
121 {
122 	if (!atomic_dec_and_test(&dev->refcnt))
123 		return;
124 	pthread_mutex_lock(&table_lock);
125 	drmHashDestroy(dev->handle_table);
126 	drmHashDelete(dev_table, dev->fd);
127 	pthread_mutex_unlock(&table_lock);
128 	free(dev);
129 }
130 
131 int
omap_get_param(struct omap_device * dev,uint64_t param,uint64_t * value)132 omap_get_param(struct omap_device *dev, uint64_t param, uint64_t *value)
133 {
134 	struct drm_omap_param req = {
135 			.param = param,
136 	};
137 	int ret;
138 
139 	ret = drmCommandWriteRead(dev->fd, DRM_OMAP_GET_PARAM, &req, sizeof(req));
140 	if (ret) {
141 		return ret;
142 	}
143 
144 	*value = req.value;
145 
146 	return 0;
147 }
148 
149 int
omap_set_param(struct omap_device * dev,uint64_t param,uint64_t value)150 omap_set_param(struct omap_device *dev, uint64_t param, uint64_t value)
151 {
152 	struct drm_omap_param req = {
153 			.param = param,
154 			.value = value,
155 	};
156 	return drmCommandWrite(dev->fd, DRM_OMAP_SET_PARAM, &req, sizeof(req));
157 }
158 
159 /* lookup a buffer from it's handle, call w/ table_lock held: */
lookup_bo(struct omap_device * dev,uint32_t handle)160 static struct omap_bo * lookup_bo(struct omap_device *dev,
161 		uint32_t handle)
162 {
163 	struct omap_bo *bo = NULL;
164 	if (!drmHashLookup(dev->handle_table, handle, (void **)&bo)) {
165 		/* found, incr refcnt and return: */
166 		bo = omap_bo_ref(bo);
167 	}
168 	return bo;
169 }
170 
171 /* allocate a new buffer object, call w/ table_lock held */
bo_from_handle(struct omap_device * dev,uint32_t handle)172 static struct omap_bo * bo_from_handle(struct omap_device *dev,
173 		uint32_t handle)
174 {
175 	struct omap_bo *bo = calloc(sizeof(*bo), 1);
176 	if (!bo) {
177 		struct drm_gem_close req = {
178 				.handle = handle,
179 		};
180 		drmIoctl(dev->fd, DRM_IOCTL_GEM_CLOSE, &req);
181 		return NULL;
182 	}
183 	bo->dev = omap_device_ref(dev);
184 	bo->handle = handle;
185 	bo->fd = -1;
186 	atomic_set(&bo->refcnt, 1);
187 	/* add ourselves to the handle table: */
188 	drmHashInsert(dev->handle_table, handle, bo);
189 	return bo;
190 }
191 
192 /* allocate a new buffer object */
omap_bo_new_impl(struct omap_device * dev,union omap_gem_size size,uint32_t flags)193 static struct omap_bo * omap_bo_new_impl(struct omap_device *dev,
194 		union omap_gem_size size, uint32_t flags)
195 {
196 	struct omap_bo *bo = NULL;
197 	struct drm_omap_gem_new req = {
198 			.size = size,
199 			.flags = flags,
200 	};
201 
202 	if (size.bytes == 0) {
203 		goto fail;
204 	}
205 
206 	if (drmCommandWriteRead(dev->fd, DRM_OMAP_GEM_NEW, &req, sizeof(req))) {
207 		goto fail;
208 	}
209 
210 	pthread_mutex_lock(&table_lock);
211 	bo = bo_from_handle(dev, req.handle);
212 	pthread_mutex_unlock(&table_lock);
213 
214 	if (flags & OMAP_BO_TILED) {
215 		bo->size = round_up(size.tiled.width, PAGE_SIZE) * size.tiled.height;
216 	} else {
217 		bo->size = size.bytes;
218 	}
219 
220 	return bo;
221 
222 fail:
223 	free(bo);
224 	return NULL;
225 }
226 
227 
228 /* allocate a new (un-tiled) buffer object */
229 struct omap_bo *
omap_bo_new(struct omap_device * dev,uint32_t size,uint32_t flags)230 omap_bo_new(struct omap_device *dev, uint32_t size, uint32_t flags)
231 {
232 	union omap_gem_size gsize = {
233 			.bytes = size,
234 	};
235 	if (flags & OMAP_BO_TILED) {
236 		return NULL;
237 	}
238 	return omap_bo_new_impl(dev, gsize, flags);
239 }
240 
241 /* allocate a new buffer object */
242 struct omap_bo *
omap_bo_new_tiled(struct omap_device * dev,uint32_t width,uint32_t height,uint32_t flags)243 omap_bo_new_tiled(struct omap_device *dev, uint32_t width,
244 		  uint32_t height, uint32_t flags)
245 {
246 	union omap_gem_size gsize = {
247 			.tiled = {
248 				.width = width,
249 				.height = height,
250 			},
251 	};
252 	if (!(flags & OMAP_BO_TILED)) {
253 		return NULL;
254 	}
255 	return omap_bo_new_impl(dev, gsize, flags);
256 }
257 
omap_bo_ref(struct omap_bo * bo)258 struct omap_bo *omap_bo_ref(struct omap_bo *bo)
259 {
260 	atomic_inc(&bo->refcnt);
261 	return bo;
262 }
263 
264 /* get buffer info */
get_buffer_info(struct omap_bo * bo)265 static int get_buffer_info(struct omap_bo *bo)
266 {
267 	struct drm_omap_gem_info req = {
268 			.handle = bo->handle,
269 	};
270 	int ret = drmCommandWriteRead(bo->dev->fd, DRM_OMAP_GEM_INFO,
271 			&req, sizeof(req));
272 	if (ret) {
273 		return ret;
274 	}
275 
276 	/* really all we need for now is mmap offset */
277 	bo->offset = req.offset;
278 	bo->size = req.size;
279 
280 	return 0;
281 }
282 
283 /* import a buffer object from DRI2 name */
284 struct omap_bo *
omap_bo_from_name(struct omap_device * dev,uint32_t name)285 omap_bo_from_name(struct omap_device *dev, uint32_t name)
286 {
287 	struct omap_bo *bo = NULL;
288 	struct drm_gem_open req = {
289 			.name = name,
290 	};
291 
292 	pthread_mutex_lock(&table_lock);
293 
294 	if (drmIoctl(dev->fd, DRM_IOCTL_GEM_OPEN, &req)) {
295 		goto fail;
296 	}
297 
298 	bo = lookup_bo(dev, req.handle);
299 	if (!bo) {
300 		bo = bo_from_handle(dev, req.handle);
301 		bo->name = name;
302 	}
303 
304 	pthread_mutex_unlock(&table_lock);
305 
306 	return bo;
307 
308 fail:
309 	pthread_mutex_unlock(&table_lock);
310 	free(bo);
311 	return NULL;
312 }
313 
314 /* import a buffer from dmabuf fd, does not take ownership of the
315  * fd so caller should close() the fd when it is otherwise done
316  * with it (even if it is still using the 'struct omap_bo *')
317  */
318 struct omap_bo *
omap_bo_from_dmabuf(struct omap_device * dev,int fd)319 omap_bo_from_dmabuf(struct omap_device *dev, int fd)
320 {
321 	struct omap_bo *bo = NULL;
322 	struct drm_prime_handle req = {
323 			.fd = fd,
324 	};
325 	int ret;
326 
327 	pthread_mutex_lock(&table_lock);
328 
329 	ret = drmIoctl(dev->fd, DRM_IOCTL_PRIME_FD_TO_HANDLE, &req);
330 	if (ret) {
331 		goto fail;
332 	}
333 
334 	bo = lookup_bo(dev, req.handle);
335 	if (!bo) {
336 		bo = bo_from_handle(dev, req.handle);
337 	}
338 
339 	pthread_mutex_unlock(&table_lock);
340 
341 	return bo;
342 
343 fail:
344 	pthread_mutex_unlock(&table_lock);
345 	free(bo);
346 	return NULL;
347 }
348 
349 /* destroy a buffer object */
omap_bo_del(struct omap_bo * bo)350 void omap_bo_del(struct omap_bo *bo)
351 {
352 	if (!bo) {
353 		return;
354 	}
355 
356 	if (!atomic_dec_and_test(&bo->refcnt))
357 		return;
358 
359 	if (bo->map) {
360 		munmap(bo->map, bo->size);
361 	}
362 
363 	if (bo->fd >= 0) {
364 		close(bo->fd);
365 	}
366 
367 	if (bo->handle) {
368 		struct drm_gem_close req = {
369 				.handle = bo->handle,
370 		};
371 		pthread_mutex_lock(&table_lock);
372 		drmHashDelete(bo->dev->handle_table, bo->handle);
373 		drmIoctl(bo->dev->fd, DRM_IOCTL_GEM_CLOSE, &req);
374 		pthread_mutex_unlock(&table_lock);
375 	}
376 
377 	omap_device_del(bo->dev);
378 
379 	free(bo);
380 }
381 
382 /* get the global flink/DRI2 buffer name */
omap_bo_get_name(struct omap_bo * bo,uint32_t * name)383 int omap_bo_get_name(struct omap_bo *bo, uint32_t *name)
384 {
385 	if (!bo->name) {
386 		struct drm_gem_flink req = {
387 				.handle = bo->handle,
388 		};
389 		int ret;
390 
391 		ret = drmIoctl(bo->dev->fd, DRM_IOCTL_GEM_FLINK, &req);
392 		if (ret) {
393 			return ret;
394 		}
395 
396 		bo->name = req.name;
397 	}
398 
399 	*name = bo->name;
400 
401 	return 0;
402 }
403 
omap_bo_handle(struct omap_bo * bo)404 uint32_t omap_bo_handle(struct omap_bo *bo)
405 {
406 	return bo->handle;
407 }
408 
409 /* caller owns the dmabuf fd that is returned and is responsible
410  * to close() it when done
411  */
omap_bo_dmabuf(struct omap_bo * bo)412 int omap_bo_dmabuf(struct omap_bo *bo)
413 {
414 	if (bo->fd < 0) {
415 		struct drm_prime_handle req = {
416 				.handle = bo->handle,
417 				.flags = DRM_CLOEXEC | DRM_RDWR,
418 		};
419 		int ret;
420 
421 		ret = drmIoctl(bo->dev->fd, DRM_IOCTL_PRIME_HANDLE_TO_FD, &req);
422 		if (ret) {
423 			return ret;
424 		}
425 
426 		bo->fd = req.fd;
427 	}
428 	return dup(bo->fd);
429 }
430 
omap_bo_size(struct omap_bo * bo)431 uint32_t omap_bo_size(struct omap_bo *bo)
432 {
433 	if (!bo->size) {
434 		get_buffer_info(bo);
435 	}
436 	return bo->size;
437 }
438 
omap_bo_map(struct omap_bo * bo)439 void *omap_bo_map(struct omap_bo *bo)
440 {
441 	if (!bo->map) {
442 		if (!bo->offset) {
443 			get_buffer_info(bo);
444 		}
445 
446 		bo->map = mmap(0, bo->size, PROT_READ | PROT_WRITE,
447 				MAP_SHARED, bo->dev->fd, bo->offset);
448 		if (bo->map == MAP_FAILED) {
449 			bo->map = NULL;
450 		}
451 	}
452 	return bo->map;
453 }
454 
omap_bo_cpu_prep(struct omap_bo * bo,enum omap_gem_op op)455 int omap_bo_cpu_prep(struct omap_bo *bo, enum omap_gem_op op)
456 {
457 	struct drm_omap_gem_cpu_prep req = {
458 			.handle = bo->handle,
459 			.op = op,
460 	};
461 	return drmCommandWrite(bo->dev->fd,
462 			DRM_OMAP_GEM_CPU_PREP, &req, sizeof(req));
463 }
464 
omap_bo_cpu_fini(struct omap_bo * bo,enum omap_gem_op op)465 int omap_bo_cpu_fini(struct omap_bo *bo, enum omap_gem_op op)
466 {
467 	struct drm_omap_gem_cpu_fini req = {
468 			.handle = bo->handle,
469 			.op = op,
470 			.nregions = 0,
471 	};
472 	return drmCommandWrite(bo->dev->fd,
473 			DRM_OMAP_GEM_CPU_FINI, &req, sizeof(req));
474 }
475