1 /**********************************************************
2 * Copyright 2009-2015 VMware, Inc. All rights reserved.
3 *
4 * Permission is hereby granted, free of charge, to any person
5 * obtaining a copy of this software and associated documentation
6 * files (the "Software"), to deal in the Software without
7 * restriction, including without limitation the rights to use, copy,
8 * modify, merge, publish, distribute, sublicense, and/or sell copies
9 * of the Software, and to permit persons to whom the Software is
10 * furnished to do so, subject to the following conditions:
11 *
12 * The above copyright notice and this permission notice shall be
13 * included in all copies or substantial portions of the Software.
14 *
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
16 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
17 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
18 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
19 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
20 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
21 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
22 * SOFTWARE.
23 *
24 **********************************************************/
25
26 /**
27 * @file
28 *
29 * Wrappers for DRM ioctl functionlaity used by the rest of the vmw
30 * drm winsys.
31 *
32 * Based on svgaicd_escape.c
33 */
34
35
36 #include "svga_cmd.h"
37 #include "util/u_memory.h"
38 #include "util/u_math.h"
39 #include "svgadump/svga_dump.h"
40 #include "frontend/drm_driver.h"
41 #include "vmw_screen.h"
42 #include "vmw_context.h"
43 #include "vmw_fence.h"
44 #include "xf86drm.h"
45 #include "vmwgfx_drm.h"
46 #include "svga3d_caps.h"
47 #include "svga3d_reg.h"
48
49 #include "os/os_mman.h"
50
51 #include <errno.h>
52 #include <unistd.h>
53
54 #define VMW_MAX_DEFAULT_TEXTURE_SIZE (128 * 1024 * 1024)
55 #define VMW_FENCE_TIMEOUT_SECONDS 3600UL
56
57 #define SVGA3D_FLAGS_64(upper32, lower32) (((uint64_t)upper32 << 32) | lower32)
58 #define SVGA3D_FLAGS_UPPER_32(svga3d_flags) (svga3d_flags >> 32)
59 #define SVGA3D_FLAGS_LOWER_32(svga3d_flags) \
60 (svga3d_flags & ((uint64_t)UINT32_MAX))
61
62 struct vmw_region
63 {
64 uint32_t handle;
65 uint64_t map_handle;
66 void *data;
67 uint32_t map_count;
68 int drm_fd;
69 uint32_t size;
70 };
71
72 uint32_t
vmw_region_size(struct vmw_region * region)73 vmw_region_size(struct vmw_region *region)
74 {
75 return region->size;
76 }
77
78 #if defined(__DragonFly__) || defined(__FreeBSD__) || \
79 defined(__NetBSD__) || defined(__OpenBSD__)
80 #define ERESTART EINTR
81 #endif
82
83 uint32
vmw_ioctl_context_create(struct vmw_winsys_screen * vws)84 vmw_ioctl_context_create(struct vmw_winsys_screen *vws)
85 {
86 struct drm_vmw_context_arg c_arg;
87 int ret;
88
89 VMW_FUNC;
90
91 ret = drmCommandRead(vws->ioctl.drm_fd, DRM_VMW_CREATE_CONTEXT,
92 &c_arg, sizeof(c_arg));
93
94 if (ret)
95 return -1;
96
97 vmw_printf("Context id is %d\n", c_arg.cid);
98 return c_arg.cid;
99 }
100
101 uint32
vmw_ioctl_extended_context_create(struct vmw_winsys_screen * vws,boolean vgpu10)102 vmw_ioctl_extended_context_create(struct vmw_winsys_screen *vws,
103 boolean vgpu10)
104 {
105 union drm_vmw_extended_context_arg c_arg;
106 int ret;
107
108 VMW_FUNC;
109 memset(&c_arg, 0, sizeof(c_arg));
110 c_arg.req = (vgpu10 ? drm_vmw_context_dx : drm_vmw_context_legacy);
111 ret = drmCommandWriteRead(vws->ioctl.drm_fd,
112 DRM_VMW_CREATE_EXTENDED_CONTEXT,
113 &c_arg, sizeof(c_arg));
114
115 if (ret)
116 return -1;
117
118 vmw_printf("Context id is %d\n", c_arg.cid);
119 return c_arg.rep.cid;
120 }
121
122 void
vmw_ioctl_context_destroy(struct vmw_winsys_screen * vws,uint32 cid)123 vmw_ioctl_context_destroy(struct vmw_winsys_screen *vws, uint32 cid)
124 {
125 struct drm_vmw_context_arg c_arg;
126
127 VMW_FUNC;
128
129 memset(&c_arg, 0, sizeof(c_arg));
130 c_arg.cid = cid;
131
132 (void)drmCommandWrite(vws->ioctl.drm_fd, DRM_VMW_UNREF_CONTEXT,
133 &c_arg, sizeof(c_arg));
134
135 }
136
137 uint32
vmw_ioctl_surface_create(struct vmw_winsys_screen * vws,SVGA3dSurface1Flags flags,SVGA3dSurfaceFormat format,unsigned usage,SVGA3dSize size,uint32_t numFaces,uint32_t numMipLevels,unsigned sampleCount)138 vmw_ioctl_surface_create(struct vmw_winsys_screen *vws,
139 SVGA3dSurface1Flags flags,
140 SVGA3dSurfaceFormat format,
141 unsigned usage,
142 SVGA3dSize size,
143 uint32_t numFaces, uint32_t numMipLevels,
144 unsigned sampleCount)
145 {
146 union drm_vmw_surface_create_arg s_arg;
147 struct drm_vmw_surface_create_req *req = &s_arg.req;
148 struct drm_vmw_surface_arg *rep = &s_arg.rep;
149 struct drm_vmw_size sizes[DRM_VMW_MAX_SURFACE_FACES*
150 DRM_VMW_MAX_MIP_LEVELS];
151 struct drm_vmw_size *cur_size;
152 uint32_t iFace;
153 uint32_t iMipLevel;
154 int ret;
155
156 vmw_printf("%s flags %d format %d\n", __FUNCTION__, flags, format);
157
158 memset(&s_arg, 0, sizeof(s_arg));
159 req->flags = (uint32_t) flags;
160 req->scanout = !!(usage & SVGA_SURFACE_USAGE_SCANOUT);
161 req->format = (uint32_t) format;
162 req->shareable = !!(usage & SVGA_SURFACE_USAGE_SHARED);
163
164 assert(numFaces * numMipLevels < DRM_VMW_MAX_SURFACE_FACES*
165 DRM_VMW_MAX_MIP_LEVELS);
166 cur_size = sizes;
167 for (iFace = 0; iFace < numFaces; ++iFace) {
168 SVGA3dSize mipSize = size;
169
170 req->mip_levels[iFace] = numMipLevels;
171 for (iMipLevel = 0; iMipLevel < numMipLevels; ++iMipLevel) {
172 cur_size->width = mipSize.width;
173 cur_size->height = mipSize.height;
174 cur_size->depth = mipSize.depth;
175 mipSize.width = MAX2(mipSize.width >> 1, 1);
176 mipSize.height = MAX2(mipSize.height >> 1, 1);
177 mipSize.depth = MAX2(mipSize.depth >> 1, 1);
178 cur_size++;
179 }
180 }
181 for (iFace = numFaces; iFace < SVGA3D_MAX_SURFACE_FACES; ++iFace) {
182 req->mip_levels[iFace] = 0;
183 }
184
185 req->size_addr = (unsigned long)&sizes;
186
187 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_CREATE_SURFACE,
188 &s_arg, sizeof(s_arg));
189
190 if (ret)
191 return -1;
192
193 vmw_printf("Surface id is %d\n", rep->sid);
194
195 return rep->sid;
196 }
197
198
199 uint32
vmw_ioctl_gb_surface_create(struct vmw_winsys_screen * vws,SVGA3dSurfaceAllFlags flags,SVGA3dSurfaceFormat format,unsigned usage,SVGA3dSize size,uint32_t numFaces,uint32_t numMipLevels,unsigned sampleCount,uint32_t buffer_handle,SVGA3dMSPattern multisamplePattern,SVGA3dMSQualityLevel qualityLevel,struct vmw_region ** p_region)200 vmw_ioctl_gb_surface_create(struct vmw_winsys_screen *vws,
201 SVGA3dSurfaceAllFlags flags,
202 SVGA3dSurfaceFormat format,
203 unsigned usage,
204 SVGA3dSize size,
205 uint32_t numFaces,
206 uint32_t numMipLevels,
207 unsigned sampleCount,
208 uint32_t buffer_handle,
209 SVGA3dMSPattern multisamplePattern,
210 SVGA3dMSQualityLevel qualityLevel,
211 struct vmw_region **p_region)
212 {
213 union {
214 union drm_vmw_gb_surface_create_ext_arg ext_arg;
215 union drm_vmw_gb_surface_create_arg arg;
216 } s_arg;
217 struct drm_vmw_gb_surface_create_rep *rep;
218 struct vmw_region *region = NULL;
219 int ret;
220
221 vmw_printf("%s flags %d format %d\n", __FUNCTION__, flags, format);
222
223 if (p_region) {
224 region = CALLOC_STRUCT(vmw_region);
225 if (!region)
226 return SVGA3D_INVALID_ID;
227 }
228
229 memset(&s_arg, 0, sizeof(s_arg));
230
231 if (vws->ioctl.have_drm_2_15) {
232 struct drm_vmw_gb_surface_create_ext_req *req = &s_arg.ext_arg.req;
233 rep = &s_arg.ext_arg.rep;
234
235 req->version = drm_vmw_gb_surface_v1;
236 req->multisample_pattern = multisamplePattern;
237 req->quality_level = qualityLevel;
238 req->buffer_byte_stride = 0;
239 req->must_be_zero = 0;
240 req->base.svga3d_flags = SVGA3D_FLAGS_LOWER_32(flags);
241 req->svga3d_flags_upper_32_bits = SVGA3D_FLAGS_UPPER_32(flags);
242 req->base.format = (uint32_t) format;
243
244 if (usage & SVGA_SURFACE_USAGE_SCANOUT)
245 req->base.drm_surface_flags |= drm_vmw_surface_flag_scanout;
246
247 if (usage & SVGA_SURFACE_USAGE_SHARED)
248 req->base.drm_surface_flags |= drm_vmw_surface_flag_shareable;
249
250 if ((usage & SVGA_SURFACE_USAGE_COHERENT) || vws->force_coherent)
251 req->base.drm_surface_flags |= drm_vmw_surface_flag_coherent;
252
253 req->base.drm_surface_flags |= drm_vmw_surface_flag_create_buffer;
254 req->base.base_size.width = size.width;
255 req->base.base_size.height = size.height;
256 req->base.base_size.depth = size.depth;
257 req->base.mip_levels = numMipLevels;
258 req->base.multisample_count = 0;
259 req->base.autogen_filter = SVGA3D_TEX_FILTER_NONE;
260
261 if (vws->base.have_vgpu10) {
262 req->base.array_size = numFaces;
263 req->base.multisample_count = sampleCount;
264 } else {
265 assert(numFaces * numMipLevels < DRM_VMW_MAX_SURFACE_FACES*
266 DRM_VMW_MAX_MIP_LEVELS);
267 req->base.array_size = 0;
268 }
269
270 req->base.buffer_handle = buffer_handle ?
271 buffer_handle : SVGA3D_INVALID_ID;
272
273 ret = drmCommandWriteRead(vws->ioctl.drm_fd,
274 DRM_VMW_GB_SURFACE_CREATE_EXT, &s_arg.ext_arg,
275 sizeof(s_arg.ext_arg));
276
277 if (ret)
278 goto out_fail_create;
279 } else {
280 struct drm_vmw_gb_surface_create_req *req = &s_arg.arg.req;
281 rep = &s_arg.arg.rep;
282
283 req->svga3d_flags = (uint32_t) flags;
284 req->format = (uint32_t) format;
285
286 if (usage & SVGA_SURFACE_USAGE_SCANOUT)
287 req->drm_surface_flags |= drm_vmw_surface_flag_scanout;
288
289 if (usage & SVGA_SURFACE_USAGE_SHARED)
290 req->drm_surface_flags |= drm_vmw_surface_flag_shareable;
291
292 req->drm_surface_flags |= drm_vmw_surface_flag_create_buffer;
293 req->base_size.width = size.width;
294 req->base_size.height = size.height;
295 req->base_size.depth = size.depth;
296 req->mip_levels = numMipLevels;
297 req->multisample_count = 0;
298 req->autogen_filter = SVGA3D_TEX_FILTER_NONE;
299
300 if (vws->base.have_vgpu10) {
301 req->array_size = numFaces;
302 req->multisample_count = sampleCount;
303 } else {
304 assert(numFaces * numMipLevels < DRM_VMW_MAX_SURFACE_FACES*
305 DRM_VMW_MAX_MIP_LEVELS);
306 req->array_size = 0;
307 }
308
309 req->buffer_handle = buffer_handle ?
310 buffer_handle : SVGA3D_INVALID_ID;
311
312 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GB_SURFACE_CREATE,
313 &s_arg.arg, sizeof(s_arg.arg));
314
315 if (ret)
316 goto out_fail_create;
317 }
318
319 if (p_region) {
320 region->handle = rep->buffer_handle;
321 region->map_handle = rep->buffer_map_handle;
322 region->drm_fd = vws->ioctl.drm_fd;
323 region->size = rep->backup_size;
324 *p_region = region;
325 }
326
327 vmw_printf("Surface id is %d\n", rep->sid);
328 return rep->handle;
329
330 out_fail_create:
331 FREE(region);
332 return SVGA3D_INVALID_ID;
333 }
334
335 /**
336 * vmw_ioctl_surface_req - Fill in a struct surface_req
337 *
338 * @vws: Winsys screen
339 * @whandle: Surface handle
340 * @req: The struct surface req to fill in
341 * @needs_unref: This call takes a kernel surface reference that needs to
342 * be unreferenced.
343 *
344 * Returns 0 on success, negative error type otherwise.
345 * Fills in the surface_req structure according to handle type and kernel
346 * capabilities.
347 */
348 static int
vmw_ioctl_surface_req(const struct vmw_winsys_screen * vws,const struct winsys_handle * whandle,struct drm_vmw_surface_arg * req,boolean * needs_unref)349 vmw_ioctl_surface_req(const struct vmw_winsys_screen *vws,
350 const struct winsys_handle *whandle,
351 struct drm_vmw_surface_arg *req,
352 boolean *needs_unref)
353 {
354 int ret;
355
356 switch(whandle->type) {
357 case WINSYS_HANDLE_TYPE_SHARED:
358 case WINSYS_HANDLE_TYPE_KMS:
359 *needs_unref = FALSE;
360 req->handle_type = DRM_VMW_HANDLE_LEGACY;
361 req->sid = whandle->handle;
362 break;
363 case WINSYS_HANDLE_TYPE_FD:
364 if (!vws->ioctl.have_drm_2_6) {
365 uint32_t handle;
366
367 ret = drmPrimeFDToHandle(vws->ioctl.drm_fd, whandle->handle, &handle);
368 if (ret) {
369 vmw_error("Failed to get handle from prime fd %d.\n",
370 (int) whandle->handle);
371 return -EINVAL;
372 }
373
374 *needs_unref = TRUE;
375 req->handle_type = DRM_VMW_HANDLE_LEGACY;
376 req->sid = handle;
377 } else {
378 *needs_unref = FALSE;
379 req->handle_type = DRM_VMW_HANDLE_PRIME;
380 req->sid = whandle->handle;
381 }
382 break;
383 default:
384 vmw_error("Attempt to import unsupported handle type %d.\n",
385 whandle->type);
386 return -EINVAL;
387 }
388
389 return 0;
390 }
391
392 /**
393 * vmw_ioctl_gb_surface_ref - Put a reference on a guest-backed surface and
394 * get surface information
395 *
396 * @vws: Screen to register the reference on
397 * @handle: Kernel handle of the guest-backed surface
398 * @flags: flags used when the surface was created
399 * @format: Format used when the surface was created
400 * @numMipLevels: Number of mipmap levels of the surface
401 * @p_region: On successful return points to a newly allocated
402 * struct vmw_region holding a reference to the surface backup buffer.
403 *
404 * Returns 0 on success, a system error on failure.
405 */
406 int
vmw_ioctl_gb_surface_ref(struct vmw_winsys_screen * vws,const struct winsys_handle * whandle,SVGA3dSurfaceAllFlags * flags,SVGA3dSurfaceFormat * format,uint32_t * numMipLevels,uint32_t * handle,struct vmw_region ** p_region)407 vmw_ioctl_gb_surface_ref(struct vmw_winsys_screen *vws,
408 const struct winsys_handle *whandle,
409 SVGA3dSurfaceAllFlags *flags,
410 SVGA3dSurfaceFormat *format,
411 uint32_t *numMipLevels,
412 uint32_t *handle,
413 struct vmw_region **p_region)
414 {
415 struct vmw_region *region = NULL;
416 boolean needs_unref = FALSE;
417 int ret;
418
419 assert(p_region != NULL);
420 region = CALLOC_STRUCT(vmw_region);
421 if (!region)
422 return -ENOMEM;
423
424 if (vws->ioctl.have_drm_2_15) {
425 union drm_vmw_gb_surface_reference_ext_arg s_arg;
426 struct drm_vmw_surface_arg *req = &s_arg.req;
427 struct drm_vmw_gb_surface_ref_ext_rep *rep = &s_arg.rep;
428
429 memset(&s_arg, 0, sizeof(s_arg));
430 ret = vmw_ioctl_surface_req(vws, whandle, req, &needs_unref);
431 if (ret)
432 goto out_fail_req;
433
434 *handle = req->sid;
435 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GB_SURFACE_REF_EXT,
436 &s_arg, sizeof(s_arg));
437
438 if (ret)
439 goto out_fail_ref;
440
441 region->handle = rep->crep.buffer_handle;
442 region->map_handle = rep->crep.buffer_map_handle;
443 region->drm_fd = vws->ioctl.drm_fd;
444 region->size = rep->crep.backup_size;
445 *p_region = region;
446
447 *handle = rep->crep.handle;
448 *flags = SVGA3D_FLAGS_64(rep->creq.svga3d_flags_upper_32_bits,
449 rep->creq.base.svga3d_flags);
450 *format = rep->creq.base.format;
451 *numMipLevels = rep->creq.base.mip_levels;
452 } else {
453 union drm_vmw_gb_surface_reference_arg s_arg;
454 struct drm_vmw_surface_arg *req = &s_arg.req;
455 struct drm_vmw_gb_surface_ref_rep *rep = &s_arg.rep;
456
457 memset(&s_arg, 0, sizeof(s_arg));
458 ret = vmw_ioctl_surface_req(vws, whandle, req, &needs_unref);
459 if (ret)
460 goto out_fail_req;
461
462 *handle = req->sid;
463 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GB_SURFACE_REF,
464 &s_arg, sizeof(s_arg));
465
466 if (ret)
467 goto out_fail_ref;
468
469 region->handle = rep->crep.buffer_handle;
470 region->map_handle = rep->crep.buffer_map_handle;
471 region->drm_fd = vws->ioctl.drm_fd;
472 region->size = rep->crep.backup_size;
473 *p_region = region;
474
475 *handle = rep->crep.handle;
476 *flags = rep->creq.svga3d_flags;
477 *format = rep->creq.format;
478 *numMipLevels = rep->creq.mip_levels;
479 }
480
481 vmw_printf("%s flags %d format %d\n", __FUNCTION__, *flags, *format);
482
483 if (needs_unref)
484 vmw_ioctl_surface_destroy(vws, *handle);
485
486 return 0;
487 out_fail_ref:
488 if (needs_unref)
489 vmw_ioctl_surface_destroy(vws, *handle);
490 out_fail_req:
491 FREE(region);
492 return ret;
493 }
494
495 void
vmw_ioctl_surface_destroy(struct vmw_winsys_screen * vws,uint32 sid)496 vmw_ioctl_surface_destroy(struct vmw_winsys_screen *vws, uint32 sid)
497 {
498 struct drm_vmw_surface_arg s_arg;
499
500 VMW_FUNC;
501
502 memset(&s_arg, 0, sizeof(s_arg));
503 s_arg.sid = sid;
504
505 (void)drmCommandWrite(vws->ioctl.drm_fd, DRM_VMW_UNREF_SURFACE,
506 &s_arg, sizeof(s_arg));
507 }
508
509 void
vmw_ioctl_command(struct vmw_winsys_screen * vws,int32_t cid,uint32_t throttle_us,void * commands,uint32_t size,struct pipe_fence_handle ** pfence,int32_t imported_fence_fd,uint32_t flags)510 vmw_ioctl_command(struct vmw_winsys_screen *vws, int32_t cid,
511 uint32_t throttle_us, void *commands, uint32_t size,
512 struct pipe_fence_handle **pfence, int32_t imported_fence_fd,
513 uint32_t flags)
514 {
515 struct drm_vmw_execbuf_arg arg;
516 struct drm_vmw_fence_rep rep;
517 int ret;
518 int argsize;
519
520 #ifdef DEBUG
521 {
522 static boolean firsttime = TRUE;
523 static boolean debug = FALSE;
524 static boolean skip = FALSE;
525 if (firsttime) {
526 debug = debug_get_bool_option("SVGA_DUMP_CMD", FALSE);
527 skip = debug_get_bool_option("SVGA_SKIP_CMD", FALSE);
528 }
529 if (debug) {
530 VMW_FUNC;
531 svga_dump_commands(commands, size);
532 }
533 firsttime = FALSE;
534 if (skip) {
535 size = 0;
536 }
537 }
538 #endif
539
540 memset(&arg, 0, sizeof(arg));
541 memset(&rep, 0, sizeof(rep));
542
543 if (flags & SVGA_HINT_FLAG_EXPORT_FENCE_FD) {
544 arg.flags |= DRM_VMW_EXECBUF_FLAG_EXPORT_FENCE_FD;
545 }
546
547 if (imported_fence_fd != -1) {
548 arg.flags |= DRM_VMW_EXECBUF_FLAG_IMPORT_FENCE_FD;
549 }
550
551 rep.error = -EFAULT;
552 if (pfence)
553 arg.fence_rep = (unsigned long)&rep;
554 arg.commands = (unsigned long)commands;
555 arg.command_size = size;
556 arg.throttle_us = throttle_us;
557 arg.version = vws->ioctl.drm_execbuf_version;
558 arg.context_handle = (vws->base.have_vgpu10 ? cid : SVGA3D_INVALID_ID);
559
560 /* Older DRM module requires this to be zero */
561 if (vws->base.have_fence_fd)
562 arg.imported_fence_fd = imported_fence_fd;
563
564 /* In DRM_VMW_EXECBUF_VERSION 1, the drm_vmw_execbuf_arg structure ends with
565 * the flags field. The structure size sent to drmCommandWrite must match
566 * the drm_execbuf_version. Otherwise, an invalid value will be returned.
567 */
568 argsize = vws->ioctl.drm_execbuf_version > 1 ? sizeof(arg) :
569 offsetof(struct drm_vmw_execbuf_arg, context_handle);
570 do {
571 ret = drmCommandWrite(vws->ioctl.drm_fd, DRM_VMW_EXECBUF, &arg, argsize);
572 if (ret == -EBUSY)
573 usleep(1000);
574 } while(ret == -ERESTART || ret == -EBUSY);
575 if (ret) {
576 vmw_error("%s error %s.\n", __FUNCTION__, strerror(-ret));
577 abort();
578 }
579
580 if (rep.error) {
581
582 /*
583 * Kernel has already synced, or caller requested no fence.
584 */
585 if (pfence)
586 *pfence = NULL;
587 } else {
588 if (pfence) {
589 vmw_fences_signal(vws->fence_ops, rep.passed_seqno, rep.seqno,
590 TRUE);
591
592 /* Older DRM module will set this to zero, but -1 is the proper FD
593 * to use for no Fence FD support */
594 if (!vws->base.have_fence_fd)
595 rep.fd = -1;
596
597 *pfence = vmw_fence_create(vws->fence_ops, rep.handle,
598 rep.seqno, rep.mask, rep.fd);
599 if (*pfence == NULL) {
600 /*
601 * Fence creation failed. Need to sync.
602 */
603 (void) vmw_ioctl_fence_finish(vws, rep.handle, rep.mask);
604 vmw_ioctl_fence_unref(vws, rep.handle);
605 }
606 }
607 }
608 }
609
610
611 struct vmw_region *
vmw_ioctl_region_create(struct vmw_winsys_screen * vws,uint32_t size)612 vmw_ioctl_region_create(struct vmw_winsys_screen *vws, uint32_t size)
613 {
614 struct vmw_region *region;
615 union drm_vmw_alloc_dmabuf_arg arg;
616 struct drm_vmw_alloc_dmabuf_req *req = &arg.req;
617 struct drm_vmw_dmabuf_rep *rep = &arg.rep;
618 int ret;
619
620 vmw_printf("%s: size = %u\n", __FUNCTION__, size);
621
622 region = CALLOC_STRUCT(vmw_region);
623 if (!region)
624 goto out_err1;
625
626 memset(&arg, 0, sizeof(arg));
627 req->size = size;
628 do {
629 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_ALLOC_DMABUF, &arg,
630 sizeof(arg));
631 } while (ret == -ERESTART);
632
633 if (ret) {
634 vmw_error("IOCTL failed %d: %s\n", ret, strerror(-ret));
635 goto out_err1;
636 }
637
638 region->data = NULL;
639 region->handle = rep->handle;
640 region->map_handle = rep->map_handle;
641 region->map_count = 0;
642 region->size = size;
643 region->drm_fd = vws->ioctl.drm_fd;
644
645 vmw_printf(" gmrId = %u, offset = %u\n",
646 region->ptr.gmrId, region->ptr.offset);
647
648 return region;
649
650 out_err1:
651 FREE(region);
652 return NULL;
653 }
654
655 void
vmw_ioctl_region_destroy(struct vmw_region * region)656 vmw_ioctl_region_destroy(struct vmw_region *region)
657 {
658 struct drm_vmw_unref_dmabuf_arg arg;
659
660 vmw_printf("%s: gmrId = %u, offset = %u\n", __FUNCTION__,
661 region->ptr.gmrId, region->ptr.offset);
662
663 if (region->data) {
664 os_munmap(region->data, region->size);
665 region->data = NULL;
666 }
667
668 memset(&arg, 0, sizeof(arg));
669 arg.handle = region->handle;
670 drmCommandWrite(region->drm_fd, DRM_VMW_UNREF_DMABUF, &arg, sizeof(arg));
671
672 FREE(region);
673 }
674
675 SVGAGuestPtr
vmw_ioctl_region_ptr(struct vmw_region * region)676 vmw_ioctl_region_ptr(struct vmw_region *region)
677 {
678 SVGAGuestPtr ptr = {region->handle, 0};
679 return ptr;
680 }
681
682 void *
vmw_ioctl_region_map(struct vmw_region * region)683 vmw_ioctl_region_map(struct vmw_region *region)
684 {
685 void *map;
686
687 vmw_printf("%s: gmrId = %u, offset = %u\n", __FUNCTION__,
688 region->ptr.gmrId, region->ptr.offset);
689
690 if (region->data == NULL) {
691 map = os_mmap(NULL, region->size, PROT_READ | PROT_WRITE, MAP_SHARED,
692 region->drm_fd, region->map_handle);
693 if (map == MAP_FAILED) {
694 vmw_error("%s: Map failed.\n", __FUNCTION__);
695 return NULL;
696 }
697
698 // MADV_HUGEPAGE only exists on Linux
699 #ifdef MADV_HUGEPAGE
700 (void) madvise(map, region->size, MADV_HUGEPAGE);
701 #endif
702 region->data = map;
703 }
704
705 ++region->map_count;
706
707 return region->data;
708 }
709
710 void
vmw_ioctl_region_unmap(struct vmw_region * region)711 vmw_ioctl_region_unmap(struct vmw_region *region)
712 {
713 vmw_printf("%s: gmrId = %u, offset = %u\n", __FUNCTION__,
714 region->ptr.gmrId, region->ptr.offset);
715
716 --region->map_count;
717 os_munmap(region->data, region->size);
718 region->data = NULL;
719 }
720
721 /**
722 * vmw_ioctl_syncforcpu - Synchronize a buffer object for CPU usage
723 *
724 * @region: Pointer to a struct vmw_region representing the buffer object.
725 * @dont_block: Dont wait for GPU idle, but rather return -EBUSY if the
726 * GPU is busy with the buffer object.
727 * @readonly: Hint that the CPU access is read-only.
728 * @allow_cs: Allow concurrent command submission while the buffer is
729 * synchronized for CPU. If FALSE command submissions referencing the
730 * buffer will block until a corresponding call to vmw_ioctl_releasefromcpu.
731 *
732 * This function idles any GPU activities touching the buffer and blocks
733 * command submission of commands referencing the buffer, even from
734 * other processes.
735 */
736 int
vmw_ioctl_syncforcpu(struct vmw_region * region,boolean dont_block,boolean readonly,boolean allow_cs)737 vmw_ioctl_syncforcpu(struct vmw_region *region,
738 boolean dont_block,
739 boolean readonly,
740 boolean allow_cs)
741 {
742 struct drm_vmw_synccpu_arg arg;
743
744 memset(&arg, 0, sizeof(arg));
745 arg.op = drm_vmw_synccpu_grab;
746 arg.handle = region->handle;
747 arg.flags = drm_vmw_synccpu_read;
748 if (!readonly)
749 arg.flags |= drm_vmw_synccpu_write;
750 if (dont_block)
751 arg.flags |= drm_vmw_synccpu_dontblock;
752 if (allow_cs)
753 arg.flags |= drm_vmw_synccpu_allow_cs;
754
755 return drmCommandWrite(region->drm_fd, DRM_VMW_SYNCCPU, &arg, sizeof(arg));
756 }
757
758 /**
759 * vmw_ioctl_releasefromcpu - Undo a previous syncforcpu.
760 *
761 * @region: Pointer to a struct vmw_region representing the buffer object.
762 * @readonly: Should hold the same value as the matching syncforcpu call.
763 * @allow_cs: Should hold the same value as the matching syncforcpu call.
764 */
765 void
vmw_ioctl_releasefromcpu(struct vmw_region * region,boolean readonly,boolean allow_cs)766 vmw_ioctl_releasefromcpu(struct vmw_region *region,
767 boolean readonly,
768 boolean allow_cs)
769 {
770 struct drm_vmw_synccpu_arg arg;
771
772 memset(&arg, 0, sizeof(arg));
773 arg.op = drm_vmw_synccpu_release;
774 arg.handle = region->handle;
775 arg.flags = drm_vmw_synccpu_read;
776 if (!readonly)
777 arg.flags |= drm_vmw_synccpu_write;
778 if (allow_cs)
779 arg.flags |= drm_vmw_synccpu_allow_cs;
780
781 (void) drmCommandWrite(region->drm_fd, DRM_VMW_SYNCCPU, &arg, sizeof(arg));
782 }
783
784 void
vmw_ioctl_fence_unref(struct vmw_winsys_screen * vws,uint32_t handle)785 vmw_ioctl_fence_unref(struct vmw_winsys_screen *vws,
786 uint32_t handle)
787 {
788 struct drm_vmw_fence_arg arg;
789 int ret;
790
791 memset(&arg, 0, sizeof(arg));
792 arg.handle = handle;
793
794 ret = drmCommandWrite(vws->ioctl.drm_fd, DRM_VMW_FENCE_UNREF,
795 &arg, sizeof(arg));
796 if (ret != 0)
797 vmw_error("%s Failed\n", __FUNCTION__);
798 }
799
800 static inline uint32_t
vmw_drm_fence_flags(uint32_t flags)801 vmw_drm_fence_flags(uint32_t flags)
802 {
803 uint32_t dflags = 0;
804
805 if (flags & SVGA_FENCE_FLAG_EXEC)
806 dflags |= DRM_VMW_FENCE_FLAG_EXEC;
807 if (flags & SVGA_FENCE_FLAG_QUERY)
808 dflags |= DRM_VMW_FENCE_FLAG_QUERY;
809
810 return dflags;
811 }
812
813
814 int
vmw_ioctl_fence_signalled(struct vmw_winsys_screen * vws,uint32_t handle,uint32_t flags)815 vmw_ioctl_fence_signalled(struct vmw_winsys_screen *vws,
816 uint32_t handle,
817 uint32_t flags)
818 {
819 struct drm_vmw_fence_signaled_arg arg;
820 uint32_t vflags = vmw_drm_fence_flags(flags);
821 int ret;
822
823 memset(&arg, 0, sizeof(arg));
824 arg.handle = handle;
825 arg.flags = vflags;
826
827 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_FENCE_SIGNALED,
828 &arg, sizeof(arg));
829
830 if (ret != 0)
831 return ret;
832
833 vmw_fences_signal(vws->fence_ops, arg.passed_seqno, 0, FALSE);
834
835 return (arg.signaled) ? 0 : -1;
836 }
837
838
839
840 int
vmw_ioctl_fence_finish(struct vmw_winsys_screen * vws,uint32_t handle,uint32_t flags)841 vmw_ioctl_fence_finish(struct vmw_winsys_screen *vws,
842 uint32_t handle,
843 uint32_t flags)
844 {
845 struct drm_vmw_fence_wait_arg arg;
846 uint32_t vflags = vmw_drm_fence_flags(flags);
847 int ret;
848
849 memset(&arg, 0, sizeof(arg));
850
851 arg.handle = handle;
852 arg.timeout_us = VMW_FENCE_TIMEOUT_SECONDS*1000000;
853 arg.lazy = 0;
854 arg.flags = vflags;
855
856 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_FENCE_WAIT,
857 &arg, sizeof(arg));
858
859 if (ret != 0)
860 vmw_error("%s Failed\n", __FUNCTION__);
861
862 return 0;
863 }
864
865 uint32
vmw_ioctl_shader_create(struct vmw_winsys_screen * vws,SVGA3dShaderType type,uint32 code_len)866 vmw_ioctl_shader_create(struct vmw_winsys_screen *vws,
867 SVGA3dShaderType type,
868 uint32 code_len)
869 {
870 struct drm_vmw_shader_create_arg sh_arg;
871 int ret;
872
873 VMW_FUNC;
874
875 memset(&sh_arg, 0, sizeof(sh_arg));
876
877 sh_arg.size = code_len;
878 sh_arg.buffer_handle = SVGA3D_INVALID_ID;
879 sh_arg.shader_handle = SVGA3D_INVALID_ID;
880 switch (type) {
881 case SVGA3D_SHADERTYPE_VS:
882 sh_arg.shader_type = drm_vmw_shader_type_vs;
883 break;
884 case SVGA3D_SHADERTYPE_PS:
885 sh_arg.shader_type = drm_vmw_shader_type_ps;
886 break;
887 default:
888 assert(!"Invalid shader type.");
889 break;
890 }
891
892 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_CREATE_SHADER,
893 &sh_arg, sizeof(sh_arg));
894
895 if (ret)
896 return SVGA3D_INVALID_ID;
897
898 return sh_arg.shader_handle;
899 }
900
901 void
vmw_ioctl_shader_destroy(struct vmw_winsys_screen * vws,uint32 shid)902 vmw_ioctl_shader_destroy(struct vmw_winsys_screen *vws, uint32 shid)
903 {
904 struct drm_vmw_shader_arg sh_arg;
905
906 VMW_FUNC;
907
908 memset(&sh_arg, 0, sizeof(sh_arg));
909 sh_arg.handle = shid;
910
911 (void)drmCommandWrite(vws->ioctl.drm_fd, DRM_VMW_UNREF_SHADER,
912 &sh_arg, sizeof(sh_arg));
913
914 }
915
916 static int
vmw_ioctl_parse_caps(struct vmw_winsys_screen * vws,const uint32_t * cap_buffer)917 vmw_ioctl_parse_caps(struct vmw_winsys_screen *vws,
918 const uint32_t *cap_buffer)
919 {
920 int i;
921
922 if (vws->base.have_gb_objects) {
923 for (i = 0; i < vws->ioctl.num_cap_3d; ++i) {
924 vws->ioctl.cap_3d[i].has_cap = TRUE;
925 vws->ioctl.cap_3d[i].result.u = cap_buffer[i];
926 }
927 return 0;
928 } else {
929 const uint32 *capsBlock;
930 const SVGA3dCapsRecord *capsRecord = NULL;
931 uint32 offset;
932 const SVGA3dCapPair *capArray;
933 int numCaps, index;
934
935 /*
936 * Search linearly through the caps block records for the specified type.
937 */
938 capsBlock = cap_buffer;
939 for (offset = 0; capsBlock[offset] != 0; offset += capsBlock[offset]) {
940 const SVGA3dCapsRecord *record;
941 assert(offset < SVGA_FIFO_3D_CAPS_SIZE);
942 record = (const SVGA3dCapsRecord *) (capsBlock + offset);
943 if ((record->header.type >= SVGA3DCAPS_RECORD_DEVCAPS_MIN) &&
944 (record->header.type <= SVGA3DCAPS_RECORD_DEVCAPS_MAX) &&
945 (!capsRecord || (record->header.type > capsRecord->header.type))) {
946 capsRecord = record;
947 }
948 }
949
950 if(!capsRecord)
951 return -1;
952
953 /*
954 * Calculate the number of caps from the size of the record.
955 */
956 capArray = (const SVGA3dCapPair *) capsRecord->data;
957 numCaps = (int) ((capsRecord->header.length * sizeof(uint32) -
958 sizeof capsRecord->header) / (2 * sizeof(uint32)));
959
960 for (i = 0; i < numCaps; i++) {
961 index = capArray[i][0];
962 if (index < vws->ioctl.num_cap_3d) {
963 vws->ioctl.cap_3d[index].has_cap = TRUE;
964 vws->ioctl.cap_3d[index].result.u = capArray[i][1];
965 } else {
966 debug_printf("Unknown devcaps seen: %d\n", index);
967 }
968 }
969 }
970 return 0;
971 }
972
973 boolean
vmw_ioctl_init(struct vmw_winsys_screen * vws)974 vmw_ioctl_init(struct vmw_winsys_screen *vws)
975 {
976 struct drm_vmw_getparam_arg gp_arg;
977 struct drm_vmw_get_3d_cap_arg cap_arg;
978 unsigned int size;
979 int ret;
980 uint32_t *cap_buffer;
981 drmVersionPtr version;
982 boolean drm_gb_capable;
983 boolean have_drm_2_5;
984 const char *getenv_val;
985
986 VMW_FUNC;
987
988 version = drmGetVersion(vws->ioctl.drm_fd);
989 if (!version)
990 goto out_no_version;
991
992 have_drm_2_5 = version->version_major > 2 ||
993 (version->version_major == 2 && version->version_minor > 4);
994 vws->ioctl.have_drm_2_6 = version->version_major > 2 ||
995 (version->version_major == 2 && version->version_minor > 5);
996 vws->ioctl.have_drm_2_9 = version->version_major > 2 ||
997 (version->version_major == 2 && version->version_minor > 8);
998 vws->ioctl.have_drm_2_15 = version->version_major > 2 ||
999 (version->version_major == 2 && version->version_minor > 14);
1000 vws->ioctl.have_drm_2_16 = version->version_major > 2 ||
1001 (version->version_major == 2 && version->version_minor > 15);
1002 vws->ioctl.have_drm_2_17 = version->version_major > 2 ||
1003 (version->version_major == 2 && version->version_minor > 16);
1004 vws->ioctl.have_drm_2_18 = version->version_major > 2 ||
1005 (version->version_major == 2 && version->version_minor > 17);
1006
1007 vws->ioctl.drm_execbuf_version = vws->ioctl.have_drm_2_9 ? 2 : 1;
1008
1009 drm_gb_capable = have_drm_2_5;
1010
1011 memset(&gp_arg, 0, sizeof(gp_arg));
1012 gp_arg.param = DRM_VMW_PARAM_3D;
1013 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
1014 &gp_arg, sizeof(gp_arg));
1015 if (ret || gp_arg.value == 0) {
1016 vmw_error("No 3D enabled (%i, %s).\n", ret, strerror(-ret));
1017 goto out_no_3d;
1018 }
1019
1020 memset(&gp_arg, 0, sizeof(gp_arg));
1021 gp_arg.param = DRM_VMW_PARAM_FIFO_HW_VERSION;
1022 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
1023 &gp_arg, sizeof(gp_arg));
1024 if (ret) {
1025 vmw_error("Failed to get fifo hw version (%i, %s).\n",
1026 ret, strerror(-ret));
1027 goto out_no_3d;
1028 }
1029 vws->ioctl.hwversion = gp_arg.value;
1030 getenv_val = getenv("SVGA_FORCE_HOST_BACKED");
1031 if (!getenv_val || strcmp(getenv_val, "0") == 0) {
1032 memset(&gp_arg, 0, sizeof(gp_arg));
1033 gp_arg.param = DRM_VMW_PARAM_HW_CAPS;
1034 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
1035 &gp_arg, sizeof(gp_arg));
1036 } else {
1037 ret = -EINVAL;
1038 }
1039 if (ret)
1040 vws->base.have_gb_objects = FALSE;
1041 else
1042 vws->base.have_gb_objects =
1043 !!(gp_arg.value & (uint64_t) SVGA_CAP_GBOBJECTS);
1044
1045 if (vws->base.have_gb_objects && !drm_gb_capable)
1046 goto out_no_3d;
1047
1048 vws->base.have_vgpu10 = FALSE;
1049 vws->base.have_sm4_1 = FALSE;
1050 vws->base.have_intra_surface_copy = FALSE;
1051
1052 if (vws->base.have_gb_objects) {
1053 memset(&gp_arg, 0, sizeof(gp_arg));
1054 gp_arg.param = DRM_VMW_PARAM_MAX_MOB_MEMORY;
1055 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
1056 &gp_arg, sizeof(gp_arg));
1057 if (ret) {
1058 /* Just guess a large enough value. */
1059 vws->ioctl.max_mob_memory = 256*1024*1024;
1060 } else {
1061 vws->ioctl.max_mob_memory = gp_arg.value;
1062 }
1063
1064 memset(&gp_arg, 0, sizeof(gp_arg));
1065 gp_arg.param = DRM_VMW_PARAM_MAX_MOB_SIZE;
1066 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
1067 &gp_arg, sizeof(gp_arg));
1068
1069 if (ret || gp_arg.value == 0) {
1070 vws->ioctl.max_texture_size = VMW_MAX_DEFAULT_TEXTURE_SIZE;
1071 } else {
1072 vws->ioctl.max_texture_size = gp_arg.value;
1073 }
1074
1075 /* Never early flush surfaces, mobs do accounting. */
1076 vws->ioctl.max_surface_memory = -1;
1077
1078 if (vws->ioctl.have_drm_2_9) {
1079 memset(&gp_arg, 0, sizeof(gp_arg));
1080 gp_arg.param = DRM_VMW_PARAM_DX;
1081 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
1082 &gp_arg, sizeof(gp_arg));
1083 if (ret == 0 && gp_arg.value != 0) {
1084 const char *vgpu10_val;
1085
1086 debug_printf("Have VGPU10 interface and hardware.\n");
1087 vws->base.have_vgpu10 = TRUE;
1088 vgpu10_val = getenv("SVGA_VGPU10");
1089 if (vgpu10_val && strcmp(vgpu10_val, "0") == 0) {
1090 debug_printf("Disabling VGPU10 interface.\n");
1091 vws->base.have_vgpu10 = FALSE;
1092 } else {
1093 debug_printf("Enabling VGPU10 interface.\n");
1094 }
1095 }
1096 }
1097
1098 if (vws->ioctl.have_drm_2_15 && vws->base.have_vgpu10) {
1099 memset(&gp_arg, 0, sizeof(gp_arg));
1100 gp_arg.param = DRM_VMW_PARAM_HW_CAPS2;
1101 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
1102 &gp_arg, sizeof(gp_arg));
1103 if (ret == 0 && gp_arg.value != 0) {
1104 vws->base.have_intra_surface_copy = TRUE;
1105 }
1106
1107 memset(&gp_arg, 0, sizeof(gp_arg));
1108 gp_arg.param = DRM_VMW_PARAM_SM4_1;
1109 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
1110 &gp_arg, sizeof(gp_arg));
1111 if (ret == 0 && gp_arg.value != 0) {
1112 vws->base.have_sm4_1 = TRUE;
1113 }
1114 }
1115
1116 if (vws->ioctl.have_drm_2_18 && vws->base.have_sm4_1) {
1117 memset(&gp_arg, 0, sizeof(gp_arg));
1118 gp_arg.param = DRM_VMW_PARAM_SM5;
1119 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
1120 &gp_arg, sizeof(gp_arg));
1121 if (ret == 0 && gp_arg.value != 0) {
1122 vws->base.have_sm5 = TRUE;
1123 }
1124 }
1125
1126 memset(&gp_arg, 0, sizeof(gp_arg));
1127 gp_arg.param = DRM_VMW_PARAM_3D_CAPS_SIZE;
1128 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
1129 &gp_arg, sizeof(gp_arg));
1130 if (ret)
1131 size = SVGA_FIFO_3D_CAPS_SIZE * sizeof(uint32_t);
1132 else
1133 size = gp_arg.value;
1134
1135 if (vws->base.have_gb_objects)
1136 vws->ioctl.num_cap_3d = size / sizeof(uint32_t);
1137 else
1138 vws->ioctl.num_cap_3d = SVGA3D_DEVCAP_MAX;
1139
1140 if (vws->ioctl.have_drm_2_16) {
1141 vws->base.have_coherent = TRUE;
1142 getenv_val = getenv("SVGA_FORCE_COHERENT");
1143 if (getenv_val && strcmp(getenv_val, "0") != 0)
1144 vws->force_coherent = TRUE;
1145 }
1146 } else {
1147 vws->ioctl.num_cap_3d = SVGA3D_DEVCAP_MAX;
1148
1149 memset(&gp_arg, 0, sizeof(gp_arg));
1150 gp_arg.param = DRM_VMW_PARAM_MAX_SURF_MEMORY;
1151 if (have_drm_2_5)
1152 ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
1153 &gp_arg, sizeof(gp_arg));
1154 if (!have_drm_2_5 || ret) {
1155 /* Just guess a large enough value, around 800mb. */
1156 vws->ioctl.max_surface_memory = 0x30000000;
1157 } else {
1158 vws->ioctl.max_surface_memory = gp_arg.value;
1159 }
1160
1161 vws->ioctl.max_texture_size = VMW_MAX_DEFAULT_TEXTURE_SIZE;
1162
1163 size = SVGA_FIFO_3D_CAPS_SIZE * sizeof(uint32_t);
1164 }
1165
1166 debug_printf("VGPU10 interface is %s.\n",
1167 vws->base.have_vgpu10 ? "on" : "off");
1168
1169 cap_buffer = calloc(1, size);
1170 if (!cap_buffer) {
1171 debug_printf("Failed alloc fifo 3D caps buffer.\n");
1172 goto out_no_3d;
1173 }
1174
1175 vws->ioctl.cap_3d = calloc(vws->ioctl.num_cap_3d,
1176 sizeof(*vws->ioctl.cap_3d));
1177 if (!vws->ioctl.cap_3d) {
1178 debug_printf("Failed alloc fifo 3D caps buffer.\n");
1179 goto out_no_caparray;
1180 }
1181
1182 memset(&cap_arg, 0, sizeof(cap_arg));
1183 cap_arg.buffer = (uint64_t) (unsigned long) (cap_buffer);
1184 cap_arg.max_size = size;
1185
1186 /*
1187 * This call must always be after DRM_VMW_PARAM_MAX_MOB_MEMORY and
1188 * DRM_VMW_PARAM_SM4_1. This is because, based on these calls, kernel
1189 * driver sends the supported cap.
1190 */
1191 ret = drmCommandWrite(vws->ioctl.drm_fd, DRM_VMW_GET_3D_CAP,
1192 &cap_arg, sizeof(cap_arg));
1193
1194 if (ret) {
1195 debug_printf("Failed to get 3D capabilities"
1196 " (%i, %s).\n", ret, strerror(-ret));
1197 goto out_no_caps;
1198 }
1199
1200 ret = vmw_ioctl_parse_caps(vws, cap_buffer);
1201 if (ret) {
1202 debug_printf("Failed to parse 3D capabilities"
1203 " (%i, %s).\n", ret, strerror(-ret));
1204 goto out_no_caps;
1205 }
1206
1207 if (((version->version_major == 2 && version->version_minor >= 10)
1208 || version->version_major > 2) && vws->base.have_vgpu10) {
1209
1210 /* support for these commands didn't make it into vmwgfx kernel
1211 * modules before 2.10.
1212 */
1213 vws->base.have_generate_mipmap_cmd = TRUE;
1214 vws->base.have_set_predication_cmd = TRUE;
1215 }
1216
1217 if (version->version_major == 2 && version->version_minor >= 14) {
1218 vws->base.have_fence_fd = TRUE;
1219 }
1220
1221 free(cap_buffer);
1222 drmFreeVersion(version);
1223 vmw_printf("%s OK\n", __FUNCTION__);
1224 return TRUE;
1225 out_no_caps:
1226 free(vws->ioctl.cap_3d);
1227 out_no_caparray:
1228 free(cap_buffer);
1229 out_no_3d:
1230 drmFreeVersion(version);
1231 out_no_version:
1232 vws->ioctl.num_cap_3d = 0;
1233 debug_printf("%s Failed\n", __FUNCTION__);
1234 return FALSE;
1235 }
1236
1237
1238
1239 void
vmw_ioctl_cleanup(struct vmw_winsys_screen * vws)1240 vmw_ioctl_cleanup(struct vmw_winsys_screen *vws)
1241 {
1242 VMW_FUNC;
1243
1244 free(vws->ioctl.cap_3d);
1245 }
1246