• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * Copyright © 2012 Intel Corporation
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a
5  * copy of this software and associated documentation files (the "Software"),
6  * to deal in the Software without restriction, including without limitation
7  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8  * and/or sell copies of the Software, and to permit persons to whom the
9  * Software is furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice (including the next
12  * paragraph) shall be included in all copies or substantial portions of the
13  * Software.
14  *
15  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
18  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
20  * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
21  * IN THE SOFTWARE.
22  */
23 
24 #include "main/glthread_marshal.h"
25 #include "main/dispatch.h"
26 #include "main/bufferobj.h"
27 
28 /**
29  * Create an upload buffer. This is called from the app thread, so everything
30  * has to be thread-safe in the driver.
31  */
32 static struct gl_buffer_object *
new_upload_buffer(struct gl_context * ctx,GLsizeiptr size,uint8_t ** ptr)33 new_upload_buffer(struct gl_context *ctx, GLsizeiptr size, uint8_t **ptr)
34 {
35    assert(ctx->GLThread.SupportsBufferUploads);
36 
37    struct gl_buffer_object *obj =
38       _mesa_bufferobj_alloc(ctx, -1);
39    if (!obj)
40       return NULL;
41 
42    obj->Immutable = true;
43 
44    if (!_mesa_bufferobj_data(ctx, GL_ARRAY_BUFFER, size, NULL,
45                           GL_WRITE_ONLY,
46                           GL_CLIENT_STORAGE_BIT | GL_MAP_WRITE_BIT,
47                           obj)) {
48       _mesa_delete_buffer_object(ctx, obj);
49       return NULL;
50    }
51 
52    *ptr = _mesa_bufferobj_map_range(ctx, 0, size,
53                                  GL_MAP_WRITE_BIT |
54                                  GL_MAP_UNSYNCHRONIZED_BIT |
55                                  MESA_MAP_THREAD_SAFE_BIT,
56                                  obj, MAP_GLTHREAD);
57    if (!*ptr) {
58       _mesa_delete_buffer_object(ctx, obj);
59       return NULL;
60    }
61 
62    return obj;
63 }
64 
65 void
_mesa_glthread_upload(struct gl_context * ctx,const void * data,GLsizeiptr size,unsigned * out_offset,struct gl_buffer_object ** out_buffer,uint8_t ** out_ptr)66 _mesa_glthread_upload(struct gl_context *ctx, const void *data,
67                       GLsizeiptr size, unsigned *out_offset,
68                       struct gl_buffer_object **out_buffer,
69                       uint8_t **out_ptr)
70 {
71    struct glthread_state *glthread = &ctx->GLThread;
72    const unsigned default_size = 1024 * 1024;
73 
74    if (unlikely(size > INT_MAX))
75       return;
76 
77    /* The alignment was chosen arbitrarily. */
78    unsigned offset = align(glthread->upload_offset, 8);
79 
80    /* Allocate a new buffer if needed. */
81    if (unlikely(!glthread->upload_buffer || offset + size > default_size)) {
82       /* If the size is greater than the buffer size, allocate a separate buffer
83        * just for this upload.
84        */
85       if (unlikely(size > default_size)) {
86          uint8_t *ptr;
87 
88          assert(*out_buffer == NULL);
89          *out_buffer = new_upload_buffer(ctx, size, &ptr);
90          if (!*out_buffer)
91             return;
92 
93          *out_offset = 0;
94          if (data)
95             memcpy(ptr, data, size);
96          else
97             *out_ptr = ptr;
98          return;
99       }
100 
101       if (glthread->upload_buffer_private_refcount > 0) {
102          p_atomic_add(&glthread->upload_buffer->RefCount,
103                       -glthread->upload_buffer_private_refcount);
104          glthread->upload_buffer_private_refcount = 0;
105       }
106       _mesa_reference_buffer_object(ctx, &glthread->upload_buffer, NULL);
107       glthread->upload_buffer =
108          new_upload_buffer(ctx, default_size, &glthread->upload_ptr);
109       glthread->upload_offset = 0;
110       offset = 0;
111 
112       /* Since atomic operations are very very slow when 2 threads are not
113        * sharing one L3 cache (which can happen on AMD Zen), prevent using
114        * atomics as follows:
115        *
116        * This function has to return a buffer reference to the caller.
117        * Instead of atomic_inc for every call, it does all possible future
118        * increments in advance when the upload buffer is allocated.
119        * The maximum number of times the function can be called per upload
120        * buffer is default_size, because the minimum allocation size is 1.
121        * Therefore the function can only return default_size number of
122        * references at most, so we will never need more. This is the number
123        * that is added to RefCount at allocation.
124        *
125        * upload_buffer_private_refcount tracks how many buffer references
126        * are left to return to callers. If the buffer is full and there are
127        * still references left, they are atomically subtracted from RefCount
128        * before the buffer is unreferenced.
129        *
130        * This can increase performance by 20%.
131        */
132       glthread->upload_buffer->RefCount += default_size;
133       glthread->upload_buffer_private_refcount = default_size;
134    }
135 
136    /* Upload data. */
137    if (data)
138       memcpy(glthread->upload_ptr + offset, data, size);
139    else
140       *out_ptr = glthread->upload_ptr + offset;
141 
142    glthread->upload_offset = offset + size;
143    *out_offset = offset;
144 
145    assert(*out_buffer == NULL);
146    assert(glthread->upload_buffer_private_refcount > 0);
147    *out_buffer = glthread->upload_buffer;
148    glthread->upload_buffer_private_refcount--;
149 }
150 
151 /** Tracks the current bindings for the vertex array and index array buffers.
152  *
153  * This is part of what we need to enable glthread on compat-GL contexts that
154  * happen to use VBOs, without also supporting the full tracking of VBO vs
155  * user vertex array bindings per attribute on each vertex array for
156  * determining what to upload at draw call time.
157  *
158  * Note that GL core makes it so that a buffer binding with an invalid handle
159  * in the "buffer" parameter will throw an error, and then a
160  * glVertexAttribPointer() that followsmight not end up pointing at a VBO.
161  * However, in GL core the draw call would throw an error as well, so we don't
162  * really care if our tracking is wrong for this case -- we never need to
163  * marshal user data for draw calls, and the unmarshal will just generate an
164  * error or not as appropriate.
165  *
166  * For compatibility GL, we do need to accurately know whether the draw call
167  * on the unmarshal side will dereference a user pointer or load data from a
168  * VBO per vertex.  That would make it seem like we need to track whether a
169  * "buffer" is valid, so that we can know when an error will be generated
170  * instead of updating the binding.  However, compat GL has the ridiculous
171  * feature that if you pass a bad name, it just gens a buffer object for you,
172  * so we escape without having to know if things are valid or not.
173  */
174 void
_mesa_glthread_BindBuffer(struct gl_context * ctx,GLenum target,GLuint buffer)175 _mesa_glthread_BindBuffer(struct gl_context *ctx, GLenum target, GLuint buffer)
176 {
177    struct glthread_state *glthread = &ctx->GLThread;
178 
179    switch (target) {
180    case GL_ARRAY_BUFFER:
181       glthread->CurrentArrayBufferName = buffer;
182       break;
183    case GL_ELEMENT_ARRAY_BUFFER:
184       /* The current element array buffer binding is actually tracked in the
185        * vertex array object instead of the context, so this would need to
186        * change on vertex array object updates.
187        */
188       glthread->CurrentVAO->CurrentElementBufferName = buffer;
189       break;
190    case GL_DRAW_INDIRECT_BUFFER:
191       glthread->CurrentDrawIndirectBufferName = buffer;
192       break;
193    case GL_PIXEL_PACK_BUFFER:
194       glthread->CurrentPixelPackBufferName = buffer;
195       break;
196    case GL_PIXEL_UNPACK_BUFFER:
197       glthread->CurrentPixelUnpackBufferName = buffer;
198       break;
199    case GL_QUERY_BUFFER:
200       glthread->CurrentQueryBufferName = buffer;
201       break;
202    }
203 }
204 
205 void
_mesa_glthread_DeleteBuffers(struct gl_context * ctx,GLsizei n,const GLuint * buffers)206 _mesa_glthread_DeleteBuffers(struct gl_context *ctx, GLsizei n,
207                              const GLuint *buffers)
208 {
209    struct glthread_state *glthread = &ctx->GLThread;
210 
211    if (!buffers || n < 0)
212       return;
213 
214    for (unsigned i = 0; i < n; i++) {
215       GLuint id = buffers[i];
216 
217       if (id == glthread->CurrentArrayBufferName)
218          _mesa_glthread_BindBuffer(ctx, GL_ARRAY_BUFFER, 0);
219       if (id == glthread->CurrentVAO->CurrentElementBufferName)
220          _mesa_glthread_BindBuffer(ctx, GL_ELEMENT_ARRAY_BUFFER, 0);
221       if (id == glthread->CurrentDrawIndirectBufferName)
222          _mesa_glthread_BindBuffer(ctx, GL_DRAW_INDIRECT_BUFFER, 0);
223       if (id == glthread->CurrentPixelPackBufferName)
224          _mesa_glthread_BindBuffer(ctx, GL_PIXEL_PACK_BUFFER, 0);
225       if (id == glthread->CurrentPixelUnpackBufferName)
226          _mesa_glthread_BindBuffer(ctx, GL_PIXEL_UNPACK_BUFFER, 0);
227    }
228 }
229 
230 /* BufferData: marshalled asynchronously */
231 struct marshal_cmd_BufferData
232 {
233    struct marshal_cmd_base cmd_base;
234    GLuint target_or_name;
235    GLsizeiptr size;
236    GLenum usage;
237    const GLvoid *data_external_mem;
238    bool data_null; /* If set, no data follows for "data" */
239    bool named;
240    bool ext_dsa;
241    /* Next size bytes are GLubyte data[size] */
242 };
243 
244 uint32_t
_mesa_unmarshal_BufferData(struct gl_context * ctx,const struct marshal_cmd_BufferData * cmd,const uint64_t * last)245 _mesa_unmarshal_BufferData(struct gl_context *ctx,
246                            const struct marshal_cmd_BufferData *cmd,
247                            const uint64_t *last)
248 {
249    const GLuint target_or_name = cmd->target_or_name;
250    const GLsizei size = cmd->size;
251    const GLenum usage = cmd->usage;
252    const void *data;
253 
254    if (cmd->data_null)
255       data = NULL;
256    else if (!cmd->named && target_or_name == GL_EXTERNAL_VIRTUAL_MEMORY_BUFFER_AMD)
257       data = cmd->data_external_mem;
258    else
259       data = (const void *) (cmd + 1);
260 
261    if (cmd->ext_dsa) {
262       CALL_NamedBufferDataEXT(ctx->CurrentServerDispatch,
263                               (target_or_name, size, data, usage));
264    } else if (cmd->named) {
265       CALL_NamedBufferData(ctx->CurrentServerDispatch,
266                            (target_or_name, size, data, usage));
267    } else {
268       CALL_BufferData(ctx->CurrentServerDispatch,
269                       (target_or_name, size, data, usage));
270    }
271    return cmd->cmd_base.cmd_size;
272 }
273 
274 uint32_t
_mesa_unmarshal_NamedBufferData(struct gl_context * ctx,const struct marshal_cmd_NamedBufferData * cmd,const uint64_t * last)275 _mesa_unmarshal_NamedBufferData(struct gl_context *ctx,
276                                 const struct marshal_cmd_NamedBufferData *cmd,
277                                 const uint64_t *last)
278 {
279    unreachable("never used - all BufferData variants use DISPATCH_CMD_BufferData");
280    return 0;
281 }
282 
283 uint32_t
_mesa_unmarshal_NamedBufferDataEXT(struct gl_context * ctx,const struct marshal_cmd_NamedBufferDataEXT * cmd,const uint64_t * last)284 _mesa_unmarshal_NamedBufferDataEXT(struct gl_context *ctx,
285                                    const struct marshal_cmd_NamedBufferDataEXT *cmd,
286                                    const uint64_t *last)
287 {
288    unreachable("never used - all BufferData variants use DISPATCH_CMD_BufferData");
289    return 0;
290 }
291 
292 static void
_mesa_marshal_BufferData_merged(GLuint target_or_name,GLsizeiptr size,const GLvoid * data,GLenum usage,bool named,bool ext_dsa,const char * func)293 _mesa_marshal_BufferData_merged(GLuint target_or_name, GLsizeiptr size,
294                                 const GLvoid *data, GLenum usage, bool named,
295                                 bool ext_dsa, const char *func)
296 {
297    GET_CURRENT_CONTEXT(ctx);
298    bool external_mem = !named &&
299                        target_or_name == GL_EXTERNAL_VIRTUAL_MEMORY_BUFFER_AMD;
300    bool copy_data = data && !external_mem;
301    size_t cmd_size = sizeof(struct marshal_cmd_BufferData) + (copy_data ? size : 0);
302 
303    if (unlikely(size < 0 || size > INT_MAX || cmd_size > MARSHAL_MAX_CMD_SIZE ||
304                 (named && target_or_name == 0))) {
305       _mesa_glthread_finish_before(ctx, func);
306       if (named) {
307          CALL_NamedBufferData(ctx->CurrentServerDispatch,
308                               (target_or_name, size, data, usage));
309       } else {
310          CALL_BufferData(ctx->CurrentServerDispatch,
311                          (target_or_name, size, data, usage));
312       }
313       return;
314    }
315 
316    struct marshal_cmd_BufferData *cmd =
317       _mesa_glthread_allocate_command(ctx, DISPATCH_CMD_BufferData,
318                                       cmd_size);
319 
320    cmd->target_or_name = target_or_name;
321    cmd->size = size;
322    cmd->usage = usage;
323    cmd->data_null = !data;
324    cmd->named = named;
325    cmd->ext_dsa = ext_dsa;
326    cmd->data_external_mem = data;
327 
328    if (copy_data) {
329       char *variable_data = (char *) (cmd + 1);
330       memcpy(variable_data, data, size);
331    }
332 }
333 
334 void GLAPIENTRY
_mesa_marshal_BufferData(GLenum target,GLsizeiptr size,const GLvoid * data,GLenum usage)335 _mesa_marshal_BufferData(GLenum target, GLsizeiptr size, const GLvoid * data,
336                          GLenum usage)
337 {
338    _mesa_marshal_BufferData_merged(target, size, data, usage, false, false,
339                                    "BufferData");
340 }
341 
342 void GLAPIENTRY
_mesa_marshal_NamedBufferData(GLuint buffer,GLsizeiptr size,const GLvoid * data,GLenum usage)343 _mesa_marshal_NamedBufferData(GLuint buffer, GLsizeiptr size,
344                               const GLvoid * data, GLenum usage)
345 {
346    _mesa_marshal_BufferData_merged(buffer, size, data, usage, true, false,
347                                    "NamedBufferData");
348 }
349 
350 void GLAPIENTRY
_mesa_marshal_NamedBufferDataEXT(GLuint buffer,GLsizeiptr size,const GLvoid * data,GLenum usage)351 _mesa_marshal_NamedBufferDataEXT(GLuint buffer, GLsizeiptr size,
352                                  const GLvoid *data, GLenum usage)
353 {
354    _mesa_marshal_BufferData_merged(buffer, size, data, usage, true, true,
355                                    "NamedBufferDataEXT");
356 }
357 
358 
359 /* BufferSubData: marshalled asynchronously */
360 struct marshal_cmd_BufferSubData
361 {
362    struct marshal_cmd_base cmd_base;
363    GLenum target_or_name;
364    GLintptr offset;
365    GLsizeiptr size;
366    bool named;
367    bool ext_dsa;
368    /* Next size bytes are GLubyte data[size] */
369 };
370 
371 uint32_t
_mesa_unmarshal_BufferSubData(struct gl_context * ctx,const struct marshal_cmd_BufferSubData * cmd,const uint64_t * last)372 _mesa_unmarshal_BufferSubData(struct gl_context *ctx,
373                               const struct marshal_cmd_BufferSubData *cmd,
374                               const uint64_t *last)
375 {
376    const GLenum target_or_name = cmd->target_or_name;
377    const GLintptr offset = cmd->offset;
378    const GLsizeiptr size = cmd->size;
379    const void *data = (const void *) (cmd + 1);
380 
381    if (cmd->ext_dsa) {
382       CALL_NamedBufferSubDataEXT(ctx->CurrentServerDispatch,
383                                  (target_or_name, offset, size, data));
384    } else if (cmd->named) {
385       CALL_NamedBufferSubData(ctx->CurrentServerDispatch,
386                               (target_or_name, offset, size, data));
387    } else {
388       CALL_BufferSubData(ctx->CurrentServerDispatch,
389                          (target_or_name, offset, size, data));
390    }
391    return cmd->cmd_base.cmd_size;
392 }
393 
394 uint32_t
_mesa_unmarshal_NamedBufferSubData(struct gl_context * ctx,const struct marshal_cmd_NamedBufferSubData * cmd,const uint64_t * last)395 _mesa_unmarshal_NamedBufferSubData(struct gl_context *ctx,
396                                    const struct marshal_cmd_NamedBufferSubData *cmd,
397                                    const uint64_t *last)
398 {
399    unreachable("never used - all BufferSubData variants use DISPATCH_CMD_BufferSubData");
400    return 0;
401 }
402 
403 uint32_t
_mesa_unmarshal_NamedBufferSubDataEXT(struct gl_context * ctx,const struct marshal_cmd_NamedBufferSubDataEXT * cmd,const uint64_t * last)404 _mesa_unmarshal_NamedBufferSubDataEXT(struct gl_context *ctx,
405                                       const struct marshal_cmd_NamedBufferSubDataEXT *cmd,
406                                       const uint64_t *last)
407 {
408    unreachable("never used - all BufferSubData variants use DISPATCH_CMD_BufferSubData");
409    return 0;
410 }
411 
412 static void
_mesa_marshal_BufferSubData_merged(GLuint target_or_name,GLintptr offset,GLsizeiptr size,const GLvoid * data,bool named,bool ext_dsa,const char * func)413 _mesa_marshal_BufferSubData_merged(GLuint target_or_name, GLintptr offset,
414                                    GLsizeiptr size, const GLvoid *data,
415                                    bool named, bool ext_dsa, const char *func)
416 {
417    GET_CURRENT_CONTEXT(ctx);
418    size_t cmd_size = sizeof(struct marshal_cmd_BufferSubData) + size;
419 
420    /* Fast path: Copy the data to an upload buffer, and use the GPU
421     * to copy the uploaded data to the destination buffer.
422     */
423    /* TODO: Handle offset == 0 && size < buffer_size.
424     *       If offset == 0 and size == buffer_size, it's better to discard
425     *       the buffer storage, but we don't know the buffer size in glthread.
426     */
427    if (ctx->GLThread.SupportsBufferUploads &&
428        data && offset > 0 && size > 0) {
429       struct gl_buffer_object *upload_buffer = NULL;
430       unsigned upload_offset = 0;
431 
432       _mesa_glthread_upload(ctx, data, size, &upload_offset, &upload_buffer,
433                             NULL);
434 
435       if (upload_buffer) {
436          _mesa_marshal_InternalBufferSubDataCopyMESA((GLintptr)upload_buffer,
437                                                      upload_offset,
438                                                      target_or_name,
439                                                      offset, size, named,
440                                                      ext_dsa);
441          return;
442       }
443    }
444 
445    if (unlikely(size < 0 || size > INT_MAX || cmd_size < 0 ||
446                 cmd_size > MARSHAL_MAX_CMD_SIZE || !data ||
447                 (named && target_or_name == 0))) {
448       _mesa_glthread_finish_before(ctx, func);
449       if (named) {
450          CALL_NamedBufferSubData(ctx->CurrentServerDispatch,
451                                  (target_or_name, offset, size, data));
452       } else {
453          CALL_BufferSubData(ctx->CurrentServerDispatch,
454                             (target_or_name, offset, size, data));
455       }
456       return;
457    }
458 
459    struct marshal_cmd_BufferSubData *cmd =
460       _mesa_glthread_allocate_command(ctx, DISPATCH_CMD_BufferSubData,
461                                       cmd_size);
462    cmd->target_or_name = target_or_name;
463    cmd->offset = offset;
464    cmd->size = size;
465    cmd->named = named;
466    cmd->ext_dsa = ext_dsa;
467 
468    char *variable_data = (char *) (cmd + 1);
469    memcpy(variable_data, data, size);
470 }
471 
472 void GLAPIENTRY
_mesa_marshal_BufferSubData(GLenum target,GLintptr offset,GLsizeiptr size,const GLvoid * data)473 _mesa_marshal_BufferSubData(GLenum target, GLintptr offset, GLsizeiptr size,
474                             const GLvoid * data)
475 {
476    _mesa_marshal_BufferSubData_merged(target, offset, size, data, false,
477                                       false, "BufferSubData");
478 }
479 
480 void GLAPIENTRY
_mesa_marshal_NamedBufferSubData(GLuint buffer,GLintptr offset,GLsizeiptr size,const GLvoid * data)481 _mesa_marshal_NamedBufferSubData(GLuint buffer, GLintptr offset,
482                                  GLsizeiptr size, const GLvoid * data)
483 {
484    _mesa_marshal_BufferSubData_merged(buffer, offset, size, data, true,
485                                       false, "NamedBufferSubData");
486 }
487 
488 void GLAPIENTRY
_mesa_marshal_NamedBufferSubDataEXT(GLuint buffer,GLintptr offset,GLsizeiptr size,const GLvoid * data)489 _mesa_marshal_NamedBufferSubDataEXT(GLuint buffer, GLintptr offset,
490                                     GLsizeiptr size, const GLvoid * data)
491 {
492    _mesa_marshal_BufferSubData_merged(buffer, offset, size, data, true,
493                                       true, "NamedBufferSubDataEXT");
494 }
495