/external/mesa3d/src/gallium/drivers/nvc0/ |
D | nvc0_context.c | 38 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_flush() local 39 struct nouveau_screen *screen = &nvc0->screen->base; in nvc0_flush() 44 PUSH_KICK(nvc0->base.pushbuf); /* fencing handled in kick_notify */ in nvc0_flush() 57 nvc0_context_unreference_resources(struct nvc0_context *nvc0) in nvc0_context_unreference_resources() argument 61 nouveau_bufctx_del(&nvc0->bufctx_3d); in nvc0_context_unreference_resources() 62 nouveau_bufctx_del(&nvc0->bufctx); in nvc0_context_unreference_resources() 64 util_unreference_framebuffer_state(&nvc0->framebuffer); in nvc0_context_unreference_resources() 66 for (i = 0; i < nvc0->num_vtxbufs; ++i) in nvc0_context_unreference_resources() 67 pipe_resource_reference(&nvc0->vtxbuf[i].buffer, NULL); in nvc0_context_unreference_resources() 69 pipe_resource_reference(&nvc0->idxbuf.buffer, NULL); in nvc0_context_unreference_resources() [all …]
|
D | nvc0_state_validate.c | 7 nvc0_validate_zcull(struct nvc0_context *nvc0) in nvc0_validate_zcull() argument 9 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_validate_zcull() 10 struct pipe_framebuffer_state *fb = &nvc0->framebuffer; in nvc0_validate_zcull() 56 nvc0_validate_fb(struct nvc0_context *nvc0) in nvc0_validate_fb() argument 58 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_validate_fb() 59 struct pipe_framebuffer_state *fb = &nvc0->framebuffer; in nvc0_validate_fb() 64 nouveau_bufctx_reset(nvc0->bufctx_3d, NVC0_BIND_FB); in nvc0_validate_fb() 120 BCTX_REFN(nvc0->bufctx_3d, FB, res, WR); in nvc0_validate_fb() 151 BCTX_REFN(nvc0->bufctx_3d, FB, &mt->base, WR); in nvc0_validate_fb() 164 nvc0_validate_blend_colour(struct nvc0_context *nvc0) in nvc0_validate_blend_colour() argument [all …]
|
D | nvc0_shader_state.c | 31 nvc0_program_update_context_state(struct nvc0_context *nvc0, in nvc0_program_update_context_state() argument 34 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_program_update_context_state() 38 if (!nvc0->state.tls_required) in nvc0_program_update_context_state() 39 BCTX_REFN_bo(nvc0->bufctx_3d, TLS, flags, nvc0->screen->tls); in nvc0_program_update_context_state() 40 nvc0->state.tls_required |= 1 << stage; in nvc0_program_update_context_state() 42 if (nvc0->state.tls_required == (1 << stage)) in nvc0_program_update_context_state() 43 nouveau_bufctx_reset(nvc0->bufctx_3d, NVC0_BIND_TLS); in nvc0_program_update_context_state() 44 nvc0->state.tls_required &= ~(1 << stage); in nvc0_program_update_context_state() 51 PUSH_DATAh(push, nvc0->screen->text->offset + prog->immd_base); in nvc0_program_update_context_state() 52 PUSH_DATA (push, nvc0->screen->text->offset + prog->immd_base); in nvc0_program_update_context_state() [all …]
|
D | nvc0_vbo.c | 166 nvc0_set_constant_vertex_attrib(struct nvc0_context *nvc0, const unsigned a) in nvc0_set_constant_vertex_attrib() argument 168 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_set_constant_vertex_attrib() 169 struct pipe_vertex_element *ve = &nvc0->vertex->element[a].pipe; in nvc0_set_constant_vertex_attrib() 170 struct pipe_vertex_buffer *vb = &nvc0->vtxbuf[ve->vertex_buffer_index]; in nvc0_set_constant_vertex_attrib() 199 nvc0_user_vbuf_range(struct nvc0_context *nvc0, int vbi, in nvc0_user_vbuf_range() argument 202 if (unlikely(nvc0->vertex->instance_bufs & (1 << vbi))) { in nvc0_user_vbuf_range() 203 const uint32_t div = nvc0->vertex->min_instance_div[vbi]; in nvc0_user_vbuf_range() 204 *base = nvc0->instance_off * nvc0->vtxbuf[vbi].stride; in nvc0_user_vbuf_range() 205 *size = (nvc0->instance_max / div) * nvc0->vtxbuf[vbi].stride + in nvc0_user_vbuf_range() 206 nvc0->vertex->vb_access_size[vbi]; in nvc0_user_vbuf_range() [all …]
|
D | nvc0_tex.c | 210 nvc0_validate_tic(struct nvc0_context *nvc0, int s) in nvc0_validate_tic() argument 213 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_validate_tic() 214 struct nouveau_bo *txc = nvc0->screen->txc; in nvc0_validate_tic() 219 for (i = 0; i < nvc0->num_textures[s]; ++i) { in nvc0_validate_tic() 220 struct nv50_tic_entry *tic = nv50_tic_entry(nvc0->textures[s][i]); in nvc0_validate_tic() 222 const boolean dirty = !!(nvc0->textures_dirty[s] & (1 << i)); in nvc0_validate_tic() 232 tic->id = nvc0_screen_tic_alloc(nvc0->screen, tic); in nvc0_validate_tic() 252 nvc0->screen->tic.lock[tic->id / 32] |= 1 << (tic->id % 32); in nvc0_validate_tic() 261 BCTX_REFN(nvc0->bufctx_3d, TEX(s, i), res, RD); in nvc0_validate_tic() 263 for (; i < nvc0->state.num_textures[s]; ++i) in nvc0_validate_tic() [all …]
|
D | nvc0_state.c | 194 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_blend_state_bind() local 196 nvc0->blend = hwcso; in nvc0_blend_state_bind() 197 nvc0->dirty |= NVC0_NEW_BLEND; in nvc0_blend_state_bind() 322 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_rasterizer_state_bind() local 324 nvc0->rast = hwcso; in nvc0_rasterizer_state_bind() 325 nvc0->dirty |= NVC0_NEW_RASTERIZER; in nvc0_rasterizer_state_bind() 393 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_zsa_state_bind() local 395 nvc0->zsa = hwcso; in nvc0_zsa_state_bind() 396 nvc0->dirty |= NVC0_NEW_ZSA; in nvc0_zsa_state_bind() 445 nvc0_stage_sampler_states_bind(struct nvc0_context *nvc0, int s, in nvc0_stage_sampler_states_bind() argument [all …]
|
D | nvc0_surface.c | 206 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_resource_copy_region() local 238 nvc0->m2mf_copy_rect(nvc0, &drect, &srect, nx, ny); in nvc0_resource_copy_region() 256 BCTX_REFN(nvc0->bufctx, 2D, nv04_resource(src), RD); in nvc0_resource_copy_region() 257 BCTX_REFN(nvc0->bufctx, 2D, nv04_resource(dst), WR); in nvc0_resource_copy_region() 258 nouveau_pushbuf_bufctx(nvc0->base.pushbuf, nvc0->bufctx); in nvc0_resource_copy_region() 259 nouveau_pushbuf_validate(nvc0->base.pushbuf); in nvc0_resource_copy_region() 262 ret = nvc0_2d_texture_do_copy(nvc0->base.pushbuf, in nvc0_resource_copy_region() 271 nouveau_bufctx_reset(nvc0->bufctx, 0); in nvc0_resource_copy_region() 281 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_clear_render_target() local 282 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_clear_render_target() [all …]
|
D | nvc0_vbo_translate.c | 40 nvc0_push_context_init(struct nvc0_context *nvc0, struct push_context *ctx) in nvc0_push_context_init() argument 42 ctx->push = nvc0->base.pushbuf; in nvc0_push_context_init() 44 ctx->translate = nvc0->vertex->translate; in nvc0_push_context_init() 45 ctx->vertex_size = nvc0->vertex->size; in nvc0_push_context_init() 48 nvc0->vertprog->vp.need_vertex_id && (nvc0->vertex->num_elements < 32); in nvc0_push_context_init() 51 ctx->edgeflag.enabled = nvc0->vertprog->vp.edgeflag < PIPE_MAX_ATTRIBS; in nvc0_push_context_init() 59 nvc0_vertex_configure_translate(struct nvc0_context *nvc0, int32_t index_bias) in nvc0_vertex_configure_translate() argument 61 struct translate *translate = nvc0->vertex->translate; in nvc0_vertex_configure_translate() 64 for (i = 0; i < nvc0->num_vtxbufs; ++i) { in nvc0_vertex_configure_translate() 66 const struct pipe_vertex_buffer *vb = &nvc0->vtxbuf[i]; in nvc0_vertex_configure_translate() [all …]
|
D | nvc0_push.c | 39 init_push_context(struct nvc0_context *nvc0, struct push_context *ctx) in init_push_context() argument 43 ctx->push = nvc0->base.pushbuf; in init_push_context() 44 ctx->translate = nvc0->vertex->translate; in init_push_context() 46 if (likely(nvc0->vertex->num_elements < 32)) in init_push_context() 47 ctx->need_vertex_id = nvc0->vertprog->vp.need_vertex_id; in init_push_context() 54 if (unlikely(nvc0->vertprog->vp.edgeflag < PIPE_MAX_ATTRIBS)) { in init_push_context() 55 ve = &nvc0->vertex->element[nvc0->vertprog->vp.edgeflag].pipe; in init_push_context() 60 ctx->packet_vertex_limit = nvc0->vertex->vtx_per_packet_max; in init_push_context() 65 ctx->vertex_words = nvc0->vertex->vtx_size; in init_push_context() 298 nvc0_push_vbo(struct nvc0_context *nvc0, const struct pipe_draw_info *info) in nvc0_push_vbo() argument [all …]
|
D | nvc0_transfer.c | 17 nvc0_m2mf_transfer_rect(struct nvc0_context *nvc0, in nvc0_m2mf_transfer_rect() argument 22 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_m2mf_transfer_rect() 23 struct nouveau_bufctx *bctx = nvc0->bufctx; in nvc0_m2mf_transfer_rect() 112 nve4_m2mf_transfer_rect(struct nvc0_context *nvc0, in nve4_m2mf_transfer_rect() argument 117 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nve4_m2mf_transfer_rect() 118 struct nouveau_bufctx *bctx = nvc0->bufctx; in nve4_m2mf_transfer_rect() 181 struct nvc0_context *nvc0 = nvc0_context(&nv->pipe); in nvc0_m2mf_push_linear() local 186 nouveau_bufctx_refn(nvc0->bufctx, 0, dst, domain | NOUVEAU_BO_WR); in nvc0_m2mf_push_linear() 187 nouveau_pushbuf_bufctx(push, nvc0->bufctx); in nvc0_m2mf_push_linear() 218 nouveau_bufctx_reset(nvc0->bufctx, 0); in nvc0_m2mf_push_linear() [all …]
|
D | nvc0_query.c | 59 nvc0_query_allocate(struct nvc0_context *nvc0, struct nvc0_query *q, int size) in nvc0_query_allocate() argument 61 struct nvc0_screen *screen = nvc0->screen; in nvc0_query_allocate() 82 nvc0_query_allocate(nvc0, q, 0); in nvc0_query_allocate() 100 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_query_create() local 138 if (!nvc0_query_allocate(nvc0, q, space)) { in nvc0_query_create() 172 nvc0_query_rotate(struct nvc0_context *nvc0, struct nvc0_query *q) in nvc0_query_rotate() argument 177 nvc0_query_allocate(nvc0, q, NVC0_QUERY_ALLOC_SPACE); in nvc0_query_rotate() 183 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_query_begin() local 184 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_query_begin() 192 nvc0_query_rotate(nvc0, q); in nvc0_query_begin() [all …]
|
D | nvc0_draw.c | 29 struct nvc0_context *nvc0; member 74 nvc0_draw_render_stage(struct nvc0_context *nvc0) in nvc0_draw_render_stage() argument 78 rs->nvc0 = nvc0; in nvc0_draw_render_stage() 79 rs->stage.draw = nvc0->draw; in nvc0_draw_render_stage()
|
/external/chromium_org/third_party/mesa/src/src/gallium/drivers/nvc0/ |
D | nvc0_context.c | 38 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_flush() local 39 struct nouveau_screen *screen = &nvc0->screen->base; in nvc0_flush() 44 PUSH_KICK(nvc0->base.pushbuf); /* fencing handled in kick_notify */ in nvc0_flush() 57 nvc0_context_unreference_resources(struct nvc0_context *nvc0) in nvc0_context_unreference_resources() argument 61 nouveau_bufctx_del(&nvc0->bufctx_3d); in nvc0_context_unreference_resources() 62 nouveau_bufctx_del(&nvc0->bufctx); in nvc0_context_unreference_resources() 64 util_unreference_framebuffer_state(&nvc0->framebuffer); in nvc0_context_unreference_resources() 66 for (i = 0; i < nvc0->num_vtxbufs; ++i) in nvc0_context_unreference_resources() 67 pipe_resource_reference(&nvc0->vtxbuf[i].buffer, NULL); in nvc0_context_unreference_resources() 69 pipe_resource_reference(&nvc0->idxbuf.buffer, NULL); in nvc0_context_unreference_resources() [all …]
|
D | nvc0_state_validate.c | 7 nvc0_validate_zcull(struct nvc0_context *nvc0) in nvc0_validate_zcull() argument 9 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_validate_zcull() 10 struct pipe_framebuffer_state *fb = &nvc0->framebuffer; in nvc0_validate_zcull() 56 nvc0_validate_fb(struct nvc0_context *nvc0) in nvc0_validate_fb() argument 58 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_validate_fb() 59 struct pipe_framebuffer_state *fb = &nvc0->framebuffer; in nvc0_validate_fb() 64 nouveau_bufctx_reset(nvc0->bufctx_3d, NVC0_BIND_FB); in nvc0_validate_fb() 120 BCTX_REFN(nvc0->bufctx_3d, FB, res, WR); in nvc0_validate_fb() 151 BCTX_REFN(nvc0->bufctx_3d, FB, &mt->base, WR); in nvc0_validate_fb() 164 nvc0_validate_blend_colour(struct nvc0_context *nvc0) in nvc0_validate_blend_colour() argument [all …]
|
D | nvc0_shader_state.c | 31 nvc0_program_update_context_state(struct nvc0_context *nvc0, in nvc0_program_update_context_state() argument 34 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_program_update_context_state() 38 if (!nvc0->state.tls_required) in nvc0_program_update_context_state() 39 BCTX_REFN_bo(nvc0->bufctx_3d, TLS, flags, nvc0->screen->tls); in nvc0_program_update_context_state() 40 nvc0->state.tls_required |= 1 << stage; in nvc0_program_update_context_state() 42 if (nvc0->state.tls_required == (1 << stage)) in nvc0_program_update_context_state() 43 nouveau_bufctx_reset(nvc0->bufctx_3d, NVC0_BIND_TLS); in nvc0_program_update_context_state() 44 nvc0->state.tls_required &= ~(1 << stage); in nvc0_program_update_context_state() 51 PUSH_DATAh(push, nvc0->screen->text->offset + prog->immd_base); in nvc0_program_update_context_state() 52 PUSH_DATA (push, nvc0->screen->text->offset + prog->immd_base); in nvc0_program_update_context_state() [all …]
|
D | nvc0_vbo.c | 166 nvc0_set_constant_vertex_attrib(struct nvc0_context *nvc0, const unsigned a) in nvc0_set_constant_vertex_attrib() argument 168 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_set_constant_vertex_attrib() 169 struct pipe_vertex_element *ve = &nvc0->vertex->element[a].pipe; in nvc0_set_constant_vertex_attrib() 170 struct pipe_vertex_buffer *vb = &nvc0->vtxbuf[ve->vertex_buffer_index]; in nvc0_set_constant_vertex_attrib() 199 nvc0_user_vbuf_range(struct nvc0_context *nvc0, int vbi, in nvc0_user_vbuf_range() argument 202 if (unlikely(nvc0->vertex->instance_bufs & (1 << vbi))) { in nvc0_user_vbuf_range() 203 const uint32_t div = nvc0->vertex->min_instance_div[vbi]; in nvc0_user_vbuf_range() 204 *base = nvc0->instance_off * nvc0->vtxbuf[vbi].stride; in nvc0_user_vbuf_range() 205 *size = (nvc0->instance_max / div) * nvc0->vtxbuf[vbi].stride + in nvc0_user_vbuf_range() 206 nvc0->vertex->vb_access_size[vbi]; in nvc0_user_vbuf_range() [all …]
|
D | nvc0_tex.c | 210 nvc0_validate_tic(struct nvc0_context *nvc0, int s) in nvc0_validate_tic() argument 213 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_validate_tic() 214 struct nouveau_bo *txc = nvc0->screen->txc; in nvc0_validate_tic() 219 for (i = 0; i < nvc0->num_textures[s]; ++i) { in nvc0_validate_tic() 220 struct nv50_tic_entry *tic = nv50_tic_entry(nvc0->textures[s][i]); in nvc0_validate_tic() 222 const boolean dirty = !!(nvc0->textures_dirty[s] & (1 << i)); in nvc0_validate_tic() 232 tic->id = nvc0_screen_tic_alloc(nvc0->screen, tic); in nvc0_validate_tic() 252 nvc0->screen->tic.lock[tic->id / 32] |= 1 << (tic->id % 32); in nvc0_validate_tic() 261 BCTX_REFN(nvc0->bufctx_3d, TEX(s, i), res, RD); in nvc0_validate_tic() 263 for (; i < nvc0->state.num_textures[s]; ++i) in nvc0_validate_tic() [all …]
|
D | nvc0_state.c | 194 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_blend_state_bind() local 196 nvc0->blend = hwcso; in nvc0_blend_state_bind() 197 nvc0->dirty |= NVC0_NEW_BLEND; in nvc0_blend_state_bind() 322 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_rasterizer_state_bind() local 324 nvc0->rast = hwcso; in nvc0_rasterizer_state_bind() 325 nvc0->dirty |= NVC0_NEW_RASTERIZER; in nvc0_rasterizer_state_bind() 393 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_zsa_state_bind() local 395 nvc0->zsa = hwcso; in nvc0_zsa_state_bind() 396 nvc0->dirty |= NVC0_NEW_ZSA; in nvc0_zsa_state_bind() 445 nvc0_stage_sampler_states_bind(struct nvc0_context *nvc0, int s, in nvc0_stage_sampler_states_bind() argument [all …]
|
D | nvc0_surface.c | 206 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_resource_copy_region() local 238 nvc0->m2mf_copy_rect(nvc0, &drect, &srect, nx, ny); in nvc0_resource_copy_region() 256 BCTX_REFN(nvc0->bufctx, 2D, nv04_resource(src), RD); in nvc0_resource_copy_region() 257 BCTX_REFN(nvc0->bufctx, 2D, nv04_resource(dst), WR); in nvc0_resource_copy_region() 258 nouveau_pushbuf_bufctx(nvc0->base.pushbuf, nvc0->bufctx); in nvc0_resource_copy_region() 259 nouveau_pushbuf_validate(nvc0->base.pushbuf); in nvc0_resource_copy_region() 262 ret = nvc0_2d_texture_do_copy(nvc0->base.pushbuf, in nvc0_resource_copy_region() 271 nouveau_bufctx_reset(nvc0->bufctx, 0); in nvc0_resource_copy_region() 281 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_clear_render_target() local 282 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_clear_render_target() [all …]
|
D | nvc0_vbo_translate.c | 40 nvc0_push_context_init(struct nvc0_context *nvc0, struct push_context *ctx) in nvc0_push_context_init() argument 42 ctx->push = nvc0->base.pushbuf; in nvc0_push_context_init() 44 ctx->translate = nvc0->vertex->translate; in nvc0_push_context_init() 45 ctx->vertex_size = nvc0->vertex->size; in nvc0_push_context_init() 48 nvc0->vertprog->vp.need_vertex_id && (nvc0->vertex->num_elements < 32); in nvc0_push_context_init() 51 ctx->edgeflag.enabled = nvc0->vertprog->vp.edgeflag < PIPE_MAX_ATTRIBS; in nvc0_push_context_init() 59 nvc0_vertex_configure_translate(struct nvc0_context *nvc0, int32_t index_bias) in nvc0_vertex_configure_translate() argument 61 struct translate *translate = nvc0->vertex->translate; in nvc0_vertex_configure_translate() 64 for (i = 0; i < nvc0->num_vtxbufs; ++i) { in nvc0_vertex_configure_translate() 66 const struct pipe_vertex_buffer *vb = &nvc0->vtxbuf[i]; in nvc0_vertex_configure_translate() [all …]
|
D | nvc0_push.c | 39 init_push_context(struct nvc0_context *nvc0, struct push_context *ctx) in init_push_context() argument 43 ctx->push = nvc0->base.pushbuf; in init_push_context() 44 ctx->translate = nvc0->vertex->translate; in init_push_context() 46 if (likely(nvc0->vertex->num_elements < 32)) in init_push_context() 47 ctx->need_vertex_id = nvc0->vertprog->vp.need_vertex_id; in init_push_context() 54 if (unlikely(nvc0->vertprog->vp.edgeflag < PIPE_MAX_ATTRIBS)) { in init_push_context() 55 ve = &nvc0->vertex->element[nvc0->vertprog->vp.edgeflag].pipe; in init_push_context() 60 ctx->packet_vertex_limit = nvc0->vertex->vtx_per_packet_max; in init_push_context() 65 ctx->vertex_words = nvc0->vertex->vtx_size; in init_push_context() 298 nvc0_push_vbo(struct nvc0_context *nvc0, const struct pipe_draw_info *info) in nvc0_push_vbo() argument [all …]
|
D | nvc0_transfer.c | 17 nvc0_m2mf_transfer_rect(struct nvc0_context *nvc0, in nvc0_m2mf_transfer_rect() argument 22 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_m2mf_transfer_rect() 23 struct nouveau_bufctx *bctx = nvc0->bufctx; in nvc0_m2mf_transfer_rect() 112 nve4_m2mf_transfer_rect(struct nvc0_context *nvc0, in nve4_m2mf_transfer_rect() argument 117 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nve4_m2mf_transfer_rect() 118 struct nouveau_bufctx *bctx = nvc0->bufctx; in nve4_m2mf_transfer_rect() 181 struct nvc0_context *nvc0 = nvc0_context(&nv->pipe); in nvc0_m2mf_push_linear() local 186 nouveau_bufctx_refn(nvc0->bufctx, 0, dst, domain | NOUVEAU_BO_WR); in nvc0_m2mf_push_linear() 187 nouveau_pushbuf_bufctx(push, nvc0->bufctx); in nvc0_m2mf_push_linear() 218 nouveau_bufctx_reset(nvc0->bufctx, 0); in nvc0_m2mf_push_linear() [all …]
|
D | nvc0_query.c | 59 nvc0_query_allocate(struct nvc0_context *nvc0, struct nvc0_query *q, int size) in nvc0_query_allocate() argument 61 struct nvc0_screen *screen = nvc0->screen; in nvc0_query_allocate() 82 nvc0_query_allocate(nvc0, q, 0); in nvc0_query_allocate() 100 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_query_create() local 138 if (!nvc0_query_allocate(nvc0, q, space)) { in nvc0_query_create() 172 nvc0_query_rotate(struct nvc0_context *nvc0, struct nvc0_query *q) in nvc0_query_rotate() argument 177 nvc0_query_allocate(nvc0, q, NVC0_QUERY_ALLOC_SPACE); in nvc0_query_rotate() 183 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_query_begin() local 184 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_query_begin() 192 nvc0_query_rotate(nvc0, q); in nvc0_query_begin() [all …]
|
D | nvc0_draw.c | 29 struct nvc0_context *nvc0; member 74 nvc0_draw_render_stage(struct nvc0_context *nvc0) in nvc0_draw_render_stage() argument 78 rs->nvc0 = nvc0; in nvc0_draw_render_stage() 79 rs->stage.draw = nvc0->draw; in nvc0_draw_render_stage()
|
D | SConscript | 5 nvc0 = env.ConvenienceLibrary( 6 target = 'nvc0', 11 Export('nvc0')
|