Home
last modified time | relevance | path

Searched refs:nvc0 (Results 1 – 25 of 57) sorted by relevance

123

/external/mesa3d/src/gallium/drivers/nvc0/
Dnvc0_context.c38 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_flush() local
39 struct nouveau_screen *screen = &nvc0->screen->base; in nvc0_flush()
44 PUSH_KICK(nvc0->base.pushbuf); /* fencing handled in kick_notify */ in nvc0_flush()
57 nvc0_context_unreference_resources(struct nvc0_context *nvc0) in nvc0_context_unreference_resources() argument
61 nouveau_bufctx_del(&nvc0->bufctx_3d); in nvc0_context_unreference_resources()
62 nouveau_bufctx_del(&nvc0->bufctx); in nvc0_context_unreference_resources()
64 util_unreference_framebuffer_state(&nvc0->framebuffer); in nvc0_context_unreference_resources()
66 for (i = 0; i < nvc0->num_vtxbufs; ++i) in nvc0_context_unreference_resources()
67 pipe_resource_reference(&nvc0->vtxbuf[i].buffer, NULL); in nvc0_context_unreference_resources()
69 pipe_resource_reference(&nvc0->idxbuf.buffer, NULL); in nvc0_context_unreference_resources()
[all …]
Dnvc0_state_validate.c7 nvc0_validate_zcull(struct nvc0_context *nvc0) in nvc0_validate_zcull() argument
9 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_validate_zcull()
10 struct pipe_framebuffer_state *fb = &nvc0->framebuffer; in nvc0_validate_zcull()
56 nvc0_validate_fb(struct nvc0_context *nvc0) in nvc0_validate_fb() argument
58 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_validate_fb()
59 struct pipe_framebuffer_state *fb = &nvc0->framebuffer; in nvc0_validate_fb()
64 nouveau_bufctx_reset(nvc0->bufctx_3d, NVC0_BIND_FB); in nvc0_validate_fb()
120 BCTX_REFN(nvc0->bufctx_3d, FB, res, WR); in nvc0_validate_fb()
151 BCTX_REFN(nvc0->bufctx_3d, FB, &mt->base, WR); in nvc0_validate_fb()
164 nvc0_validate_blend_colour(struct nvc0_context *nvc0) in nvc0_validate_blend_colour() argument
[all …]
Dnvc0_shader_state.c31 nvc0_program_update_context_state(struct nvc0_context *nvc0, in nvc0_program_update_context_state() argument
34 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_program_update_context_state()
38 if (!nvc0->state.tls_required) in nvc0_program_update_context_state()
39 BCTX_REFN_bo(nvc0->bufctx_3d, TLS, flags, nvc0->screen->tls); in nvc0_program_update_context_state()
40 nvc0->state.tls_required |= 1 << stage; in nvc0_program_update_context_state()
42 if (nvc0->state.tls_required == (1 << stage)) in nvc0_program_update_context_state()
43 nouveau_bufctx_reset(nvc0->bufctx_3d, NVC0_BIND_TLS); in nvc0_program_update_context_state()
44 nvc0->state.tls_required &= ~(1 << stage); in nvc0_program_update_context_state()
51 PUSH_DATAh(push, nvc0->screen->text->offset + prog->immd_base); in nvc0_program_update_context_state()
52 PUSH_DATA (push, nvc0->screen->text->offset + prog->immd_base); in nvc0_program_update_context_state()
[all …]
Dnvc0_vbo.c166 nvc0_set_constant_vertex_attrib(struct nvc0_context *nvc0, const unsigned a) in nvc0_set_constant_vertex_attrib() argument
168 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_set_constant_vertex_attrib()
169 struct pipe_vertex_element *ve = &nvc0->vertex->element[a].pipe; in nvc0_set_constant_vertex_attrib()
170 struct pipe_vertex_buffer *vb = &nvc0->vtxbuf[ve->vertex_buffer_index]; in nvc0_set_constant_vertex_attrib()
199 nvc0_user_vbuf_range(struct nvc0_context *nvc0, int vbi, in nvc0_user_vbuf_range() argument
202 if (unlikely(nvc0->vertex->instance_bufs & (1 << vbi))) { in nvc0_user_vbuf_range()
203 const uint32_t div = nvc0->vertex->min_instance_div[vbi]; in nvc0_user_vbuf_range()
204 *base = nvc0->instance_off * nvc0->vtxbuf[vbi].stride; in nvc0_user_vbuf_range()
205 *size = (nvc0->instance_max / div) * nvc0->vtxbuf[vbi].stride + in nvc0_user_vbuf_range()
206 nvc0->vertex->vb_access_size[vbi]; in nvc0_user_vbuf_range()
[all …]
Dnvc0_tex.c210 nvc0_validate_tic(struct nvc0_context *nvc0, int s) in nvc0_validate_tic() argument
213 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_validate_tic()
214 struct nouveau_bo *txc = nvc0->screen->txc; in nvc0_validate_tic()
219 for (i = 0; i < nvc0->num_textures[s]; ++i) { in nvc0_validate_tic()
220 struct nv50_tic_entry *tic = nv50_tic_entry(nvc0->textures[s][i]); in nvc0_validate_tic()
222 const boolean dirty = !!(nvc0->textures_dirty[s] & (1 << i)); in nvc0_validate_tic()
232 tic->id = nvc0_screen_tic_alloc(nvc0->screen, tic); in nvc0_validate_tic()
252 nvc0->screen->tic.lock[tic->id / 32] |= 1 << (tic->id % 32); in nvc0_validate_tic()
261 BCTX_REFN(nvc0->bufctx_3d, TEX(s, i), res, RD); in nvc0_validate_tic()
263 for (; i < nvc0->state.num_textures[s]; ++i) in nvc0_validate_tic()
[all …]
Dnvc0_state.c194 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_blend_state_bind() local
196 nvc0->blend = hwcso; in nvc0_blend_state_bind()
197 nvc0->dirty |= NVC0_NEW_BLEND; in nvc0_blend_state_bind()
322 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_rasterizer_state_bind() local
324 nvc0->rast = hwcso; in nvc0_rasterizer_state_bind()
325 nvc0->dirty |= NVC0_NEW_RASTERIZER; in nvc0_rasterizer_state_bind()
393 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_zsa_state_bind() local
395 nvc0->zsa = hwcso; in nvc0_zsa_state_bind()
396 nvc0->dirty |= NVC0_NEW_ZSA; in nvc0_zsa_state_bind()
445 nvc0_stage_sampler_states_bind(struct nvc0_context *nvc0, int s, in nvc0_stage_sampler_states_bind() argument
[all …]
Dnvc0_surface.c206 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_resource_copy_region() local
238 nvc0->m2mf_copy_rect(nvc0, &drect, &srect, nx, ny); in nvc0_resource_copy_region()
256 BCTX_REFN(nvc0->bufctx, 2D, nv04_resource(src), RD); in nvc0_resource_copy_region()
257 BCTX_REFN(nvc0->bufctx, 2D, nv04_resource(dst), WR); in nvc0_resource_copy_region()
258 nouveau_pushbuf_bufctx(nvc0->base.pushbuf, nvc0->bufctx); in nvc0_resource_copy_region()
259 nouveau_pushbuf_validate(nvc0->base.pushbuf); in nvc0_resource_copy_region()
262 ret = nvc0_2d_texture_do_copy(nvc0->base.pushbuf, in nvc0_resource_copy_region()
271 nouveau_bufctx_reset(nvc0->bufctx, 0); in nvc0_resource_copy_region()
281 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_clear_render_target() local
282 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_clear_render_target()
[all …]
Dnvc0_vbo_translate.c40 nvc0_push_context_init(struct nvc0_context *nvc0, struct push_context *ctx) in nvc0_push_context_init() argument
42 ctx->push = nvc0->base.pushbuf; in nvc0_push_context_init()
44 ctx->translate = nvc0->vertex->translate; in nvc0_push_context_init()
45 ctx->vertex_size = nvc0->vertex->size; in nvc0_push_context_init()
48 nvc0->vertprog->vp.need_vertex_id && (nvc0->vertex->num_elements < 32); in nvc0_push_context_init()
51 ctx->edgeflag.enabled = nvc0->vertprog->vp.edgeflag < PIPE_MAX_ATTRIBS; in nvc0_push_context_init()
59 nvc0_vertex_configure_translate(struct nvc0_context *nvc0, int32_t index_bias) in nvc0_vertex_configure_translate() argument
61 struct translate *translate = nvc0->vertex->translate; in nvc0_vertex_configure_translate()
64 for (i = 0; i < nvc0->num_vtxbufs; ++i) { in nvc0_vertex_configure_translate()
66 const struct pipe_vertex_buffer *vb = &nvc0->vtxbuf[i]; in nvc0_vertex_configure_translate()
[all …]
Dnvc0_push.c39 init_push_context(struct nvc0_context *nvc0, struct push_context *ctx) in init_push_context() argument
43 ctx->push = nvc0->base.pushbuf; in init_push_context()
44 ctx->translate = nvc0->vertex->translate; in init_push_context()
46 if (likely(nvc0->vertex->num_elements < 32)) in init_push_context()
47 ctx->need_vertex_id = nvc0->vertprog->vp.need_vertex_id; in init_push_context()
54 if (unlikely(nvc0->vertprog->vp.edgeflag < PIPE_MAX_ATTRIBS)) { in init_push_context()
55 ve = &nvc0->vertex->element[nvc0->vertprog->vp.edgeflag].pipe; in init_push_context()
60 ctx->packet_vertex_limit = nvc0->vertex->vtx_per_packet_max; in init_push_context()
65 ctx->vertex_words = nvc0->vertex->vtx_size; in init_push_context()
298 nvc0_push_vbo(struct nvc0_context *nvc0, const struct pipe_draw_info *info) in nvc0_push_vbo() argument
[all …]
Dnvc0_transfer.c17 nvc0_m2mf_transfer_rect(struct nvc0_context *nvc0, in nvc0_m2mf_transfer_rect() argument
22 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_m2mf_transfer_rect()
23 struct nouveau_bufctx *bctx = nvc0->bufctx; in nvc0_m2mf_transfer_rect()
112 nve4_m2mf_transfer_rect(struct nvc0_context *nvc0, in nve4_m2mf_transfer_rect() argument
117 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nve4_m2mf_transfer_rect()
118 struct nouveau_bufctx *bctx = nvc0->bufctx; in nve4_m2mf_transfer_rect()
181 struct nvc0_context *nvc0 = nvc0_context(&nv->pipe); in nvc0_m2mf_push_linear() local
186 nouveau_bufctx_refn(nvc0->bufctx, 0, dst, domain | NOUVEAU_BO_WR); in nvc0_m2mf_push_linear()
187 nouveau_pushbuf_bufctx(push, nvc0->bufctx); in nvc0_m2mf_push_linear()
218 nouveau_bufctx_reset(nvc0->bufctx, 0); in nvc0_m2mf_push_linear()
[all …]
Dnvc0_query.c59 nvc0_query_allocate(struct nvc0_context *nvc0, struct nvc0_query *q, int size) in nvc0_query_allocate() argument
61 struct nvc0_screen *screen = nvc0->screen; in nvc0_query_allocate()
82 nvc0_query_allocate(nvc0, q, 0); in nvc0_query_allocate()
100 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_query_create() local
138 if (!nvc0_query_allocate(nvc0, q, space)) { in nvc0_query_create()
172 nvc0_query_rotate(struct nvc0_context *nvc0, struct nvc0_query *q) in nvc0_query_rotate() argument
177 nvc0_query_allocate(nvc0, q, NVC0_QUERY_ALLOC_SPACE); in nvc0_query_rotate()
183 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_query_begin() local
184 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_query_begin()
192 nvc0_query_rotate(nvc0, q); in nvc0_query_begin()
[all …]
Dnvc0_draw.c29 struct nvc0_context *nvc0; member
74 nvc0_draw_render_stage(struct nvc0_context *nvc0) in nvc0_draw_render_stage() argument
78 rs->nvc0 = nvc0; in nvc0_draw_render_stage()
79 rs->stage.draw = nvc0->draw; in nvc0_draw_render_stage()
/external/chromium_org/third_party/mesa/src/src/gallium/drivers/nvc0/
Dnvc0_context.c38 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_flush() local
39 struct nouveau_screen *screen = &nvc0->screen->base; in nvc0_flush()
44 PUSH_KICK(nvc0->base.pushbuf); /* fencing handled in kick_notify */ in nvc0_flush()
57 nvc0_context_unreference_resources(struct nvc0_context *nvc0) in nvc0_context_unreference_resources() argument
61 nouveau_bufctx_del(&nvc0->bufctx_3d); in nvc0_context_unreference_resources()
62 nouveau_bufctx_del(&nvc0->bufctx); in nvc0_context_unreference_resources()
64 util_unreference_framebuffer_state(&nvc0->framebuffer); in nvc0_context_unreference_resources()
66 for (i = 0; i < nvc0->num_vtxbufs; ++i) in nvc0_context_unreference_resources()
67 pipe_resource_reference(&nvc0->vtxbuf[i].buffer, NULL); in nvc0_context_unreference_resources()
69 pipe_resource_reference(&nvc0->idxbuf.buffer, NULL); in nvc0_context_unreference_resources()
[all …]
Dnvc0_state_validate.c7 nvc0_validate_zcull(struct nvc0_context *nvc0) in nvc0_validate_zcull() argument
9 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_validate_zcull()
10 struct pipe_framebuffer_state *fb = &nvc0->framebuffer; in nvc0_validate_zcull()
56 nvc0_validate_fb(struct nvc0_context *nvc0) in nvc0_validate_fb() argument
58 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_validate_fb()
59 struct pipe_framebuffer_state *fb = &nvc0->framebuffer; in nvc0_validate_fb()
64 nouveau_bufctx_reset(nvc0->bufctx_3d, NVC0_BIND_FB); in nvc0_validate_fb()
120 BCTX_REFN(nvc0->bufctx_3d, FB, res, WR); in nvc0_validate_fb()
151 BCTX_REFN(nvc0->bufctx_3d, FB, &mt->base, WR); in nvc0_validate_fb()
164 nvc0_validate_blend_colour(struct nvc0_context *nvc0) in nvc0_validate_blend_colour() argument
[all …]
Dnvc0_shader_state.c31 nvc0_program_update_context_state(struct nvc0_context *nvc0, in nvc0_program_update_context_state() argument
34 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_program_update_context_state()
38 if (!nvc0->state.tls_required) in nvc0_program_update_context_state()
39 BCTX_REFN_bo(nvc0->bufctx_3d, TLS, flags, nvc0->screen->tls); in nvc0_program_update_context_state()
40 nvc0->state.tls_required |= 1 << stage; in nvc0_program_update_context_state()
42 if (nvc0->state.tls_required == (1 << stage)) in nvc0_program_update_context_state()
43 nouveau_bufctx_reset(nvc0->bufctx_3d, NVC0_BIND_TLS); in nvc0_program_update_context_state()
44 nvc0->state.tls_required &= ~(1 << stage); in nvc0_program_update_context_state()
51 PUSH_DATAh(push, nvc0->screen->text->offset + prog->immd_base); in nvc0_program_update_context_state()
52 PUSH_DATA (push, nvc0->screen->text->offset + prog->immd_base); in nvc0_program_update_context_state()
[all …]
Dnvc0_vbo.c166 nvc0_set_constant_vertex_attrib(struct nvc0_context *nvc0, const unsigned a) in nvc0_set_constant_vertex_attrib() argument
168 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_set_constant_vertex_attrib()
169 struct pipe_vertex_element *ve = &nvc0->vertex->element[a].pipe; in nvc0_set_constant_vertex_attrib()
170 struct pipe_vertex_buffer *vb = &nvc0->vtxbuf[ve->vertex_buffer_index]; in nvc0_set_constant_vertex_attrib()
199 nvc0_user_vbuf_range(struct nvc0_context *nvc0, int vbi, in nvc0_user_vbuf_range() argument
202 if (unlikely(nvc0->vertex->instance_bufs & (1 << vbi))) { in nvc0_user_vbuf_range()
203 const uint32_t div = nvc0->vertex->min_instance_div[vbi]; in nvc0_user_vbuf_range()
204 *base = nvc0->instance_off * nvc0->vtxbuf[vbi].stride; in nvc0_user_vbuf_range()
205 *size = (nvc0->instance_max / div) * nvc0->vtxbuf[vbi].stride + in nvc0_user_vbuf_range()
206 nvc0->vertex->vb_access_size[vbi]; in nvc0_user_vbuf_range()
[all …]
Dnvc0_tex.c210 nvc0_validate_tic(struct nvc0_context *nvc0, int s) in nvc0_validate_tic() argument
213 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_validate_tic()
214 struct nouveau_bo *txc = nvc0->screen->txc; in nvc0_validate_tic()
219 for (i = 0; i < nvc0->num_textures[s]; ++i) { in nvc0_validate_tic()
220 struct nv50_tic_entry *tic = nv50_tic_entry(nvc0->textures[s][i]); in nvc0_validate_tic()
222 const boolean dirty = !!(nvc0->textures_dirty[s] & (1 << i)); in nvc0_validate_tic()
232 tic->id = nvc0_screen_tic_alloc(nvc0->screen, tic); in nvc0_validate_tic()
252 nvc0->screen->tic.lock[tic->id / 32] |= 1 << (tic->id % 32); in nvc0_validate_tic()
261 BCTX_REFN(nvc0->bufctx_3d, TEX(s, i), res, RD); in nvc0_validate_tic()
263 for (; i < nvc0->state.num_textures[s]; ++i) in nvc0_validate_tic()
[all …]
Dnvc0_state.c194 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_blend_state_bind() local
196 nvc0->blend = hwcso; in nvc0_blend_state_bind()
197 nvc0->dirty |= NVC0_NEW_BLEND; in nvc0_blend_state_bind()
322 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_rasterizer_state_bind() local
324 nvc0->rast = hwcso; in nvc0_rasterizer_state_bind()
325 nvc0->dirty |= NVC0_NEW_RASTERIZER; in nvc0_rasterizer_state_bind()
393 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_zsa_state_bind() local
395 nvc0->zsa = hwcso; in nvc0_zsa_state_bind()
396 nvc0->dirty |= NVC0_NEW_ZSA; in nvc0_zsa_state_bind()
445 nvc0_stage_sampler_states_bind(struct nvc0_context *nvc0, int s, in nvc0_stage_sampler_states_bind() argument
[all …]
Dnvc0_surface.c206 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_resource_copy_region() local
238 nvc0->m2mf_copy_rect(nvc0, &drect, &srect, nx, ny); in nvc0_resource_copy_region()
256 BCTX_REFN(nvc0->bufctx, 2D, nv04_resource(src), RD); in nvc0_resource_copy_region()
257 BCTX_REFN(nvc0->bufctx, 2D, nv04_resource(dst), WR); in nvc0_resource_copy_region()
258 nouveau_pushbuf_bufctx(nvc0->base.pushbuf, nvc0->bufctx); in nvc0_resource_copy_region()
259 nouveau_pushbuf_validate(nvc0->base.pushbuf); in nvc0_resource_copy_region()
262 ret = nvc0_2d_texture_do_copy(nvc0->base.pushbuf, in nvc0_resource_copy_region()
271 nouveau_bufctx_reset(nvc0->bufctx, 0); in nvc0_resource_copy_region()
281 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_clear_render_target() local
282 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_clear_render_target()
[all …]
Dnvc0_vbo_translate.c40 nvc0_push_context_init(struct nvc0_context *nvc0, struct push_context *ctx) in nvc0_push_context_init() argument
42 ctx->push = nvc0->base.pushbuf; in nvc0_push_context_init()
44 ctx->translate = nvc0->vertex->translate; in nvc0_push_context_init()
45 ctx->vertex_size = nvc0->vertex->size; in nvc0_push_context_init()
48 nvc0->vertprog->vp.need_vertex_id && (nvc0->vertex->num_elements < 32); in nvc0_push_context_init()
51 ctx->edgeflag.enabled = nvc0->vertprog->vp.edgeflag < PIPE_MAX_ATTRIBS; in nvc0_push_context_init()
59 nvc0_vertex_configure_translate(struct nvc0_context *nvc0, int32_t index_bias) in nvc0_vertex_configure_translate() argument
61 struct translate *translate = nvc0->vertex->translate; in nvc0_vertex_configure_translate()
64 for (i = 0; i < nvc0->num_vtxbufs; ++i) { in nvc0_vertex_configure_translate()
66 const struct pipe_vertex_buffer *vb = &nvc0->vtxbuf[i]; in nvc0_vertex_configure_translate()
[all …]
Dnvc0_push.c39 init_push_context(struct nvc0_context *nvc0, struct push_context *ctx) in init_push_context() argument
43 ctx->push = nvc0->base.pushbuf; in init_push_context()
44 ctx->translate = nvc0->vertex->translate; in init_push_context()
46 if (likely(nvc0->vertex->num_elements < 32)) in init_push_context()
47 ctx->need_vertex_id = nvc0->vertprog->vp.need_vertex_id; in init_push_context()
54 if (unlikely(nvc0->vertprog->vp.edgeflag < PIPE_MAX_ATTRIBS)) { in init_push_context()
55 ve = &nvc0->vertex->element[nvc0->vertprog->vp.edgeflag].pipe; in init_push_context()
60 ctx->packet_vertex_limit = nvc0->vertex->vtx_per_packet_max; in init_push_context()
65 ctx->vertex_words = nvc0->vertex->vtx_size; in init_push_context()
298 nvc0_push_vbo(struct nvc0_context *nvc0, const struct pipe_draw_info *info) in nvc0_push_vbo() argument
[all …]
Dnvc0_transfer.c17 nvc0_m2mf_transfer_rect(struct nvc0_context *nvc0, in nvc0_m2mf_transfer_rect() argument
22 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_m2mf_transfer_rect()
23 struct nouveau_bufctx *bctx = nvc0->bufctx; in nvc0_m2mf_transfer_rect()
112 nve4_m2mf_transfer_rect(struct nvc0_context *nvc0, in nve4_m2mf_transfer_rect() argument
117 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nve4_m2mf_transfer_rect()
118 struct nouveau_bufctx *bctx = nvc0->bufctx; in nve4_m2mf_transfer_rect()
181 struct nvc0_context *nvc0 = nvc0_context(&nv->pipe); in nvc0_m2mf_push_linear() local
186 nouveau_bufctx_refn(nvc0->bufctx, 0, dst, domain | NOUVEAU_BO_WR); in nvc0_m2mf_push_linear()
187 nouveau_pushbuf_bufctx(push, nvc0->bufctx); in nvc0_m2mf_push_linear()
218 nouveau_bufctx_reset(nvc0->bufctx, 0); in nvc0_m2mf_push_linear()
[all …]
Dnvc0_query.c59 nvc0_query_allocate(struct nvc0_context *nvc0, struct nvc0_query *q, int size) in nvc0_query_allocate() argument
61 struct nvc0_screen *screen = nvc0->screen; in nvc0_query_allocate()
82 nvc0_query_allocate(nvc0, q, 0); in nvc0_query_allocate()
100 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_query_create() local
138 if (!nvc0_query_allocate(nvc0, q, space)) { in nvc0_query_create()
172 nvc0_query_rotate(struct nvc0_context *nvc0, struct nvc0_query *q) in nvc0_query_rotate() argument
177 nvc0_query_allocate(nvc0, q, NVC0_QUERY_ALLOC_SPACE); in nvc0_query_rotate()
183 struct nvc0_context *nvc0 = nvc0_context(pipe); in nvc0_query_begin() local
184 struct nouveau_pushbuf *push = nvc0->base.pushbuf; in nvc0_query_begin()
192 nvc0_query_rotate(nvc0, q); in nvc0_query_begin()
[all …]
Dnvc0_draw.c29 struct nvc0_context *nvc0; member
74 nvc0_draw_render_stage(struct nvc0_context *nvc0) in nvc0_draw_render_stage() argument
78 rs->nvc0 = nvc0; in nvc0_draw_render_stage()
79 rs->stage.draw = nvc0->draw; in nvc0_draw_render_stage()
DSConscript5 nvc0 = env.ConvenienceLibrary(
6 target = 'nvc0',
11 Export('nvc0')

123