1 /**************************************************************************
3 * Copyright 2007 Tungsten Graphics, Inc., Cedar Park, Texas.
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the
8 * "Software"), to deal in the Software without restriction, including
9 * without limitation the rights to use, copy, modify, merge, publish,
10 * distribute, sub license, and/or sell copies of the Software, and to
11 * permit persons to whom the Software is furnished to do so, subject to
12 * the following conditions:
14 * The above copyright notice and this permission notice (including the
15 * next paragraph) shall be included in all copies or substantial portions
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
19 * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
20 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
21 * IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR
22 * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
23 * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
24 * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
26 **************************************************************************/
30 * Build post-transformation, post-clipping vertex buffers and element
31 * lists by hooking into the end of the primitive pipeline and
32 * manipulating the vertex_id field in the vertex headers.
34 * XXX: work in progress
36 * \author José Fonseca <jrfonseca@tungstengraphics.com>
37 * \author Keith Whitwell <keith@tungstengraphics.com>
41 #include "pipe/p_debug.h"
42 #include "pipe/p_inlines.h"
43 #include "pipe/p_winsys.h"
45 #include "nv20_context.h"
46 #include "nv20_state.h"
48 #include "draw/draw_vbuf.h"
51 * Primitive renderer for nv20.
53 struct nv20_vbuf_render {
54 struct vbuf_render base;
56 struct nv20_context *nv20;
58 /** Vertex buffer in VRAM */
59 struct pipe_buffer *pbuffer;
61 /** Vertex buffer in normal memory */
64 /** Vertex size in bytes */
65 /*unsigned vertex_size;*/
67 /** Hardware primitive */
72 * Basically a cast wrapper.
74 static INLINE struct nv20_vbuf_render *
75 nv20_vbuf_render(struct vbuf_render *render)
78 return (struct nv20_vbuf_render *)render;
81 void nv20_vtxbuf_bind( struct nv20_context* nv20 )
85 for(i = 0; i < NV20TCL_VTXBUF_ADDRESS__SIZE; i++) {
86 BEGIN_RING(kelvin, NV20TCL_VTXBUF_ADDRESS(i), 1);
87 OUT_RING(0/*nv20->vtxbuf*/);
88 BEGIN_RING(kelvin, NV20TCL_VTXFMT(i) ,1);
94 static const struct vertex_info *
95 nv20_vbuf_render_get_vertex_info( struct vbuf_render *render )
97 struct nv20_vbuf_render *nv20_render = nv20_vbuf_render(render);
98 struct nv20_context *nv20 = nv20_render->nv20;
100 nv20_emit_hw_state(nv20);
102 return &nv20->vertex_info;
106 nv20__allocate_mbuffer(struct nv20_vbuf_render *nv20_render, size_t size)
108 nv20_render->mbuffer = MALLOC(size);
109 return nv20_render->mbuffer;
113 nv20__allocate_pbuffer(struct nv20_vbuf_render *nv20_render, size_t size)
115 struct pipe_winsys *winsys = nv20_render->nv20->pipe.winsys;
116 nv20_render->pbuffer = winsys->buffer_create(winsys, 64,
117 PIPE_BUFFER_USAGE_VERTEX, size);
118 return winsys->buffer_map(winsys,
119 nv20_render->pbuffer,
120 PIPE_BUFFER_USAGE_CPU_WRITE);
124 nv20_vbuf_render_allocate_vertices( struct vbuf_render *render,
128 struct nv20_vbuf_render *nv20_render = nv20_vbuf_render(render);
129 size_t size = (size_t)vertex_size * (size_t)nr_vertices;
132 assert(!nv20_render->pbuffer);
133 assert(!nv20_render->mbuffer);
136 * For small amount of vertices, don't bother with pipe vertex
137 * buffer, the data will be passed directly via the fifo.
139 /* XXX: Pipe vertex buffers don't work. */
140 if (0 && size > 16 * 1024)
141 buf = nv20__allocate_pbuffer(nv20_render, size);
143 buf = nv20__allocate_mbuffer(nv20_render, size);
146 nv20_render->nv20->dirty |= NV20_NEW_VTXARRAYS;
152 nv20_vbuf_render_set_primitive( struct vbuf_render *render,
155 struct nv20_vbuf_render *nv20_render = nv20_vbuf_render(render);
156 unsigned hwp = nvgl_primitive(prim);
160 nv20_render->hwprim = hwp;
165 nv20__vtxhwformat(unsigned stride, unsigned fields, unsigned type)
167 return (stride << NV20TCL_VTXFMT_STRIDE_SHIFT) |
168 (fields << NV20TCL_VTXFMT_SIZE_SHIFT) |
169 (type << NV20TCL_VTXFMT_TYPE_SHIFT);
173 nv20__emit_format(struct nv20_context *nv20, enum attrib_emit type, int hwattr)
180 hwfmt = nv20__vtxhwformat(0, 0, 2);
184 hwfmt = nv20__vtxhwformat(4, 1, 2);
188 hwfmt = nv20__vtxhwformat(8, 2, 2);
192 hwfmt = nv20__vtxhwformat(12, 3, 2);
196 hwfmt = nv20__vtxhwformat(16, 4, 2);
200 NOUVEAU_ERR("unhandled attrib_emit %d\n", type);
204 BEGIN_RING(kelvin, NV20TCL_VTXFMT(hwattr), 1);
210 nv20__emit_vertex_array_format(struct nv20_context *nv20)
212 struct vertex_info *vinfo = &nv20->vertex_info;
213 int hwattr = NV20TCL_VTXFMT__SIZE;
215 unsigned nr_fields = 0;
217 while (hwattr-- > 0) {
218 if (vinfo->hwfmt[0] & (1 << hwattr)) {
219 nr_fields += nv20__emit_format(nv20,
220 vinfo->attrib[attr].emit, hwattr);
223 nv20__emit_format(nv20, EMIT_OMIT, hwattr);
230 nv20__draw_mbuffer(struct nv20_vbuf_render *nv20_render,
231 const ushort *indices,
234 struct nv20_context *nv20 = nv20_render->nv20;
235 struct vertex_info *vinfo = &nv20->vertex_info;
238 ubyte *data = nv20_render->mbuffer;
239 int vsz = 4 * vinfo->size;
241 nr_fields = nv20__emit_vertex_array_format(nv20);
243 BEGIN_RING(kelvin, NV20TCL_VERTEX_BEGIN_END, 1);
244 OUT_RING(nv20_render->hwprim);
246 max_push = 1200 / nr_fields;
249 int push = MIN2(nr_indices, max_push);
251 BEGIN_RING_NI(kelvin, NV20TCL_VERTEX_DATA, push * nr_fields);
252 for (i = 0; i < push; i++) {
253 /* XXX: fixme to handle other than floats? */
255 float *attrv = (float*)&data[indices[i] * vsz];
264 BEGIN_RING(kelvin, NV20TCL_VERTEX_BEGIN_END, 1);
265 OUT_RING(NV20TCL_VERTEX_BEGIN_END_STOP);
269 nv20__draw_pbuffer(struct nv20_vbuf_render *nv20_render,
270 const ushort *indices,
273 struct nv20_context *nv20 = nv20_render->nv20;
276 NOUVEAU_ERR("nv20__draw_pbuffer: this path is broken.\n");
278 BEGIN_RING(kelvin, NV10TCL_VERTEX_ARRAY_OFFSET_POS, 1);
279 OUT_RELOCl(nv20_render->pbuffer, 0,
280 NOUVEAU_BO_VRAM | NOUVEAU_BO_GART | NOUVEAU_BO_RD);
282 BEGIN_RING(kelvin, NV10TCL_VERTEX_BUFFER_BEGIN_END, 1);
283 OUT_RING(nv20_render->hwprim);
285 if (nr_indices & 1) {
286 BEGIN_RING(kelvin, NV10TCL_VB_ELEMENT_U32, 1);
287 OUT_RING (indices[0]);
288 indices++; nr_indices--;
292 // XXX too big/small ? check the size
293 push = MIN2(nr_indices, 1200 * 2);
295 BEGIN_RING_NI(kelvin, NV10TCL_VB_ELEMENT_U16, push >> 1);
296 for (i = 0; i < push; i+=2)
297 OUT_RING((indices[i+1] << 16) | indices[i]);
303 BEGIN_RING(kelvin, NV10TCL_VERTEX_BUFFER_BEGIN_END, 1);
308 nv20_vbuf_render_draw( struct vbuf_render *render,
309 const ushort *indices,
312 struct nv20_vbuf_render *nv20_render = nv20_vbuf_render(render);
314 nv20_emit_hw_state(nv20_render->nv20);
316 if (nv20_render->pbuffer)
317 nv20__draw_pbuffer(nv20_render, indices, nr_indices);
318 else if (nv20_render->mbuffer)
319 nv20__draw_mbuffer(nv20_render, indices, nr_indices);
326 nv20_vbuf_render_release_vertices( struct vbuf_render *render,
328 unsigned vertex_size,
329 unsigned vertices_used )
331 struct nv20_vbuf_render *nv20_render = nv20_vbuf_render(render);
332 struct nv20_context *nv20 = nv20_render->nv20;
333 struct pipe_winsys *winsys = nv20->pipe.winsys;
334 struct pipe_screen *pscreen = &nv20->screen->pipe;
336 if (nv20_render->pbuffer) {
337 winsys->buffer_unmap(winsys, nv20_render->pbuffer);
338 pipe_buffer_reference(pscreen, &nv20_render->pbuffer, NULL);
339 } else if (nv20_render->mbuffer) {
340 FREE(nv20_render->mbuffer);
341 nv20_render->mbuffer = NULL;
348 nv20_vbuf_render_destroy( struct vbuf_render *render )
350 struct nv20_vbuf_render *nv20_render = nv20_vbuf_render(render);
352 assert(!nv20_render->pbuffer);
353 assert(!nv20_render->mbuffer);
360 * Create a new primitive render.
362 static struct vbuf_render *
363 nv20_vbuf_render_create( struct nv20_context *nv20 )
365 struct nv20_vbuf_render *nv20_render = CALLOC_STRUCT(nv20_vbuf_render);
367 nv20_render->nv20 = nv20;
369 nv20_render->base.max_vertex_buffer_bytes = 16*1024;
370 nv20_render->base.max_indices = 1024;
371 nv20_render->base.get_vertex_info = nv20_vbuf_render_get_vertex_info;
372 nv20_render->base.allocate_vertices =
373 nv20_vbuf_render_allocate_vertices;
374 nv20_render->base.set_primitive = nv20_vbuf_render_set_primitive;
375 nv20_render->base.draw = nv20_vbuf_render_draw;
376 nv20_render->base.release_vertices = nv20_vbuf_render_release_vertices;
377 nv20_render->base.destroy = nv20_vbuf_render_destroy;
379 return &nv20_render->base;
384 * Create a new primitive vbuf/render stage.
386 struct draw_stage *nv20_draw_vbuf_stage( struct nv20_context *nv20 )
388 struct vbuf_render *render;
389 struct draw_stage *stage;
391 render = nv20_vbuf_render_create(nv20);
395 stage = draw_vbuf_stage( nv20->draw, render );
397 render->destroy(render);