1 /**********************************************************
2 * Copyright 2008-2009 VMware, Inc. All rights reserved.
4 * Permission is hereby granted, free of charge, to any person
5 * obtaining a copy of this software and associated documentation
6 * files (the "Software"), to deal in the Software without
7 * restriction, including without limitation the rights to use, copy,
8 * modify, merge, publish, distribute, sublicense, and/or sell copies
9 * of the Software, and to permit persons to whom the Software is
10 * furnished to do so, subject to the following conditions:
12 * The above copyright notice and this permission notice shall be
13 * included in all copies or substantial portions of the Software.
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
16 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
17 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
18 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
19 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
20 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
21 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
24 **********************************************************/
26 #include "draw/draw_vbuf.h"
27 #include "draw/draw_context.h"
28 #include "draw/draw_vertex.h"
30 #include "util/u_debug.h"
31 #include "util/u_inlines.h"
32 #include "util/u_math.h"
33 #include "util/u_memory.h"
35 #include "svga_context.h"
36 #include "svga_state.h"
37 #include "svga_swtnl.h"
39 #include "svga_types.h"
41 #include "svga3d_reg.h"
42 #include "svga_draw.h"
43 #include "svga_swtnl_private.h"
46 static const struct vertex_info *
47 svga_vbuf_render_get_vertex_info( struct vbuf_render *render )
49 struct svga_vbuf_render *svga_render = svga_vbuf_render(render);
50 struct svga_context *svga = svga_render->svga;
52 svga_swtnl_update_vdecl(svga);
54 return &svga_render->vertex_info;
59 svga_vbuf_render_allocate_vertices( struct vbuf_render *render,
63 struct svga_vbuf_render *svga_render = svga_vbuf_render(render);
64 struct svga_context *svga = svga_render->svga;
65 struct pipe_screen *screen = svga->pipe.screen;
66 size_t size = (size_t)nr_vertices * (size_t)vertex_size;
67 boolean new_vbuf = FALSE;
68 boolean new_ibuf = FALSE;
70 if (svga_render->vertex_size != vertex_size)
71 svga->swtnl.new_vdecl = TRUE;
72 svga_render->vertex_size = (size_t)vertex_size;
74 if (svga->swtnl.new_vbuf)
75 new_ibuf = new_vbuf = TRUE;
76 svga->swtnl.new_vbuf = FALSE;
78 if (svga_render->vbuf_size < svga_render->vbuf_offset + svga_render->vbuf_used + size)
82 pipe_resource_reference(&svga_render->vbuf, NULL);
84 pipe_resource_reference(&svga_render->ibuf, NULL);
86 if (!svga_render->vbuf) {
87 svga_render->vbuf_size = MAX2(size, svga_render->vbuf_alloc_size);
88 svga_render->vbuf = pipe_buffer_create(screen,
89 PIPE_BIND_VERTEX_BUFFER,
91 svga_render->vbuf_size);
92 if(!svga_render->vbuf) {
93 svga_context_flush(svga, NULL);
94 assert(!svga_render->vbuf);
95 svga_render->vbuf = pipe_buffer_create(screen,
96 PIPE_BIND_VERTEX_BUFFER,
98 svga_render->vbuf_size);
99 assert(svga_render->vbuf);
102 svga->swtnl.new_vdecl = TRUE;
103 svga_render->vbuf_offset = 0;
105 svga_render->vbuf_offset += svga_render->vbuf_used;
108 svga_render->vbuf_used = 0;
110 if (svga->swtnl.new_vdecl)
111 svga_render->vdecl_offset = svga_render->vbuf_offset;
117 svga_vbuf_render_map_vertices( struct vbuf_render *render )
119 struct svga_vbuf_render *svga_render = svga_vbuf_render(render);
120 struct svga_context *svga = svga_render->svga;
122 char *ptr = (char*)pipe_buffer_map(&svga->pipe,
124 PIPE_TRANSFER_WRITE |
125 PIPE_TRANSFER_FLUSH_EXPLICIT |
126 PIPE_TRANSFER_DISCARD |
127 PIPE_TRANSFER_UNSYNCHRONIZED,
128 &svga_render->vbuf_transfer);
129 return ptr + svga_render->vbuf_offset;
133 svga_vbuf_render_unmap_vertices( struct vbuf_render *render,
137 struct svga_vbuf_render *svga_render = svga_vbuf_render(render);
138 struct svga_context *svga = svga_render->svga;
139 unsigned offset, length;
140 size_t used = svga_render->vertex_size * ((size_t)max_index + 1);
142 offset = svga_render->vbuf_offset + svga_render->vertex_size * min_index;
143 length = svga_render->vertex_size * (max_index + 1 - min_index);
144 pipe_buffer_flush_mapped_range(&svga->pipe,
145 svga_render->vbuf_transfer,
147 pipe_buffer_unmap(&svga->pipe, svga_render->vbuf_transfer);
148 svga_render->min_index = min_index;
149 svga_render->max_index = max_index;
150 svga_render->vbuf_used = MAX2(svga_render->vbuf_used, used);
154 svga_vbuf_render_set_primitive( struct vbuf_render *render,
157 struct svga_vbuf_render *svga_render = svga_vbuf_render(render);
158 svga_render->prim = prim;
164 svga_vbuf_submit_state( struct svga_vbuf_render *svga_render )
166 struct svga_context *svga = svga_render->svga;
167 SVGA3dVertexDecl vdecl[PIPE_MAX_ATTRIBS];
171 /* if the vdecl or vbuf hasn't changed do nothing */
172 if (!svga->swtnl.new_vdecl)
175 memcpy(vdecl, svga_render->vdecl, sizeof(vdecl));
177 /* flush the hw state */
178 ret = svga_hwtnl_flush(svga->hwtnl);
180 svga_context_flush(svga, NULL);
181 ret = svga_hwtnl_flush(svga->hwtnl);
182 /* if we hit this path we might become synced with hw */
183 svga->swtnl.new_vbuf = TRUE;
187 svga_hwtnl_reset_vdecl(svga->hwtnl, svga_render->vdecl_count);
189 for (i = 0; i < svga_render->vdecl_count; i++) {
190 vdecl[i].array.offset += svga_render->vdecl_offset;
192 svga_hwtnl_vdecl( svga->hwtnl,
198 /* We have already taken care of flatshading, so let the hwtnl
199 * module use whatever is most convenient:
201 if (svga->state.sw.need_pipeline) {
202 svga_hwtnl_set_flatshade(svga->hwtnl, FALSE, FALSE);
203 svga_hwtnl_set_unfilled(svga->hwtnl, PIPE_POLYGON_MODE_FILL);
206 svga_hwtnl_set_flatshade( svga->hwtnl,
207 svga->curr.rast->templ.flatshade,
208 svga->curr.rast->templ.flatshade_first );
210 svga_hwtnl_set_unfilled( svga->hwtnl,
211 svga->curr.rast->hw_unfilled );
214 svga->swtnl.new_vdecl = FALSE;
218 svga_vbuf_render_draw_arrays( struct vbuf_render *render,
222 struct svga_vbuf_render *svga_render = svga_vbuf_render(render);
223 struct svga_context *svga = svga_render->svga;
224 unsigned bias = (svga_render->vbuf_offset - svga_render->vdecl_offset) / svga_render->vertex_size;
225 enum pipe_error ret = 0;
227 /* off to hardware */
228 svga_vbuf_submit_state(svga_render);
230 /* Need to call update_state() again as the draw module may have
231 * altered some of our state behind our backs. Testcase:
234 svga_update_state_retry( svga, SVGA_STATE_HW_DRAW );
236 ret = svga_hwtnl_draw_arrays(svga->hwtnl, svga_render->prim, start + bias, nr);
237 if (ret != PIPE_OK) {
238 svga_context_flush(svga, NULL);
239 ret = svga_hwtnl_draw_arrays(svga->hwtnl, svga_render->prim, start + bias, nr);
240 svga->swtnl.new_vbuf = TRUE;
241 assert(ret == PIPE_OK);
247 svga_vbuf_render_draw_elements( struct vbuf_render *render,
248 const ushort *indices,
251 struct svga_vbuf_render *svga_render = svga_vbuf_render(render);
252 struct svga_context *svga = svga_render->svga;
253 struct pipe_screen *screen = svga->pipe.screen;
254 int bias = (svga_render->vbuf_offset - svga_render->vdecl_offset) / svga_render->vertex_size;
256 size_t size = 2 * nr_indices;
258 assert(( svga_render->vbuf_offset - svga_render->vdecl_offset) % svga_render->vertex_size == 0);
260 if (svga_render->ibuf_size < svga_render->ibuf_offset + size)
261 pipe_resource_reference(&svga_render->ibuf, NULL);
263 if (!svga_render->ibuf) {
264 svga_render->ibuf_size = MAX2(size, svga_render->ibuf_alloc_size);
265 svga_render->ibuf = pipe_buffer_create(screen,
266 PIPE_BIND_INDEX_BUFFER,
268 svga_render->ibuf_size);
269 svga_render->ibuf_offset = 0;
272 pipe_buffer_write_nooverlap(&svga->pipe, svga_render->ibuf,
273 svga_render->ibuf_offset, 2 * nr_indices, indices);
275 /* off to hardware */
276 svga_vbuf_submit_state(svga_render);
278 /* Need to call update_state() again as the draw module may have
279 * altered some of our state behind our backs. Testcase:
282 svga_update_state_retry( svga, SVGA_STATE_HW_DRAW );
284 ret = svga_hwtnl_draw_range_elements(svga->hwtnl,
288 svga_render->min_index,
289 svga_render->max_index,
291 svga_render->ibuf_offset / 2, nr_indices);
293 svga_context_flush(svga, NULL);
294 ret = svga_hwtnl_draw_range_elements(svga->hwtnl,
298 svga_render->min_index,
299 svga_render->max_index,
301 svga_render->ibuf_offset / 2, nr_indices);
302 svga->swtnl.new_vbuf = TRUE;
303 assert(ret == PIPE_OK);
306 svga_render->ibuf_offset += size;
311 svga_vbuf_render_release_vertices( struct vbuf_render *render )
318 svga_vbuf_render_destroy( struct vbuf_render *render )
320 struct svga_vbuf_render *svga_render = svga_vbuf_render(render);
322 pipe_resource_reference(&svga_render->vbuf, NULL);
323 pipe_resource_reference(&svga_render->ibuf, NULL);
329 * Create a new primitive render.
332 svga_vbuf_render_create( struct svga_context *svga )
334 struct svga_vbuf_render *svga_render = CALLOC_STRUCT(svga_vbuf_render);
336 svga_render->svga = svga;
337 svga_render->ibuf_size = 0;
338 svga_render->vbuf_size = 0;
339 svga_render->ibuf_alloc_size = 4*1024;
340 svga_render->vbuf_alloc_size = 64*1024;
341 svga_render->base.max_vertex_buffer_bytes = 64*1024/10;
342 svga_render->base.max_indices = 65536;
343 svga_render->base.get_vertex_info = svga_vbuf_render_get_vertex_info;
344 svga_render->base.allocate_vertices = svga_vbuf_render_allocate_vertices;
345 svga_render->base.map_vertices = svga_vbuf_render_map_vertices;
346 svga_render->base.unmap_vertices = svga_vbuf_render_unmap_vertices;
347 svga_render->base.set_primitive = svga_vbuf_render_set_primitive;
348 svga_render->base.draw_elements = svga_vbuf_render_draw_elements;
349 svga_render->base.draw_arrays = svga_vbuf_render_draw_arrays;
350 svga_render->base.release_vertices = svga_vbuf_render_release_vertices;
351 svga_render->base.destroy = svga_vbuf_render_destroy;
353 return &svga_render->base;