1 /**************************************************************************
3 * Copyright 2009 VMware, Inc.
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the
8 * "Software"), to deal in the Software without restriction, including
9 * without limitation the rights to use, copy, modify, merge, publish,
10 * distribute, sub license, and/or sell copies of the Software, and to
11 * permit persons to whom the Software is furnished to do so, subject to
12 * the following conditions:
14 * The above copyright notice and this permission notice (including the
15 * next paragraph) shall be included in all copies or substantial portions
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
19 * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
20 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
21 * IN NO EVENT SHALL VMWARE AND/OR ITS SUPPLIERS BE LIABLE FOR
22 * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
23 * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
24 * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
26 **************************************************************************/
30 * Texture sampling -- common code.
32 * @author Jose Fonseca <jfonseca@vmware.com>
35 #include "pipe/p_defines.h"
36 #include "pipe/p_state.h"
37 #include "util/u_format.h"
38 #include "util/u_math.h"
39 #include "lp_bld_arit.h"
40 #include "lp_bld_const.h"
41 #include "lp_bld_debug.h"
42 #include "lp_bld_flow.h"
43 #include "lp_bld_sample.h"
44 #include "lp_bld_swizzle.h"
45 #include "lp_bld_type.h"
49 * Initialize lp_sampler_static_state object with the gallium sampler
51 * The former is considered to be static and the later dynamic.
54 lp_sampler_static_state(struct lp_sampler_static_state *state,
55 const struct pipe_sampler_view *view,
56 const struct pipe_sampler_state *sampler)
58 const struct pipe_resource *texture = view->texture;
60 memset(state, 0, sizeof *state);
69 * We don't copy sampler state over unless it is actually enabled, to avoid
70 * spurious recompiles, as the sampler static state is part of the shader
73 * Ideally the state tracker or cso_cache module would make all state
74 * canonical, but until that happens it's better to be safe than sorry here.
76 * XXX: Actually there's much more than can be done here, especially
77 * regarding 1D/2D/3D/CUBE textures, wrap modes, etc.
80 state->format = view->format;
81 state->swizzle_r = view->swizzle_r;
82 state->swizzle_g = view->swizzle_g;
83 state->swizzle_b = view->swizzle_b;
84 state->swizzle_a = view->swizzle_a;
86 state->target = texture->target;
87 state->pot_width = util_is_power_of_two(texture->width0);
88 state->pot_height = util_is_power_of_two(texture->height0);
89 state->pot_depth = util_is_power_of_two(texture->depth0);
91 state->wrap_s = sampler->wrap_s;
92 state->wrap_t = sampler->wrap_t;
93 state->wrap_r = sampler->wrap_r;
94 state->min_img_filter = sampler->min_img_filter;
95 state->mag_img_filter = sampler->mag_img_filter;
96 if (view->last_level) {
97 state->min_mip_filter = sampler->min_mip_filter;
99 state->min_mip_filter = PIPE_TEX_MIPFILTER_NONE;
102 /* If min_lod == max_lod we can greatly simplify mipmap selection.
103 * This is a case that occurs during automatic mipmap generation.
105 if (sampler->min_lod == sampler->max_lod) {
106 state->min_max_lod_equal = 1;
107 state->min_max_lod = sampler->min_lod;
110 state->compare_mode = sampler->compare_mode;
111 if (sampler->compare_mode != PIPE_TEX_COMPARE_NONE) {
112 state->compare_func = sampler->compare_func;
115 state->normalized_coords = sampler->normalized_coords;
118 * FIXME: Handle the remainder of pipe_sampler_view.
124 * Generate code to compute texture level of detail (lambda).
125 * \param ddx partial derivatives of (s, t, r, q) with respect to X
126 * \param ddy partial derivatives of (s, t, r, q) with respect to Y
127 * \param lod_bias optional float vector with the shader lod bias
128 * \param explicit_lod optional float vector with the explicit lod
129 * \param width scalar int texture width
130 * \param height scalar int texture height
131 * \param depth scalar int texture depth
133 * XXX: The resulting lod is scalar, so ignore all but the first element of
134 * derivatives, lod_bias, etc that are passed by the shader.
137 lp_build_lod_selector(struct lp_build_sample_context *bld,
139 const LLVMValueRef ddx[4],
140 const LLVMValueRef ddy[4],
141 LLVMValueRef lod_bias, /* optional */
142 LLVMValueRef explicit_lod, /* optional */
148 if (bld->static_state->min_max_lod_equal) {
149 /* User is forcing sampling from a particular mipmap level.
150 * This is hit during mipmap generation.
152 return LLVMConstReal(LLVMFloatType(), bld->static_state->min_max_lod);
155 struct lp_build_context *float_bld = &bld->float_bld;
156 LLVMValueRef sampler_lod_bias =
157 bld->dynamic_state->lod_bias(bld->dynamic_state, bld->builder, unit);
158 LLVMValueRef min_lod =
159 bld->dynamic_state->min_lod(bld->dynamic_state, bld->builder, unit);
160 LLVMValueRef max_lod =
161 bld->dynamic_state->max_lod(bld->dynamic_state, bld->builder, unit);
162 LLVMValueRef index0 = LLVMConstInt(LLVMInt32Type(), 0, 0);
166 lod = LLVMBuildExtractElement(bld->builder, explicit_lod,
170 const int dims = texture_dims(bld->static_state->target);
171 LLVMValueRef dsdx, dsdy;
172 LLVMValueRef dtdx = NULL, dtdy = NULL, drdx = NULL, drdy = NULL;
175 dsdx = LLVMBuildExtractElement(bld->builder, ddx[0], index0, "dsdx");
176 dsdx = lp_build_abs(float_bld, dsdx);
177 dsdy = LLVMBuildExtractElement(bld->builder, ddy[0], index0, "dsdy");
178 dsdy = lp_build_abs(float_bld, dsdy);
180 dtdx = LLVMBuildExtractElement(bld->builder, ddx[1], index0, "dtdx");
181 dtdx = lp_build_abs(float_bld, dtdx);
182 dtdy = LLVMBuildExtractElement(bld->builder, ddy[1], index0, "dtdy");
183 dtdy = lp_build_abs(float_bld, dtdy);
185 drdx = LLVMBuildExtractElement(bld->builder, ddx[2], index0, "drdx");
186 drdx = lp_build_abs(float_bld, drdx);
187 drdy = LLVMBuildExtractElement(bld->builder, ddy[2], index0, "drdy");
188 drdy = lp_build_abs(float_bld, drdy);
192 /* Compute rho = max of all partial derivatives scaled by texture size.
193 * XXX this could be vectorized somewhat
195 rho = LLVMBuildFMul(bld->builder,
196 lp_build_max(float_bld, dsdx, dsdy),
197 lp_build_int_to_float(float_bld, width), "");
200 max = LLVMBuildFMul(bld->builder,
201 lp_build_max(float_bld, dtdx, dtdy),
202 lp_build_int_to_float(float_bld, height), "");
203 rho = lp_build_max(float_bld, rho, max);
205 max = LLVMBuildFMul(bld->builder,
206 lp_build_max(float_bld, drdx, drdy),
207 lp_build_int_to_float(float_bld, depth), "");
208 rho = lp_build_max(float_bld, rho, max);
212 /* compute lod = log2(rho) */
213 lod = lp_build_log2(float_bld, rho);
215 /* add shader lod bias */
217 lod_bias = LLVMBuildExtractElement(bld->builder, lod_bias,
219 lod = LLVMBuildFAdd(bld->builder, lod, lod_bias, "shader_lod_bias");
223 /* add sampler lod bias */
224 lod = LLVMBuildFAdd(bld->builder, lod, sampler_lod_bias, "sampler_lod_bias");
227 lod = lp_build_clamp(float_bld, lod, min_lod, max_lod);
235 * For PIPE_TEX_MIPFILTER_NEAREST, convert float LOD to integer
236 * mipmap level index.
237 * Note: this is all scalar code.
238 * \param lod scalar float texture level of detail
239 * \param level_out returns integer
242 lp_build_nearest_mip_level(struct lp_build_sample_context *bld,
245 LLVMValueRef *level_out)
247 struct lp_build_context *float_bld = &bld->float_bld;
248 struct lp_build_context *int_bld = &bld->int_bld;
249 LLVMValueRef last_level, level;
251 LLVMValueRef zero = LLVMConstInt(LLVMInt32Type(), 0, 0);
253 last_level = bld->dynamic_state->last_level(bld->dynamic_state,
256 /* convert float lod to integer */
257 level = lp_build_iround(float_bld, lod);
259 /* clamp level to legal range of levels */
260 *level_out = lp_build_clamp(int_bld, level, zero, last_level);
265 * For PIPE_TEX_MIPFILTER_LINEAR, convert float LOD to integer to
266 * two (adjacent) mipmap level indexes. Later, we'll sample from those
267 * two mipmap levels and interpolate between them.
270 lp_build_linear_mip_levels(struct lp_build_sample_context *bld,
273 LLVMValueRef *level0_out,
274 LLVMValueRef *level1_out,
275 LLVMValueRef *weight_out)
277 struct lp_build_context *float_bld = &bld->float_bld;
278 struct lp_build_context *int_bld = &bld->int_bld;
279 LLVMValueRef last_level, level;
281 last_level = bld->dynamic_state->last_level(bld->dynamic_state,
284 /* convert float lod to integer */
285 level = lp_build_ifloor(float_bld, lod);
287 /* compute level 0 and clamp to legal range of levels */
288 *level0_out = lp_build_clamp(int_bld, level,
291 /* compute level 1 and clamp to legal range of levels */
292 level = lp_build_add(int_bld, level, int_bld->one);
293 *level1_out = lp_build_clamp(int_bld, level,
297 *weight_out = lp_build_fract(float_bld, lod);
302 lp_build_get_mipmap_level(struct lp_build_sample_context *bld,
303 LLVMValueRef data_array, LLVMValueRef level)
305 LLVMValueRef indexes[2], data_ptr;
306 indexes[0] = LLVMConstInt(LLVMInt32Type(), 0, 0);
308 data_ptr = LLVMBuildGEP(bld->builder, data_array, indexes, 2, "");
309 data_ptr = LLVMBuildLoad(bld->builder, data_ptr, "");
315 lp_build_get_const_mipmap_level(struct lp_build_sample_context *bld,
316 LLVMValueRef data_array, int level)
318 LLVMValueRef lvl = LLVMConstInt(LLVMInt32Type(), level, 0);
319 return lp_build_get_mipmap_level(bld, data_array, lvl);
324 * Codegen equivalent for u_minify().
325 * Return max(1, base_size >> level);
328 lp_build_minify(struct lp_build_sample_context *bld,
329 LLVMValueRef base_size,
332 LLVMValueRef size = LLVMBuildLShr(bld->builder, base_size, level, "minify");
333 size = lp_build_max(&bld->int_coord_bld, size, bld->int_coord_bld.one);
339 * Dereference stride_array[mipmap_level] array to get a stride.
340 * Return stride as a vector.
343 lp_build_get_level_stride_vec(struct lp_build_sample_context *bld,
344 LLVMValueRef stride_array, LLVMValueRef level)
346 LLVMValueRef indexes[2], stride;
347 indexes[0] = LLVMConstInt(LLVMInt32Type(), 0, 0);
349 stride = LLVMBuildGEP(bld->builder, stride_array, indexes, 2, "");
350 stride = LLVMBuildLoad(bld->builder, stride, "");
351 stride = lp_build_broadcast_scalar(&bld->int_coord_bld, stride);
357 * When sampling a mipmap, we need to compute the width, height, depth
358 * of the source levels from the level indexes. This helper function
362 lp_build_mipmap_level_sizes(struct lp_build_sample_context *bld,
364 LLVMValueRef width_vec,
365 LLVMValueRef height_vec,
366 LLVMValueRef depth_vec,
367 LLVMValueRef ilevel0,
368 LLVMValueRef ilevel1,
369 LLVMValueRef row_stride_array,
370 LLVMValueRef img_stride_array,
371 LLVMValueRef *width0_vec,
372 LLVMValueRef *width1_vec,
373 LLVMValueRef *height0_vec,
374 LLVMValueRef *height1_vec,
375 LLVMValueRef *depth0_vec,
376 LLVMValueRef *depth1_vec,
377 LLVMValueRef *row_stride0_vec,
378 LLVMValueRef *row_stride1_vec,
379 LLVMValueRef *img_stride0_vec,
380 LLVMValueRef *img_stride1_vec)
382 const unsigned mip_filter = bld->static_state->min_mip_filter;
383 LLVMValueRef ilevel0_vec, ilevel1_vec;
385 ilevel0_vec = lp_build_broadcast_scalar(&bld->int_coord_bld, ilevel0);
386 if (mip_filter == PIPE_TEX_MIPFILTER_LINEAR)
387 ilevel1_vec = lp_build_broadcast_scalar(&bld->int_coord_bld, ilevel1);
390 * Compute width, height, depth at mipmap level 'ilevel0'
392 *width0_vec = lp_build_minify(bld, width_vec, ilevel0_vec);
394 *height0_vec = lp_build_minify(bld, height_vec, ilevel0_vec);
395 *row_stride0_vec = lp_build_get_level_stride_vec(bld,
398 if (dims == 3 || bld->static_state->target == PIPE_TEXTURE_CUBE) {
399 *img_stride0_vec = lp_build_get_level_stride_vec(bld,
403 *depth0_vec = lp_build_minify(bld, depth_vec, ilevel0_vec);
407 if (mip_filter == PIPE_TEX_MIPFILTER_LINEAR) {
408 /* compute width, height, depth for second mipmap level at 'ilevel1' */
409 *width1_vec = lp_build_minify(bld, width_vec, ilevel1_vec);
411 *height1_vec = lp_build_minify(bld, height_vec, ilevel1_vec);
412 *row_stride1_vec = lp_build_get_level_stride_vec(bld,
415 if (dims == 3 || bld->static_state->target == PIPE_TEXTURE_CUBE) {
416 *img_stride1_vec = lp_build_get_level_stride_vec(bld,
420 *depth1_vec = lp_build_minify(bld, depth_vec, ilevel1_vec);
429 /** Helper used by lp_build_cube_lookup() */
431 lp_build_cube_ima(struct lp_build_context *coord_bld, LLVMValueRef coord)
433 /* ima = -0.5 / abs(coord); */
434 LLVMValueRef negHalf = lp_build_const_vec(coord_bld->type, -0.5);
435 LLVMValueRef absCoord = lp_build_abs(coord_bld, coord);
436 LLVMValueRef ima = lp_build_div(coord_bld, negHalf, absCoord);
442 * Helper used by lp_build_cube_lookup()
443 * \param sign scalar +1 or -1
444 * \param coord float vector
445 * \param ima float vector
448 lp_build_cube_coord(struct lp_build_context *coord_bld,
449 LLVMValueRef sign, int negate_coord,
450 LLVMValueRef coord, LLVMValueRef ima)
452 /* return negate(coord) * ima * sign + 0.5; */
453 LLVMValueRef half = lp_build_const_vec(coord_bld->type, 0.5);
456 assert(negate_coord == +1 || negate_coord == -1);
458 if (negate_coord == -1) {
459 coord = lp_build_negate(coord_bld, coord);
462 res = lp_build_mul(coord_bld, coord, ima);
464 sign = lp_build_broadcast_scalar(coord_bld, sign);
465 res = lp_build_mul(coord_bld, res, sign);
467 res = lp_build_add(coord_bld, res, half);
473 /** Helper used by lp_build_cube_lookup()
474 * Return (major_coord >= 0) ? pos_face : neg_face;
477 lp_build_cube_face(struct lp_build_sample_context *bld,
478 LLVMValueRef major_coord,
479 unsigned pos_face, unsigned neg_face)
481 LLVMValueRef cmp = LLVMBuildFCmp(bld->builder, LLVMRealUGE,
483 bld->float_bld.zero, "");
484 LLVMValueRef pos = LLVMConstInt(LLVMInt32Type(), pos_face, 0);
485 LLVMValueRef neg = LLVMConstInt(LLVMInt32Type(), neg_face, 0);
486 LLVMValueRef res = LLVMBuildSelect(bld->builder, cmp, pos, neg, "");
493 * Generate code to do cube face selection and compute per-face texcoords.
496 lp_build_cube_lookup(struct lp_build_sample_context *bld,
501 LLVMValueRef *face_s,
502 LLVMValueRef *face_t)
504 struct lp_build_context *float_bld = &bld->float_bld;
505 struct lp_build_context *coord_bld = &bld->coord_bld;
506 LLVMValueRef rx, ry, rz;
507 LLVMValueRef arx, ary, arz;
508 LLVMValueRef c25 = LLVMConstReal(LLVMFloatType(), 0.25);
509 LLVMValueRef arx_ge_ary, arx_ge_arz;
510 LLVMValueRef ary_ge_arx, ary_ge_arz;
511 LLVMValueRef arx_ge_ary_arz, ary_ge_arx_arz;
512 LLVMValueRef rx_pos, ry_pos, rz_pos;
514 assert(bld->coord_bld.type.length == 4);
517 * Use the average of the four pixel's texcoords to choose the face.
519 rx = lp_build_mul(float_bld, c25,
520 lp_build_sum_vector(&bld->coord_bld, s));
521 ry = lp_build_mul(float_bld, c25,
522 lp_build_sum_vector(&bld->coord_bld, t));
523 rz = lp_build_mul(float_bld, c25,
524 lp_build_sum_vector(&bld->coord_bld, r));
526 arx = lp_build_abs(float_bld, rx);
527 ary = lp_build_abs(float_bld, ry);
528 arz = lp_build_abs(float_bld, rz);
531 * Compare sign/magnitude of rx,ry,rz to determine face
533 arx_ge_ary = LLVMBuildFCmp(bld->builder, LLVMRealUGE, arx, ary, "");
534 arx_ge_arz = LLVMBuildFCmp(bld->builder, LLVMRealUGE, arx, arz, "");
535 ary_ge_arx = LLVMBuildFCmp(bld->builder, LLVMRealUGE, ary, arx, "");
536 ary_ge_arz = LLVMBuildFCmp(bld->builder, LLVMRealUGE, ary, arz, "");
538 arx_ge_ary_arz = LLVMBuildAnd(bld->builder, arx_ge_ary, arx_ge_arz, "");
539 ary_ge_arx_arz = LLVMBuildAnd(bld->builder, ary_ge_arx, ary_ge_arz, "");
541 rx_pos = LLVMBuildFCmp(bld->builder, LLVMRealUGE, rx, float_bld->zero, "");
542 ry_pos = LLVMBuildFCmp(bld->builder, LLVMRealUGE, ry, float_bld->zero, "");
543 rz_pos = LLVMBuildFCmp(bld->builder, LLVMRealUGE, rz, float_bld->zero, "");
546 struct lp_build_flow_context *flow_ctx;
547 struct lp_build_if_state if_ctx;
549 flow_ctx = lp_build_flow_create(bld->builder);
550 lp_build_flow_scope_begin(flow_ctx);
552 *face_s = bld->coord_bld.undef;
553 *face_t = bld->coord_bld.undef;
554 *face = bld->int_bld.undef;
556 lp_build_name(*face_s, "face_s");
557 lp_build_name(*face_t, "face_t");
558 lp_build_name(*face, "face");
560 lp_build_flow_scope_declare(flow_ctx, face_s);
561 lp_build_flow_scope_declare(flow_ctx, face_t);
562 lp_build_flow_scope_declare(flow_ctx, face);
564 lp_build_if(&if_ctx, flow_ctx, bld->builder, arx_ge_ary_arz);
567 LLVMValueRef sign = lp_build_sgn(float_bld, rx);
568 LLVMValueRef ima = lp_build_cube_ima(coord_bld, s);
569 *face_s = lp_build_cube_coord(coord_bld, sign, +1, r, ima);
570 *face_t = lp_build_cube_coord(coord_bld, NULL, +1, t, ima);
571 *face = lp_build_cube_face(bld, rx,
573 PIPE_TEX_FACE_NEG_X);
575 lp_build_else(&if_ctx);
577 struct lp_build_flow_context *flow_ctx2;
578 struct lp_build_if_state if_ctx2;
580 LLVMValueRef face_s2 = bld->coord_bld.undef;
581 LLVMValueRef face_t2 = bld->coord_bld.undef;
582 LLVMValueRef face2 = bld->int_bld.undef;
584 flow_ctx2 = lp_build_flow_create(bld->builder);
585 lp_build_flow_scope_begin(flow_ctx2);
586 lp_build_flow_scope_declare(flow_ctx2, &face_s2);
587 lp_build_flow_scope_declare(flow_ctx2, &face_t2);
588 lp_build_flow_scope_declare(flow_ctx2, &face2);
590 ary_ge_arx_arz = LLVMBuildAnd(bld->builder, ary_ge_arx, ary_ge_arz, "");
592 lp_build_if(&if_ctx2, flow_ctx2, bld->builder, ary_ge_arx_arz);
595 LLVMValueRef sign = lp_build_sgn(float_bld, ry);
596 LLVMValueRef ima = lp_build_cube_ima(coord_bld, t);
597 face_s2 = lp_build_cube_coord(coord_bld, NULL, -1, s, ima);
598 face_t2 = lp_build_cube_coord(coord_bld, sign, -1, r, ima);
599 face2 = lp_build_cube_face(bld, ry,
601 PIPE_TEX_FACE_NEG_Y);
603 lp_build_else(&if_ctx2);
606 LLVMValueRef sign = lp_build_sgn(float_bld, rz);
607 LLVMValueRef ima = lp_build_cube_ima(coord_bld, r);
608 face_s2 = lp_build_cube_coord(coord_bld, sign, -1, s, ima);
609 face_t2 = lp_build_cube_coord(coord_bld, NULL, +1, t, ima);
610 face2 = lp_build_cube_face(bld, rz,
612 PIPE_TEX_FACE_NEG_Z);
614 lp_build_endif(&if_ctx2);
615 lp_build_flow_scope_end(flow_ctx2);
616 lp_build_flow_destroy(flow_ctx2);
622 lp_build_endif(&if_ctx);
623 lp_build_flow_scope_end(flow_ctx);
624 lp_build_flow_destroy(flow_ctx);
630 * Compute the partial offset of a pixel block along an arbitrary axis.
632 * @param coord coordinate in pixels
633 * @param stride number of bytes between rows of successive pixel blocks
634 * @param block_length number of pixels in a pixels block along the coordinate
636 * @param out_offset resulting relative offset of the pixel block in bytes
637 * @param out_subcoord resulting sub-block pixel coordinate
640 lp_build_sample_partial_offset(struct lp_build_context *bld,
641 unsigned block_length,
644 LLVMValueRef *out_offset,
645 LLVMValueRef *out_subcoord)
648 LLVMValueRef subcoord;
650 if (block_length == 1) {
651 subcoord = bld->zero;
655 * Pixel blocks have power of two dimensions. LLVM should convert the
656 * rem/div to bit arithmetic.
658 * It does indeed BUT it does transform it to scalar (and back) when doing so
659 * (using roughly extract, shift/and, mov, unpack) (llvm 2.7).
660 * The generated code looks seriously unfunny and is quite expensive.
663 LLVMValueRef block_width = lp_build_const_int_vec(bld->type, block_length);
664 subcoord = LLVMBuildURem(bld->builder, coord, block_width, "");
665 coord = LLVMBuildUDiv(bld->builder, coord, block_width, "");
667 unsigned logbase2 = util_unsigned_logbase2(block_length);
668 LLVMValueRef block_shift = lp_build_const_int_vec(bld->type, logbase2);
669 LLVMValueRef block_mask = lp_build_const_int_vec(bld->type, block_length - 1);
670 subcoord = LLVMBuildAnd(bld->builder, coord, block_mask, "");
671 coord = LLVMBuildLShr(bld->builder, coord, block_shift, "");
675 offset = lp_build_mul(bld, coord, stride);
678 assert(out_subcoord);
680 *out_offset = offset;
681 *out_subcoord = subcoord;
686 * Compute the offset of a pixel block.
688 * x, y, z, y_stride, z_stride are vectors, and they refer to pixels.
690 * Returns the relative offset and i,j sub-block coordinates
693 lp_build_sample_offset(struct lp_build_context *bld,
694 const struct util_format_description *format_desc,
698 LLVMValueRef y_stride,
699 LLVMValueRef z_stride,
700 LLVMValueRef *out_offset,
704 LLVMValueRef x_stride;
707 x_stride = lp_build_const_vec(bld->type, format_desc->block.bits/8);
709 lp_build_sample_partial_offset(bld,
710 format_desc->block.width,
715 LLVMValueRef y_offset;
716 lp_build_sample_partial_offset(bld,
717 format_desc->block.height,
720 offset = lp_build_add(bld, offset, y_offset);
727 LLVMValueRef z_offset;
729 lp_build_sample_partial_offset(bld,
730 1, /* pixel blocks are always 2D */
733 offset = lp_build_add(bld, offset, z_offset);
736 *out_offset = offset;