bool
mem_vectorize_callback(unsigned align_mul, unsigned align_offset,
unsigned bit_size,
- unsigned num_components, unsigned high_offset,
+ unsigned num_components,
nir_intrinsic_instr *low, nir_intrinsic_instr *high)
{
if (num_components > 4)
typedef bool (*nir_should_vectorize_mem_func)(unsigned align_mul,
unsigned align_offset,
unsigned bit_size,
- unsigned num_components, unsigned high_offset,
+ unsigned num_components,
nir_intrinsic_instr *low, nir_intrinsic_instr *high);
bool nir_opt_load_store_vectorize(nir_shader *shader, nir_variable_mode modes,
if (!ctx->callback(low->align_mul,
low->align_offset,
new_bit_size, new_num_components,
- high_offset, low->intrin, high->intrin))
+ low->intrin, high->intrin))
return false;
if (low->is_store) {
static bool mem_vectorize_callback(unsigned align_mul, unsigned align_offset,
unsigned bit_size,
- unsigned num_components, unsigned high_offset,
+ unsigned num_components,
nir_intrinsic_instr *low, nir_intrinsic_instr *high);
static void shared_type_info(const struct glsl_type *type, unsigned *size, unsigned *align);
bool nir_load_store_vectorize_test::mem_vectorize_callback(
unsigned align_mul, unsigned align_offset, unsigned bit_size,
- unsigned num_components, unsigned high_offset,
+ unsigned num_components,
nir_intrinsic_instr *low, nir_intrinsic_instr *high)
{
return bit_size / 8;
static bool
brw_nir_should_vectorize_mem(unsigned align_mul, unsigned align_offset,
unsigned bit_size,
- unsigned num_components, unsigned high_offset,
+ unsigned num_components,
nir_intrinsic_instr *low,
nir_intrinsic_instr *high)
{