mem_vectorize_callback(unsigned align_mul, unsigned align_offset,
unsigned bit_size,
unsigned num_components,
- nir_intrinsic_instr *low, nir_intrinsic_instr *high)
+ nir_intrinsic_instr *low, nir_intrinsic_instr *high,
+ void *data)
{
if (num_components > 4)
return false;
unsigned align_offset,
unsigned bit_size,
unsigned num_components,
- nir_intrinsic_instr *low, nir_intrinsic_instr *high);
+ nir_intrinsic_instr *low, nir_intrinsic_instr *high,
+ void *data);
typedef struct {
nir_should_vectorize_mem_func callback;
nir_variable_mode modes;
nir_variable_mode robust_modes;
+ void *cb_data;
} nir_load_store_vectorize_options;
bool nir_opt_load_store_vectorize(nir_shader *shader, const nir_load_store_vectorize_options *options);
if (!ctx->options->callback(low->align_mul,
low->align_offset,
new_bit_size, new_num_components,
- low->intrin, high->intrin))
+ low->intrin, high->intrin,
+ ctx->options->cb_data))
return false;
if (low->is_store) {
static bool mem_vectorize_callback(unsigned align_mul, unsigned align_offset,
unsigned bit_size,
unsigned num_components,
- nir_intrinsic_instr *low, nir_intrinsic_instr *high);
+ nir_intrinsic_instr *low, nir_intrinsic_instr *high,
+ void *data);
static void shared_type_info(const struct glsl_type *type, unsigned *size, unsigned *align);
std::string swizzle(nir_alu_instr *instr, int src);
bool nir_load_store_vectorize_test::mem_vectorize_callback(
unsigned align_mul, unsigned align_offset, unsigned bit_size,
unsigned num_components,
- nir_intrinsic_instr *low, nir_intrinsic_instr *high)
+ nir_intrinsic_instr *low, nir_intrinsic_instr *high,
+ void *data)
{
/* Calculate a simple alignment, like how nir_intrinsic_align() does. */
uint32_t align = align_mul;
unsigned bit_size,
unsigned num_components,
nir_intrinsic_instr *low,
- nir_intrinsic_instr *high)
+ nir_intrinsic_instr *high,
+ void *data)
{
assert(bit_size >= 8);
if (bit_size != 32)
static bool
ntt_should_vectorize_io(unsigned align, unsigned bit_size,
unsigned num_components, unsigned high_offset,
- nir_intrinsic_instr *low, nir_intrinsic_instr *high)
+ nir_intrinsic_instr *low, nir_intrinsic_instr *high,
+ void *data)
{
if (bit_size != 32)
return false;
unsigned bit_size,
unsigned num_components,
nir_intrinsic_instr *low,
- nir_intrinsic_instr *high)
+ nir_intrinsic_instr *high,
+ void *data)
{
/* Don't combine things to generate 64-bit loads/stores. We have to split
* those back into 32-bit ones anyway and UBO loads aren't split in NIR so