}
FALLTHROUGH;
default:
- nir_src_copy(&tex->src[i].src, &old_tex->src[i].src, tex);
+ nir_src_copy(&tex->src[i].src, &old_tex->src[i].src, &tex->instr);
break;
}
}
nir_ssa_def *val = evaluate_rvalue(param_rvalue);
nir_src src = nir_src_for_ssa(val);
- nir_src_copy(&call->params[i], &src, call);
+ nir_src_copy(&call->params[i], &src, &call->instr);
} else if (sig_param->data.mode == ir_var_function_inout) {
unreachable("unimplemented: inout parameters");
}
}
}
-/* NOTE: if the instruction you are copying a src to is already added
- * to the IR, use nir_instr_rewrite_src() instead.
- */
-void nir_src_copy(nir_src *dest, const nir_src *src, void *mem_ctx)
+static void
+src_copy(nir_src *dest, const nir_src *src)
{
src_free_indirects(dest);
dest->reg.reg = src->reg.reg;
if (src->reg.indirect) {
dest->reg.indirect = calloc(1, sizeof(nir_src));
- nir_src_copy(dest->reg.indirect, src->reg.indirect, mem_ctx);
+ src_copy(dest->reg.indirect, src->reg.indirect);
} else {
dest->reg.indirect = NULL;
}
}
}
+/* NOTE: if the instruction you are copying a src to is already added
+ * to the IR, use nir_instr_rewrite_src() instead.
+ */
+void nir_src_copy(nir_src *dest, const nir_src *src, nir_instr *instr)
+{
+ src_copy(dest, src);
+}
+
void nir_dest_copy(nir_dest *dest, const nir_dest *src, nir_instr *instr)
{
/* Copying an SSA definition makes no sense whatsoever. */
nir_alu_src_copy(nir_alu_src *dest, const nir_alu_src *src,
nir_alu_instr *instr)
{
- nir_src_copy(&dest->src, &src->src, &instr->instr);
+ nir_src_copy(&dest->src, &src->src, instr ? &instr->instr : NULL);
dest->abs = src->abs;
dest->negate = src->negate;
for (unsigned i = 0; i < NIR_MAX_VEC_COMPONENTS; i++)
assert(!src_is_valid(src) || src->parent_if == if_stmt);
src_remove_all_uses(src);
- nir_src_copy(src, &new_src, if_stmt);
+ src_copy(src, &new_src);
src_add_all_uses(src, NULL, if_stmt);
}
return true;
}
-void nir_src_copy(nir_src *dest, const nir_src *src, void *instr_or_if);
+void nir_src_copy(nir_src *dest, const nir_src *src, nir_instr *instr);
void nir_dest_copy(nir_dest *dest, const nir_dest *src, nir_instr *instr);
typedef struct {
tex->src[i].src_type == nir_tex_src_sampler_offset ||
tex->src[i].src_type == nir_tex_src_texture_handle ||
tex->src[i].src_type == nir_tex_src_sampler_handle) {
- nir_src_copy(&txs->src[idx].src, &tex->src[i].src, txs);
+ nir_src_copy(&txs->src[idx].src, &tex->src[i].src, &txs->instr);
txs->src[idx].src_type = tex->src[i].src_type;
idx++;
}
tex->src[i].src_type == nir_tex_src_sampler_offset ||
tex->src[i].src_type == nir_tex_src_texture_handle ||
tex->src[i].src_type == nir_tex_src_sampler_handle) {
- nir_src_copy(&tql->src[idx].src, &tex->src[i].src, tql);
+ nir_src_copy(&tql->src[idx].src, &tex->src[i].src, &tql->instr);
tql->src[idx].src_type = tex->src[i].src_type;
idx++;
}
parent = rematerialize_deref_in_block(parent, state);
new_deref->parent = nir_src_for_ssa(&parent->dest.ssa);
} else {
- nir_src_copy(&new_deref->parent, &deref->parent, new_deref);
+ nir_src_copy(&new_deref->parent, &deref->parent, &new_deref->instr);
}
}
case nir_deref_type_array:
case nir_deref_type_ptr_as_array:
assert(!nir_src_as_deref(deref->arr.index));
- nir_src_copy(&new_deref->arr.index, &deref->arr.index, new_deref);
+ nir_src_copy(&new_deref->arr.index, &deref->arr.index, &new_deref->instr);
break;
case nir_deref_type_struct:
assert(src.reg.reg->num_components >= dest_src.reg.reg->num_components);
nir_alu_instr *mov = nir_alu_instr_create(b->shader, nir_op_mov);
- nir_src_copy(&mov->src[0].src, &src, mov);
+ nir_src_copy(&mov->src[0].src, &src, &mov->instr);
mov->dest.dest = nir_dest_for_reg(dest_src.reg.reg);
mov->dest.write_mask = (1 << dest_src.reg.reg->num_components) - 1;
/* remapped to ssbo_atomic_add: { buffer_idx, offset, +1 } */
temp = nir_imm_int(b, +1);
new_instr->src[0] = nir_src_for_ssa(buffer);
- nir_src_copy(&new_instr->src[1], &instr->src[0], new_instr);
+ nir_src_copy(&new_instr->src[1], &instr->src[0], &new_instr->instr);
new_instr->src[2] = nir_src_for_ssa(temp);
break;
case nir_intrinsic_atomic_counter_pre_dec:
/* NOTE semantic difference so we adjust the return value below */
temp = nir_imm_int(b, -1);
new_instr->src[0] = nir_src_for_ssa(buffer);
- nir_src_copy(&new_instr->src[1], &instr->src[0], new_instr);
+ nir_src_copy(&new_instr->src[1], &instr->src[0], &new_instr->instr);
new_instr->src[2] = nir_src_for_ssa(temp);
break;
case nir_intrinsic_atomic_counter_read:
/* remapped to load_ssbo: { buffer_idx, offset } */
new_instr->src[0] = nir_src_for_ssa(buffer);
- nir_src_copy(&new_instr->src[1], &instr->src[0], new_instr);
+ nir_src_copy(&new_instr->src[1], &instr->src[0], &new_instr->instr);
break;
default:
/* remapped to ssbo_atomic_x: { buffer_idx, offset, data, (compare)? } */
new_instr->src[0] = nir_src_for_ssa(buffer);
- nir_src_copy(&new_instr->src[1], &instr->src[0], new_instr);
- nir_src_copy(&new_instr->src[2], &instr->src[1], new_instr);
+ nir_src_copy(&new_instr->src[1], &instr->src[0], &new_instr->instr);
+ nir_src_copy(&new_instr->src[2], &instr->src[1], &new_instr->instr);
if (op == nir_intrinsic_ssbo_atomic_comp_swap ||
op == nir_intrinsic_ssbo_atomic_fcomp_swap)
- nir_src_copy(&new_instr->src[3], &instr->src[2], new_instr);
+ nir_src_copy(&new_instr->src[3], &instr->src[2], &new_instr->instr);
break;
}
/* Copy over any other sources. This is needed for interp_deref_at */
for (unsigned i = 1;
i < nir_intrinsic_infos[orig_instr->intrinsic].num_srcs; i++)
- nir_src_copy(&load->src[i], &orig_instr->src[i], load);
+ nir_src_copy(&load->src[i], &orig_instr->src[i], &load->instr);
nir_ssa_dest_init(&load->instr, &load->dest,
orig_instr->dest.ssa.num_components,
if (intrin->intrinsic == nir_intrinsic_interp_deref_at_sample ||
intrin->intrinsic == nir_intrinsic_interp_deref_at_offset ||
intrin->intrinsic == nir_intrinsic_interp_deref_at_vertex)
- nir_src_copy(&bary_setup->src[0], &intrin->src[1], bary_setup);
+ nir_src_copy(&bary_setup->src[0], &intrin->src[1], &bary_setup->instr);
nir_builder_instr_insert(b, &bary_setup->instr);
nir_alu_instr *mov = nir_alu_instr_create(b->shader, nir_op_mov);
- nir_src_copy(&mov->src[0].src, &intrin->src[1], mov);
+ nir_src_copy(&mov->src[0].src, &intrin->src[1], &mov->instr);
/* The normal NIR SSA copy propagate pass can't happen after this pass,
* so do an ad-hoc copy propagate since this ALU op can do swizzles
{
nir_intrinsic_instr *load = nir_intrinsic_instr_create(b->shader, op);
load->num_components = 1;
- nir_src_copy(&load->src[0], idx, load);
+ nir_src_copy(&load->src[0], idx, &load->instr);
nir_ssa_dest_init(&load->instr, &load->dest, 1, bitsize, NULL);
nir_builder_instr_insert(b, &load->instr);
return &load->dest.ssa;
}
if (is_store) {
- nir_src_copy(&global->src[0], &intr->src[0], global);
+ nir_src_copy(&global->src[0], &intr->src[0], &global->instr);
nir_intrinsic_set_write_mask(global, nir_intrinsic_write_mask(intr));
} else {
nir_ssa_dest_init(&global->instr, &global->dest,
intr->dest.ssa.bit_size, NULL);
if (is_atomic) {
- nir_src_copy(&global->src[1], &intr->src[2], global);
+ nir_src_copy(&global->src[1], &intr->src[2], &global->instr);
if (nir_intrinsic_infos[op].num_srcs > 2)
- nir_src_copy(&global->src[2], &intr->src[3], global);
+ nir_src_copy(&global->src[2], &intr->src[3], &global->instr);
}
}
intr->const_index[1] = intrin->const_index[1];
intr->src[0] = nir_src_for_ssa(comp);
if (nir_intrinsic_infos[intrin->intrinsic].num_srcs == 2)
- nir_src_copy(&intr->src[1], &intrin->src[1], intr);
+ nir_src_copy(&intr->src[1], &intrin->src[1], &intr->instr);
intr->num_components = 1;
nir_builder_instr_insert(b, &intr->instr);
/* invocation */
if (nir_intrinsic_infos[intrin->intrinsic].num_srcs > 1) {
assert(nir_intrinsic_infos[intrin->intrinsic].num_srcs == 2);
- nir_src_copy(&chan_intrin->src[1], &intrin->src[1], chan_intrin);
+ nir_src_copy(&chan_intrin->src[1], &intrin->src[1], &chan_intrin->instr);
}
chan_intrin->const_index[0] = intrin->const_index[0];
nir_intrinsic_instr *swizzle = nir_intrinsic_instr_create(
b->shader, nir_intrinsic_masked_swizzle_amd);
swizzle->num_components = intrin->num_components;
- nir_src_copy(&swizzle->src[0], &intrin->src[0], swizzle);
+ nir_src_copy(&swizzle->src[0], &intrin->src[0], &swizzle->instr);
nir_intrinsic_set_swizzle_mask(swizzle, (mask << 10) | 0x1f);
nir_ssa_dest_init(&swizzle->instr, &swizzle->dest,
intrin->dest.ssa.num_components,
nir_intrinsic_instr *shuffle =
nir_intrinsic_instr_create(b->shader, nir_intrinsic_shuffle);
shuffle->num_components = intrin->num_components;
- nir_src_copy(&shuffle->src[0], &intrin->src[0], shuffle);
+ nir_src_copy(&shuffle->src[0], &intrin->src[0], &shuffle->instr);
shuffle->src[1] = nir_src_for_ssa(index);
nir_ssa_dest_init(&shuffle->instr, &shuffle->dest,
intrin->dest.ssa.num_components,
qbcst->num_components = intrin->num_components;
qbcst->src[1] = nir_src_for_ssa(nir_imm_int(b, i));
- nir_src_copy(&qbcst->src[0], &intrin->src[0], qbcst);
+ nir_src_copy(&qbcst->src[0], &intrin->src[0], &qbcst->instr);
nir_ssa_dest_init(&qbcst->instr, &qbcst->dest,
intrin->dest.ssa.num_components,
intrin->dest.ssa.bit_size, NULL);
nir_tex_instr *plane_tex =
nir_tex_instr_create(b->shader, tex->num_srcs + 1);
for (unsigned i = 0; i < tex->num_srcs; i++) {
- nir_src_copy(&plane_tex->src[i].src, &tex->src[i].src, plane_tex);
+ nir_src_copy(&plane_tex->src[i].src, &tex->src[i].src, &plane_tex->instr);
plane_tex->src[i].src_type = tex->src[i].src_type;
}
plane_tex->src[tex->num_srcs].src = nir_src_for_ssa(nir_imm_int(b, plane));
/* reuse existing srcs */
for (unsigned i = 0; i < tex->num_srcs; i++) {
- nir_src_copy(&txd->src[i].src, &tex->src[i].src, txd);
+ nir_src_copy(&txd->src[i].src, &tex->src[i].src, &txd->instr);
txd->src[i].src_type = tex->src[i].src_type;
}
int coord = nir_tex_instr_src_index(tex, nir_tex_src_coord);
/* reuse all but bias src */
for (int i = 0; i < 2; i++) {
if (tex->src[i].src_type != nir_tex_src_bias) {
- nir_src_copy(&txl->src[i].src, &tex->src[i].src, txl);
+ nir_src_copy(&txl->src[i].src, &tex->src[i].src, &txl->instr);
txl->src[i].src_type = tex->src[i].src_type;
}
}
tex_copy->dest_type = tex->dest_type;
for (unsigned j = 0; j < tex->num_srcs; ++j) {
- nir_src_copy(&tex_copy->src[j].src, &tex->src[j].src, tex_copy);
+ nir_src_copy(&tex_copy->src[j].src, &tex->src[j].src, &tex_copy->instr);
tex_copy->src[j].src_type = tex->src[j].src_type;
}
nir_phi_instr *phi = nir_instr_as_phi(instr);
nir_alu_instr *sel = nir_alu_instr_create(shader, nir_op_bcsel);
- nir_src_copy(&sel->src[0].src, &if_stmt->condition, sel);
+ nir_src_copy(&sel->src[0].src, &if_stmt->condition, &sel->instr);
/* Splat the condition to all channels */
memset(sel->src[0].swizzle, 0, sizeof sel->src[0].swizzle);
assert(src->src.is_ssa);
unsigned idx = src->pred == then_block ? 1 : 2;
- nir_src_copy(&sel->src[idx].src, &src->src, sel);
+ nir_src_copy(&sel->src[idx].src, &src->src, &sel->instr);
}
nir_ssa_dest_init(&sel->instr, &sel->dest.dest,
assert(state->variables_seen & (1 << var->variable));
nir_alu_src val = { NIR_SRC_INIT };
- nir_alu_src_copy(&val, &state->variables[var->variable],
- (void *)build->shader);
+ nir_alu_src_copy(&val, &state->variables[var->variable], NULL);
assert(!var->is_constant);
for (unsigned i = 0; i < NIR_MAX_VEC_COMPONENTS; i++)
nir_intrinsic_set_dest_type(new_intrin, nir_intrinsic_dest_type(intrin));
/* offset */
- nir_src_copy(&new_intrin->src[0], &intrin->src[0], new_intrin);
+ nir_src_copy(&new_intrin->src[0], &intrin->src[0], &new_intrin->instr);
nir_builder_instr_insert(b, &new_intrin->instr);
nir_ssa_def_rewrite_uses(&alu->dest.dest.ssa,
mov->dest.dest.reg.base_offset = 0;
mov->dest.dest.reg.indirect = (nir_src *)calloc(1, sizeof(nir_src));
nir_src addr = nir_src_for_ssa(c2);
- nir_src_copy(mov->dest.dest.reg.indirect, &addr, mov);
+ nir_src_copy(mov->dest.dest.reg.indirect, &addr, &mov->instr);
nir_builder_instr_insert(&b, &mov->instr);
auto addr_reg = factory->src(addr, 0);
new_instr->src[0] = nir_src_for_ssa(&deref_arr->dest.ssa);
/* deref ops have no offset src, so copy the srcs after it */
for (unsigned i = 2; i < nir_intrinsic_infos[intr->intrinsic].num_srcs; i++)
- nir_src_copy(&new_instr->src[i - 1], &intr->src[i], new_instr);
+ nir_src_copy(&new_instr->src[i - 1], &intr->src[i], &new_instr->instr);
nir_builder_instr_insert(b, &new_instr->instr);
result[i] = &new_instr->dest.ssa;
nir_ssa_def *c = nir_channels(b, psrc->ssa, BITFIELD_MASK(nir_tex_instr_src_size(array_tex, s)));
array_tex->src[s].src = nir_src_for_ssa(c);
} else
- nir_src_copy(&array_tex->src[s].src, psrc, array_tex);
+ nir_src_copy(&array_tex->src[s].src, psrc, &array_tex->instr);
s++;
}
for (int i = 0; i < tex->num_srcs; i++) {
if (i == bias_idx)
continue;
- nir_src_copy(&txl->src[s].src, &tex->src[i].src, txl);
+ nir_src_copy(&txl->src[s].src, &tex->src[i].src, &txl->instr);
txl->src[s].src_type = tex->src[i].src_type;
s++;
}
}
FALLTHROUGH;
default:
- nir_src_copy(&tex->src[i].src, &old_tex->src[i].src, tex);
+ nir_src_copy(&tex->src[i].src, &old_tex->src[i].src, &tex->instr);
break;
}
}
nir_src *psrc = (tex->src[i].src_type == nir_tex_src_coord) ?
&coord_src : &tex->src[i].src;
- nir_src_copy(&array_tex->src[i].src, psrc, array_tex);
+ nir_src_copy(&array_tex->src[i].src, psrc, &array_tex->instr);
array_tex->src[i].src_type = tex->src[i].src_type;
}
nir_ssa_def *ssa_src = nir_channels(b, tex->src[coord_index].src.ssa,
(1 << coord_components) - 1);
nir_src src = nir_src_for_ssa(ssa_src);
- nir_src_copy(&tql->src[0].src, &src, tql);
+ nir_src_copy(&tql->src[0].src, &src, &tql->instr);
tql->src[0].src_type = nir_tex_src_coord;
unsigned idx = 1;
tex->src[i].src_type == nir_tex_src_sampler_offset ||
tex->src[i].src_type == nir_tex_src_texture_handle ||
tex->src[i].src_type == nir_tex_src_sampler_handle) {
- nir_src_copy(&tql->src[idx].src, &tex->src[i].src, tql);
+ nir_src_copy(&tql->src[idx].src, &tex->src[i].src, &tql->instr);
tql->src[idx].src_type = tex->src[i].src_type;
idx++;
}
if (tex->src[i].src_type == nir_tex_src_texture_deref ||
tex->src[i].src_type == nir_tex_src_texture_offset ||
tex->src[i].src_type == nir_tex_src_texture_handle) {
- nir_src_copy(&txf->src[idx].src, &tex->src[i].src, txf);
+ nir_src_copy(&txf->src[idx].src, &tex->src[i].src, &txf->instr);
txf->src[idx].src_type = tex->src[i].src_type;
idx++;
}
/* TODO: Indirect samplers, separate sampler objects XXX */
nir_src idx = nir_src_for_ssa(nir_imm_int(b, tex->texture_index));
- nir_src_copy(&l->src[0], &idx, l);
+ nir_src_copy(&l->src[0], &idx, &l->instr);
nir_builder_instr_insert(b, &l->instr);
nir_ssa_def *params = &l->dest.ssa;