void Planner::visit(const graph::operation::Concat::Node &node)
{
- const ::neurun::graph::operand::Index ofm_index{node.getOutputs().at(0)};
-
// NOTE This implementation assumes concat over feature depth
// TODO Remove this assumption
assert(_ctx.at(::neurun::graph::operand::Index{node.param().axis_index}).asScalar<int32_t>() ==
void Planner::visit(const graph::operation::Reshape::Node &node)
{
- const ::neurun::graph::operand::Index output_index{node.getOutputs().at(0)};
- const ::neurun::graph::operand::Index input_index{node.getInputs().at(0)};
+ const auto output_index{node.getOutputs().at(0)};
+ const auto input_index{node.getInputs().at(0)};
// NOTE The content of a tensor specified by shape_index should be aligned with
// output tensor shape
{
VERBOSE(Softmax) << "Configure SOFTMAX operation" << std::endl;
- const ::neurun::graph::operand::Index output_index{node.getOutputs().at(0)};
- const ::neurun::graph::operand::Index input_index{node.getInputs().at(0)};
+ const auto output_index{node.getOutputs().at(0)};
+ const auto input_index{node.getInputs().at(0)};
assert(_ctx.at(output_index).shape().rank() == _ctx.at(input_index).shape().rank());
{
VERBOSE(Permute) << "Configure Permute operation" << std::endl;
- const ::neurun::graph::operand::Index output_index{node.getOutputs().at(0)};
- const ::neurun::graph::operand::Index input_index{node.getInputs().at(0)};
+ const auto output_index{node.getOutputs().at(0)};
+ const auto input_index{node.getInputs().at(0)};
assert(_ctx.at(output_index).shape().rank() == _ctx.at(input_index).shape().rank());