From 5ecddd222588be17d0694e8b85b181f89fcf2c28 Mon Sep 17 00:00:00 2001 From: =?utf8?q?=EC=9D=B4=ED=95=9C=EC=A2=85/On-Device=20Lab=28SR=29/Enginee?= =?utf8?q?r/=EC=82=BC=EC=84=B1=EC=A0=84=EC=9E=90?= Date: Thu, 7 Mar 2019 13:33:02 +0900 Subject: [PATCH] [neurun] Fix build break due to namespace change (#4625) Remove specifying `compiler::` namespace for `IExecutionBuilder`. Use `using` at the top of the source code instead. Signed-off-by: Hanjoung Lee --- .../neurun/src/backend/acl_cl/StageGenerator.cc | 38 ++++++++++++---------- 1 file changed, 20 insertions(+), 18 deletions(-) diff --git a/runtimes/neurun/src/backend/acl_cl/StageGenerator.cc b/runtimes/neurun/src/backend/acl_cl/StageGenerator.cc index a0d7ad7..c3f711b 100644 --- a/runtimes/neurun/src/backend/acl_cl/StageGenerator.cc +++ b/runtimes/neurun/src/backend/acl_cl/StageGenerator.cc @@ -52,6 +52,8 @@ #include "Swizzle.h" +using ::neurun::compiler::IExecutionBuilder; + template std::unique_ptr make_layer(void) { return std::unique_ptr{new T}; } std::unique_ptr<::neurun::backend::acl_cl::kernel::CLFunction> @@ -86,7 +88,7 @@ namespace acl_cl class ActivationBuilder { public: - ActivationBuilder(compiler::IExecutionBuilder &builder) : _builder(builder) + ActivationBuilder(IExecutionBuilder &builder) : _builder(builder) { // DO NOTHING } @@ -100,7 +102,7 @@ public: void append(FuseCode code, ::arm_compute::ICLTensor *tensor); private: - compiler::IExecutionBuilder &_builder; + IExecutionBuilder &_builder; }; void ActivationBuilder::appendReLU(::arm_compute::ICLTensor *ifm_alloc) @@ -205,7 +207,7 @@ void StageGenerator::visit(const model::operation::CastNode &node) auto tensors = _tensor_builder; - returnStage([tensors, param](compiler::IExecutionBuilder &builder) { + returnStage([tensors, param](IExecutionBuilder &builder) { auto ofm_alloc = tensors->at(param.ofm_index).get(); auto ifm_alloc = tensors->at(param.ifm_index).get(); @@ -303,7 +305,7 @@ void StageGenerator::visit(const model::operation::Conv2DNode &node) auto tensors = _tensor_builder; - returnStage([tensors, param](compiler::IExecutionBuilder &builder) { + returnStage([tensors, param](IExecutionBuilder &builder) { auto ofm_alloc = tensors->at(param.ofm_index).get(); auto ifm_alloc = tensors->at(param.ifm_index).get(); auto ker_alloc = tensors->at(param.ker_index).get(); @@ -407,7 +409,7 @@ void StageGenerator::visit(const model::operation::DepthwiseConv2DNode &node) auto tensors = _tensor_builder; - returnStage([tensors, param](compiler::IExecutionBuilder &builder) { + returnStage([tensors, param](IExecutionBuilder &builder) { auto ofm_alloc = tensors->at(param.ofm_index).get(); auto ifm_alloc = tensors->at(param.ifm_index).get(); auto ker_alloc = tensors->at(param.ker_index).get(); @@ -524,7 +526,7 @@ void StageGenerator::visit(const model::operation::MaxPool2DNode &node) auto tensors = _tensor_builder; - returnStage([tensors, param](compiler::IExecutionBuilder &builder) { + returnStage([tensors, param](IExecutionBuilder &builder) { auto ofm_alloc = tensors->at(param.ofm_index).get(); auto ifm_alloc = tensors->at(param.ifm_index).get(); @@ -639,7 +641,7 @@ void StageGenerator::visit(const model::operation::AvgPool2DNode &node) auto tensors = _tensor_builder; - returnStage([tensors, param](compiler::IExecutionBuilder &builder) { + returnStage([tensors, param](IExecutionBuilder &builder) { auto ofm_alloc = tensors->at(param.ofm_index).get(); auto ifm_alloc = tensors->at(param.ifm_index).get(); @@ -685,7 +687,7 @@ void StageGenerator::visit(const model::operation::ConcatNode &node) auto tensors = _tensor_builder; - returnStage([tensors, param](compiler::IExecutionBuilder &builder) { + returnStage([tensors, param](IExecutionBuilder &builder) { // If tensor allocator allocate as subtensor bool canEliminate = true; for (auto ifm_ind : param.input_indexes) @@ -794,7 +796,7 @@ void StageGenerator::visit(const model::operation::FullyConnectedNode &node) param.needs_reshape = needs_reshape; param.reshape = reshape; - returnStage([tensors, param](compiler::IExecutionBuilder &builder) { + returnStage([tensors, param](IExecutionBuilder &builder) { auto output_alloc = tensors->at(param.output_index).get(); auto input_alloc = tensors->at(param.input_index).get(); auto weight_alloc = tensors->at(param.weight_index).get(); @@ -852,7 +854,7 @@ void StageGenerator::visit(const model::operation::MulNode &node) auto tensors = _tensor_builder; - returnStage([tensors, param](compiler::IExecutionBuilder &builder) { + returnStage([tensors, param](IExecutionBuilder &builder) { auto ofm_alloc = tensors->at(param.ofm_index).get(); auto lhs_alloc = tensors->at(param.lhs_index).get(); auto rhs_alloc = tensors->at(param.rhs_index).get(); @@ -896,7 +898,7 @@ void StageGenerator::visit(const model::operation::ReshapeNode &node) tensors->dimCorrection(input_index, false); tensors->dimCorrection(output_index, false); - returnStage([tensors, param](compiler::IExecutionBuilder &builder) { + returnStage([tensors, param](IExecutionBuilder &builder) { auto output_alloc = tensors->at(param.output_index).get(); auto input_alloc = tensors->at(param.input_index).get(); @@ -928,7 +930,7 @@ void StageGenerator::visit(const model::operation::TanhNode &node) auto tensors = _tensor_builder; - returnStage([tensors, param](compiler::IExecutionBuilder &builder) { + returnStage([tensors, param](IExecutionBuilder &builder) { auto output_alloc = tensors->at(param.output_index).get(); auto input_alloc = tensors->at(param.input_index).get(); @@ -966,7 +968,7 @@ void StageGenerator::visit(const model::operation::SoftmaxNode &node) auto tensors = _tensor_builder; - returnStage([tensors, param](compiler::IExecutionBuilder &builder) { + returnStage([tensors, param](IExecutionBuilder &builder) { auto output_alloc = tensors->at(param.output_index).get(); auto input_alloc = tensors->at(param.input_index).get(); @@ -1067,7 +1069,7 @@ void StageGenerator::visit(const model::operation::StridedSliceNode &node) auto tensors = _tensor_builder; - returnStage([tensors, param](compiler::IExecutionBuilder &builder) { + returnStage([tensors, param](IExecutionBuilder &builder) { auto outputData_alloc = tensors->at(param.outputData_index).get(); auto inputData_alloc = tensors->at(param.inputData_index).get(); @@ -1121,7 +1123,7 @@ void StageGenerator::visit(const model::operation::TransposeNode &node) auto tensors = _tensor_builder; - returnStage([tensors, param](compiler::IExecutionBuilder &builder) { + returnStage([tensors, param](IExecutionBuilder &builder) { auto ofm_alloc = tensors->at(param.ofm_idx).get(); auto ifm_alloc = tensors->at(param.ifm_idx).get(); const auto pv = getARMComputePermutationVector(param.rank, param.pv); @@ -1176,7 +1178,7 @@ void StageGenerator::visit(const model::operation::AddNode &node) auto tensors = _tensor_builder; - returnStage([tensors, param](compiler::IExecutionBuilder &builder) { + returnStage([tensors, param](IExecutionBuilder &builder) { auto ofm_alloc = tensors->at(param.ofm_index).get(); auto lhs_alloc = tensors->at(param.lhs_index).get(); auto rhs_alloc = tensors->at(param.rhs_index).get(); @@ -1237,7 +1239,7 @@ void StageGenerator::visit(const model::operation::SubNode &node) auto tensors = _tensor_builder; - returnStage([tensors, param](compiler::IExecutionBuilder &builder) { + returnStage([tensors, param](IExecutionBuilder &builder) { auto ofm_alloc = tensors->at(param.ofm_index).get(); auto lhs_alloc = tensors->at(param.lhs_index).get(); auto rhs_alloc = tensors->at(param.rhs_index).get(); @@ -1299,7 +1301,7 @@ void StageGenerator::visit(const model::operation::DivNode &node) auto tensors = _tensor_builder; - returnStage([tensors, param](compiler::IExecutionBuilder &builder) { + returnStage([tensors, param](IExecutionBuilder &builder) { auto ofm_alloc = tensors->at(param.ofm_index).get(); auto lhs_alloc = tensors->at(param.lhs_index).get(); auto rhs_alloc = tensors->at(param.rhs_index).get(); -- 2.7.4