From 824d0a98e4a70e8cead065526a7f878787b98302 Mon Sep 17 00:00:00 2001 From: =?utf8?q?=EB=B0=95=EC=84=B8=ED=9D=AC/On-Device=20Lab=28SR=29/Princip?= =?utf8?q?al=20Engineer/=EC=82=BC=EC=84=B1=EC=A0=84=EC=9E=90?= Date: Tue, 27 Aug 2019 13:18:51 +0900 Subject: [PATCH] [logo] Introduce ResolveRedundantReshapePass (#6940) This will introduce ResolveRedundantReshapePass copied from moco Signed-off-by: SaeHie Park --- .../include/logo/ResolveRedundantReshapePass.h | 44 +++++++++ .../src/Passes/ResolveRedundantReshapePass.cpp | 103 +++++++++++++++++++++ 2 files changed, 147 insertions(+) create mode 100644 compiler/logo/include/logo/ResolveRedundantReshapePass.h create mode 100644 compiler/logo/src/Passes/ResolveRedundantReshapePass.cpp diff --git a/compiler/logo/include/logo/ResolveRedundantReshapePass.h b/compiler/logo/include/logo/ResolveRedundantReshapePass.h new file mode 100644 index 0000000..3a2dc4f --- /dev/null +++ b/compiler/logo/include/logo/ResolveRedundantReshapePass.h @@ -0,0 +1,44 @@ +/* + * Copyright (c) 2019 Samsung Electronics Co., Ltd. All Rights Reserved + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +#ifndef __LOGO_RESOLVE_REDUNDANT_RESHAPE_PASS_H__ +#define __LOGO_RESOLVE_REDUNDANT_RESHAPE_PASS_H__ + +#include + +#include + +namespace logo +{ + +/** + * @brief Remove redundant canonical FixedReshape + * + * @note To effectively run this transform, canonical shape inference should be + * done ahead + */ +class ResolveRedundantReshapePass final : public Pass +{ +public: + const char *name(void) const final { return "ResolveRedundantReshapePass"; } + +public: + bool run(loco::Graph *graph) override; +}; + +} // namespace logo + +#endif // __LOGO_RESOLVE_REDUNDANT_RESHAPE_PASS_H__ diff --git a/compiler/logo/src/Passes/ResolveRedundantReshapePass.cpp b/compiler/logo/src/Passes/ResolveRedundantReshapePass.cpp new file mode 100644 index 0000000..da4af15 --- /dev/null +++ b/compiler/logo/src/Passes/ResolveRedundantReshapePass.cpp @@ -0,0 +1,103 @@ +/* + * Copyright (c) 2019 Samsung Electronics Co., Ltd. All Rights Reserved + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +#include + +#include + +#include + +#include + +namespace +{ + +bool shape_inference_done(loco::FixedReshape *reshape) +{ + return loco::shape_known(reshape) && loco::shape_known(reshape->input()); +} + +bool are_same_tensor_shapes(const loco::NodeShape &lhs, const loco::NodeShape &rhs) +{ + assert(lhs.domain() == loco::Domain::Tensor); + assert(rhs.domain() == loco::Domain::Tensor); + + auto lts = lhs.as(); + auto rts = rhs.as(); + + if (lts.rank() != rts.rank()) + return false; + + for (uint32_t axis = 0; axis < lts.rank(); ++axis) + { + assert(lts.dim(axis).known()); + assert(rts.dim(axis).known()); + if (lts.dim(axis).value() != rts.dim(axis).value()) + return false; + } + return true; +} + +/// @return true when 'reshape' has same input and output shape +bool is_redundant_reshape(loco::FixedReshape *reshape) +{ + auto input_shape = loco::shape_get(reshape->input()); + auto output_shape = loco::shape_get(reshape); + + // Note that FixedReshape's input and output are always tensor + return are_same_tensor_shapes(input_shape, output_shape); +} + +} // namespace + +namespace logo +{ + +/** + * @brief Bypass redundant FixedReshape + * + * Before: + * + * In ----- FixedReshape ----- [Out]* + * + * After: + * + * In ------------------------ [Out]* + * \ + * ------ FixedReshape + */ +bool ResolveRedundantReshapePass::run(loco::Graph *graph) +{ + bool changed = false; + for (auto node : loco::postorder_traversal(loco::output_nodes(graph))) + { + if (auto reshape = dynamic_cast(node)) + { + if (shape_inference_done(reshape)) + { + if (is_redundant_reshape(reshape)) + { + replace(reshape).with(reshape->input()); + changed = true; + } + } + } + } + + return changed; +} + +} // namespace logo -- 2.7.4