}
void TensorBuilder::registerSubTensorInfo(const graph::operand::Index &ind,
- const backend::operand::SubTensorInfo &info)
+ const compiler::SubTensorInfo &info)
{
assert(_tensors.size() == 0);
* @param[in] info Tensor information
*/
virtual void registerSubTensorInfo(const graph::operand::Index &ind,
- const backend::operand::SubTensorInfo &info) override;
+ const compiler::SubTensorInfo &info) override;
virtual void notifyFirstUse(const graph::operand::Index &) override;
virtual void notifyLastUse(const graph::operand::Index &) override;
private:
std::unordered_map<graph::operand::Index, ::arm_compute::TensorInfo> _tensor_info_map;
- std::unordered_map<graph::operand::Index, backend::operand::SubTensorInfo> _subtensor_info_map;
+ std::unordered_map<graph::operand::Index, compiler::SubTensorInfo> _subtensor_info_map;
std::unordered_map<graph::operand::Index, std::shared_ptr<::arm_compute::CLTensor>> _tensors;
std::unordered_map<graph::operand::Index, std::shared_ptr<::arm_compute::CLSubTensor>>
_subtensors;
}
void TensorBuilder::registerSubTensorInfo(const graph::operand::Index &,
- const backend::operand::SubTensorInfo &)
+ const compiler::SubTensorInfo &)
{
// Not supported yet
assert(false);
* @param[in] info Tensor information
*/
virtual void registerSubTensorInfo(const graph::operand::Index &ind,
- const backend::operand::SubTensorInfo &info) override;
+ const compiler::SubTensorInfo &info) override;
virtual void notifyFirstUse(const graph::operand::Index &) override;
virtual void notifyLastUse(const graph::operand::Index &) override;
#include "graph/operand/Index.h"
#include "operand/IObject.h"
-#include "backend/common/operand/SubTensorInfo.h"
+#include "compiler/SubTensorInfo.h"
namespace neurun
{
* @brief Register subtensor information to allocate on backend
*/
virtual void registerSubTensorInfo(const graph::operand::Index &,
- const backend::operand::SubTensorInfo &) = 0;
+ const compiler::SubTensorInfo &) = 0;
virtual void notifyFirstUse(const graph::operand::Index &) = 0;
virtual void notifyLastUse(const graph::operand::Index &) = 0;
* @brief This file contains SubTensorInfo to represent subsumption between tensors
* for backend tensor allocation
*/
-#ifndef __NEURUN_BACKEND_OPERAND_SUBTENSOR_INFO_H__
-#define __NEURUN_BACKEND_OPERAND_SUBTENSOR_INFO_H__
-
-#include <arm_compute/core/Coordinates.h>
-#include <arm_compute/core/TensorShape.h>
+#ifndef __NEURUN_COMPILER_SUBTENSOR_INFO_H__
+#define __NEURUN_COMPILER_SUBTENSOR_INFO_H__
#include "graph/operand/Object.h"
namespace neurun
{
-namespace backend
-{
-namespace operand
+namespace compiler
{
/**
const graph::operand::TypeInfo _type;
const graph::operand::ParentInfo::Coordinate4D _offset;
};
-} // operand
-} // backend
+
+} // compiler
} // neurun
-#endif // __NEURUN_BACKEND_OPERAND_SUBTENSOR_INFO_H__
+#endif // __NEURUN_COMPILER_SUBTENSOR_INFO_H__
#include "backend/interface/IStageGenerator.h"
#include "internal/Convert.h"
#include "backend/interface/IConfig.h"
-#include "backend/common/operand/SubTensorInfo.h"
+#include "compiler/SubTensorInfo.h"
#include "logging.h"
if (isSubTensor)
{
- const backend::operand::SubTensorInfo info(obj);
+ const compiler::SubTensorInfo info(obj);
tensor_builder->registerSubTensorInfo(ind, info);
}
else