1 // Copyright (C) 2019 Intel Corporation
2 // SPDX-License-Identifier: Apache-2.0
12 #include <ie_layers.h>
13 #include <ie_memcpy.h>
14 #include "ie_const_infer_impl.hpp"
16 namespace InferenceEngine {
17 namespace ShapeInfer {
20 *@brief Implementation of Const inference for Tile layer
22 class TileConstInfer : public ConstInferImpl {
24 explicit TileConstInfer(const std::string& type) : ConstInferImpl(type) {}
26 void inferImpl(const std::vector<Blob::CPtr>& inData,
27 const std::map<std::string, std::string>& params,
28 const std::map<std::string, Blob::Ptr>& blobs,
29 std::vector<Blob::Ptr>& outData) override {
32 layer.params = params;
34 _validator->parseParams(&layer);
36 auto inBlob = *inData.begin();
37 SizeVector inShape = inBlob->getTensorDesc().getDims();
38 const auto* inBuffer = inBlob->cbuffer().as<float*>();
40 auto outBlob = *outData.begin();
41 auto* outBuffer = outBlob->buffer().as<float*>();
46 for (int i = 0; i < layer.axis; i++) m_outer_dim *= inShape[i];
47 for (int i = layer.axis; i < inShape.size(); i++) m_inner_dim *= inShape[i];
49 for (int i = 0; i < m_outer_dim; ++i) {
50 for (int t = 0; t < layer.tiles; ++t) {
51 ie_memcpy(outBuffer, outBlob->byteSize(), inBuffer, m_inner_dim * sizeof(float));
52 outBuffer += m_inner_dim;
54 inBuffer += m_inner_dim;
59 } // namespace ShapeInfer
60 } // namespace InferenceEngine