arm_compute v17.12
[platform/upstream/armcl.git] / src / runtime / TensorAllocator.cpp
1 /*
2  * Copyright (c) 2016, 2017 ARM Limited.
3  *
4  * SPDX-License-Identifier: MIT
5  *
6  * Permission is hereby granted, free of charge, to any person obtaining a copy
7  * of this software and associated documentation files (the "Software"), to
8  * deal in the Software without restriction, including without limitation the
9  * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
10  * sell copies of the Software, and to permit persons to whom the Software is
11  * furnished to do so, subject to the following conditions:
12  *
13  * The above copyright notice and this permission notice shall be included in all
14  * copies or substantial portions of the Software.
15  *
16  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
19  * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
22  * SOFTWARE.
23  */
24 #include "arm_compute/runtime/TensorAllocator.h"
25
26 #include "arm_compute/core/Coordinates.h"
27 #include "arm_compute/core/Error.h"
28 #include "arm_compute/core/TensorInfo.h"
29 #include "arm_compute/runtime/MemoryGroup.h"
30 #include "support/ToolchainSupport.h"
31
32 #include <cstddef>
33
34 using namespace arm_compute;
35
36 namespace
37 {
38 bool validate_subtensor_shape(const TensorInfo &parent_info, const TensorInfo &child_info, const Coordinates &coords)
39 {
40     bool               is_valid     = true;
41     const TensorShape &parent_shape = parent_info.tensor_shape();
42     const TensorShape &child_shape  = child_info.tensor_shape();
43     const size_t       parent_dims  = parent_info.num_dimensions();
44     const size_t       child_dims   = child_info.num_dimensions();
45
46     if(child_dims <= parent_dims)
47     {
48         for(size_t num_dimensions = child_dims; num_dimensions > 0; --num_dimensions)
49         {
50             const size_t child_dim_size = coords[num_dimensions - 1] + child_shape[num_dimensions - 1];
51
52             if((coords[num_dimensions - 1] < 0) || (child_dim_size > parent_shape[num_dimensions - 1]))
53             {
54                 is_valid = false;
55                 break;
56             }
57         }
58     }
59     else
60     {
61         is_valid = false;
62     }
63
64     return is_valid;
65 }
66 } // namespace
67
68 TensorAllocator::TensorAllocator(Tensor *owner)
69     : _associated_memory_group(nullptr), _memory(), _owner(owner)
70 {
71 }
72
73 TensorAllocator::~TensorAllocator()
74 {
75     info().set_is_resizable(true);
76 }
77
78 TensorAllocator::TensorAllocator(TensorAllocator &&o) noexcept
79     : ITensorAllocator(std::move(o)),
80       _associated_memory_group(o._associated_memory_group),
81       _memory(std::move(o._memory)),
82       _owner(o._owner)
83 {
84     o._associated_memory_group = nullptr;
85     o._memory                  = Memory();
86     o._owner                   = nullptr;
87 }
88
89 TensorAllocator &TensorAllocator::operator=(TensorAllocator &&o) noexcept
90 {
91     if(&o != this)
92     {
93         _associated_memory_group   = o._associated_memory_group;
94         o._associated_memory_group = nullptr;
95
96         _memory   = std::move(o._memory);
97         o._memory = Memory();
98
99         _owner   = o._owner;
100         o._owner = nullptr;
101
102         ITensorAllocator::operator=(std::move(o));
103     }
104     return *this;
105 }
106
107 void TensorAllocator::init(const TensorAllocator &allocator, const Coordinates &coords, TensorInfo sub_info)
108 {
109     // Get parent info
110     const TensorInfo parent_info = allocator.info();
111
112     // Check if coordinates and new shape are within the parent tensor
113     ARM_COMPUTE_ERROR_ON(!validate_subtensor_shape(parent_info, sub_info, coords));
114     ARM_COMPUTE_UNUSED(validate_subtensor_shape);
115
116     // Copy pointer to buffer
117     _memory = Memory(allocator._memory.buffer());
118
119     // Init tensor info with new dimensions
120     size_t total_size = parent_info.offset_element_in_bytes(coords) + sub_info.total_size() - sub_info.offset_first_element_in_bytes();
121     sub_info.init(sub_info.tensor_shape(), sub_info.format(), parent_info.strides_in_bytes(), parent_info.offset_element_in_bytes(coords), total_size);
122
123     // Set TensorInfo
124     init(sub_info);
125 }
126
127 uint8_t *TensorAllocator::data() const
128 {
129     return _memory.buffer();
130 }
131
132 void TensorAllocator::allocate()
133 {
134     ARM_COMPUTE_ERROR_ON(_memory.buffer() != nullptr);
135     if(_associated_memory_group == nullptr)
136     {
137         _memory = Memory(std::shared_ptr<uint8_t>(new uint8_t[info().total_size()](), [](uint8_t *ptr)
138         {
139             delete[] ptr;
140         }));
141     }
142     else
143     {
144         _associated_memory_group->finalize_memory(_owner, reinterpret_cast<void **>(_memory.handle()), info().total_size());
145     }
146     info().set_is_resizable(false);
147 }
148
149 void TensorAllocator::free()
150 {
151     _memory = Memory();
152     info().set_is_resizable(true);
153 }
154
155 arm_compute::Status TensorAllocator::import_memory(Memory memory)
156 {
157     ARM_COMPUTE_RETURN_ERROR_ON(memory.buffer() == nullptr);
158     ARM_COMPUTE_RETURN_ERROR_ON(_associated_memory_group != nullptr);
159     _memory = memory;
160     info().set_is_resizable(false);
161
162     return Status{};
163 }
164
165 void TensorAllocator::set_associated_memory_group(MemoryGroup *associated_memory_group)
166 {
167     ARM_COMPUTE_ERROR_ON(associated_memory_group == nullptr);
168     ARM_COMPUTE_ERROR_ON(_associated_memory_group != nullptr);
169     ARM_COMPUTE_ERROR_ON(_memory.buffer() != nullptr);
170     _associated_memory_group = associated_memory_group;
171 }
172
173 uint8_t *TensorAllocator::lock()
174 {
175     return _memory.buffer();
176 }
177
178 void TensorAllocator::unlock()
179 {
180 }