const T * inputData,
std::vector<T>& outputData)
{
+ boost::ignore_unused(memoryManager);
BOOST_ASSERT_MSG(inputData != nullptr, "inputData must not be null");
if (inputData == nullptr)
{
unsigned int & concatDim,
TensorInfo & outputTensorInfo)
{
+ boost::ignore_unused(memoryManager);
BOOST_ASSERT_MSG(inputTensorInfos.size() > 1,
"Expecting more than one tensor to be concatenated here");
const IBackendInternal::IMemoryManagerSharedPtr& memoryManager,
bool useSubtensor)
{
+ boost::ignore_unused(memoryManager);
+
// Defines the tensor descriptors.
TensorInfo outputTensorInfo({ 3, 6, 3 }, ArmnnType);
TensorInfo inputTensorInfo1({ 3, 6, 2 }, ArmnnType);
IWorkloadFactory& workloadFactory,
const IBackendInternal::IMemoryManagerSharedPtr& memoryManager)
{
+ boost::ignore_unused(memoryManager);
+
unsigned int outputWidth = 3;
unsigned int outputHeight = 6;
unsigned int outputChannels = 3;
IWorkloadFactory& workloadFactory,
const IBackendInternal::IMemoryManagerSharedPtr& memoryManager)
{
+ boost::ignore_unused(memoryManager);
+
unsigned int outputWidth = 3;
unsigned int outputHeight = 6;
unsigned int outputChannels = 3;
IWorkloadFactory& workloadFactory,
const IBackendInternal::IMemoryManagerSharedPtr& memoryManager)
{
+ boost::ignore_unused(memoryManager);
+
unsigned int outputWidth = 3;
unsigned int outputHeight = 6;
unsigned int outputChannels = 3;
IWorkloadFactory& workloadFactory,
const IBackendInternal::IMemoryManagerSharedPtr& memoryManager)
{
+ boost::ignore_unused(memoryManager);
+
unsigned int outputWidth = 3;
unsigned int outputHeight = 6;
unsigned int outputChannels = 3;