inputTensorDataLayout == armnn::DataLayout::NCHW ? inputTensorShape[3] : inputTensorShape[2];
const unsigned int inputTensorHeight =
inputTensorDataLayout == armnn::DataLayout::NCHW ? inputTensorShape[2] : inputTensorShape[1];
+ // Get output tensor info
+ const unsigned int outputNumElements = model.GetOutputSize();
+
const unsigned int batchSize = 1;
// Get normalisation parameters
SupportedFrontend modelFrontend;
normParams,
batchSize,
inputTensorDataLayout));
- outputDataContainers = {vector<int>(1001)};
+ outputDataContainers = { vector<int>(outputNumElements) };
break;
case armnn::DataType::QuantisedAsymm8:
inputDataContainers.push_back(
normParams,
batchSize,
inputTensorDataLayout));
- outputDataContainers = {vector<uint8_t>(1001)};
+ outputDataContainers = { vector<uint8_t>(outputNumElements) };
break;
case armnn::DataType::Float32:
default:
normParams,
batchSize,
inputTensorDataLayout));
- outputDataContainers = {vector<float>(1001)};
+ outputDataContainers = { vector<float>(outputNumElements) };
break;
}