#include <armnn/TypesUtils.hpp>
#include "InferenceModel.hpp"
-
+#include <boost/core/ignore_unused.hpp>
#include <boost/program_options.hpp>
public:
virtual ~IInferenceTestCaseProvider() {}
- virtual void AddCommandLineOptions(boost::program_options::options_description& options) {};
- virtual bool ProcessCommandLineOptions(const InferenceTestOptions &commonOptions) { return true; };
+ virtual void AddCommandLineOptions(boost::program_options::options_description& options)
+ {
+ boost::ignore_unused(options);
+ };
+ virtual bool ProcessCommandLineOptions(const InferenceTestOptions &commonOptions)
+ {
+ boost::ignore_unused(commonOptions);
+ return true;
+ };
virtual std::unique_ptr<IInferenceTestCase> GetTestCase(unsigned int testCaseId) = 0;
virtual bool OnInferenceTestFinished() { return true; };
};
void operator()(const std::vector<int>& values)
{
+ boost::ignore_unused(values);
BOOST_ASSERT_MSG(false, "Non-float predictions output not supported.");
}
TestCaseResult ProcessResult(const InferenceTestOptions& options) override
{
+ boost::ignore_unused(options);
+
const std::vector<float>& output1 = boost::get<std::vector<float>>(this->GetOutputs()[0]); // bounding boxes
BOOST_ASSERT(output1.size() == k_OutputSize1);
const bool enableProfiling, const bool enableFp16TurboMode, const double& thresholdTime,
const bool printIntermediate, bool enableLayerDetails = false, bool parseUnuspported = false)
{
+ boost::ignore_unused(runtime);
std::string modelFormat;
std::string modelPath;
std::string inputNames;
"input", // input tensor name
"output", // output tensor name
{ 0, 1, 2 }, // test images to test with as above
- [&imageSet](const char* dataDir, const ModelType & model) {
+ [&imageSet](const char* dataDir, const ModelType &) {
// we need to get the input quantization parameters from
// the parsed model
return DatabaseType(
"input", // input tensor name
"InceptionV4/Logits/Predictions", // output tensor name
{ 0, 1, 2 }, // test images to test with as above
- [&imageSet](const char* dataDir, const ModelType & model) {
+ [&imageSet](const char* dataDir, const ModelType &) {
// we need to get the input quantization parameters from
// the parsed model
return DatabaseType(
"input", // input tensor name
"output", // output tensor name
{ 0, 1, 2 }, // test images to test with as above
- [&imageSet](const char* dataDir, const ModelType & model) {
+ [&imageSet](const char* dataDir, const ModelType &) {
return DatabaseType(
dataDir,
224,
"input", // input tensor name
"MobilenetV1/Predictions/Reshape_1", // output tensor name
{ 0, 1, 2 }, // test images to test with as above
- [&imageSet](const char* dataDir, const ModelType & model) {
+ [&imageSet](const char* dataDir, const ModelType &) {
// we need to get the input quantization parameters from
// the parsed model
return DatabaseType(
"input", // input tensor name
"MobilenetV1/Predictions/Reshape_1", // output tensor name
indices, // vector of indices to select which images to validate
- [&imageSet](const char* dataDir, const ModelType & model) {
+ [&imageSet](const char* dataDir, const ModelType &) {
// we need to get the input quantization parameters from
// the parsed model
return DatabaseType(
"input", // input tensor name
"output", // output tensor name
{ 0, 1, 2 }, // test images to test with as above
- [&imageSet](const char* dataDir, const ModelType & model) {
+ [&imageSet](const char* dataDir, const ModelType &) {
// we need to get the input quantization parameters from
// the parsed model
return DatabaseType(
"input", // input tensor name
"resnet_v2_50/predictions/Reshape_1", // output tensor name
{ 0, 1, 2 }, // test images to test with as above
- [&imageSet](const char* dataDir, const ModelType & model) {
+ [&imageSet](const char* dataDir, const ModelType &) {
// we need to get the input quantization parameters from
// the parsed model
return DatabaseType(
"input", // input tensor name
"output", // output tensor name
{ 0, 1, 2 }, // test images to test with as above
- [&imageSet](const char* dataDir, const ModelType & model) {
+ [&imageSet](const char* dataDir, const ModelType &) {
return DatabaseType(
dataDir,
299,
"input", // input tensor name
"vgg_16/fc8/squeezed", // output tensor name
{ 0, 1, 2 }, // test images to test with as above
- [&imageSet](const char* dataDir, const ModelType & model) {
+ [&imageSet](const char* dataDir, const ModelType &) {
// we need to get the input quantization parameters from
// the parsed model
return DatabaseType(
virtual TestCaseResult ProcessResult(const InferenceTestOptions& options) override
{
+ boost::ignore_unused(options);
+
using Boost3dArray = boost::multi_array<float, 3>;
const std::vector<float>& output = boost::get<std::vector<float>>(this->GetOutputs()[0]);