IVGCVSW-4482 Remove boost::ignore_unused
[platform/upstream/armnn.git] / src / backends / backendsCommon / test / layerTests / LogSoftmaxTestImpl.cpp
index 0b73d37..208bed2 100644 (file)
@@ -6,15 +6,14 @@
 #include "LogSoftmaxTestImpl.hpp"
 
 #include <Half.hpp>
+#include <QuantizeHelper.hpp>
 #include <ResolveType.hpp>
 
-#include <armnn/ArmNN.hpp>
 
 #include <backendsCommon/CpuTensorHandle.hpp>
-#include <backendsCommon/IBackendInternal.hpp>
+#include <armnn/backends/IBackendInternal.hpp>
 #include <backendsCommon/WorkloadFactory.hpp>
 
-#include <backendsCommon/test/QuantizeHelper.hpp>
 #include <backendsCommon/test/TensorCopyUtils.hpp>
 #include <backendsCommon/test/WorkloadTestUtils.hpp>
 
@@ -37,9 +36,10 @@ LayerTestResult<T, NumDims> LogSoftmaxTestImpl(
     float qScale = 1.0f,
     int32_t qOffset = 0)
 {
+    IgnoreUnused(memoryManager);
     LayerTestResult<T, NumDims> result(outputInfo);
     result.outputExpected =
-        MakeTensor<T, NumDims>(outputInfo, QuantizedVector<T>(qScale, qOffset, expectedOutputValues));
+        MakeTensor<T, NumDims>(outputInfo, armnnUtils::QuantizedVector<T>(expectedOutputValues, qScale, qOffset));
 
     std::unique_ptr<armnn::ITensorHandle> inputHandle  = workloadFactory.CreateTensorHandle(inputInfo);
     std::unique_ptr<armnn::ITensorHandle> outputHandle = workloadFactory.CreateTensorHandle(outputInfo);
@@ -54,7 +54,7 @@ LayerTestResult<T, NumDims> LogSoftmaxTestImpl(
     inputHandle->Allocate();
     outputHandle->Allocate();
 
-    auto inputTensor = MakeTensor<T, NumDims>(inputInfo, QuantizedVector<T>(qScale, qOffset, inputValues));
+    auto inputTensor = MakeTensor<T, NumDims>(inputInfo, armnnUtils::QuantizedVector<T>(inputValues, qScale, qOffset));
     CopyDataToITensorHandle(inputHandle.get(), inputTensor.origin());
 
     workload->Execute();