2 * Copyright (c) 2018 Samsung Electronics Co., Ltd. All Rights Reserved
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
8 * http://www.apache.org/licenses/LICENSE-2.0
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
17 #include "gtest/gtest.h"
19 #include "tflite/ext/kernels/register.h"
20 #include "tensorflow/lite/model.h"
21 #include "tensorflow/lite/builtin_op_data.h"
25 #include "misc/environment.h"
27 #include "tflite/Diff.h"
28 #include "tflite/Quantization.h"
29 #include "tflite/interp/FunctionBuilder.h"
34 using namespace tflite;
35 using namespace nnfw::tflite;
37 TEST(NNAPI_Quickcheck_resize_bilinear_1, simple_test)
40 int SEED = std::chrono::system_clock::now().time_since_epoch().count();
42 nnfw::misc::env::IntAccessor("SEED").access(SEED);
44 // Set random test parameters
48 nnfw::misc::env::IntAccessor("VERBOSE").access(verbose);
49 nnfw::misc::env::IntAccessor("TOLERANCE").access(tolerance);
51 #define INT_VALUE(NAME, VALUE) IntVar NAME##_Value(#NAME, VALUE);
52 #include "resize_bilinear_1.lst"
55 const int32_t IFM_C = IFM_C_Value();
56 const int32_t IFM_H = IFM_H_Value();
57 const int32_t IFM_W = IFM_W_Value();
59 const int32_t OFM_C = IFM_C;
60 const int32_t OFM_H = OFM_H_Value();
61 const int32_t OFM_W = OFM_W_Value();
63 std::cout << "Configurations:" << std::endl;
64 #define PRINT_NEWLINE() \
66 std::cout << std::endl; \
68 #define PRINT_VALUE(value) \
70 std::cout << " " << #value << ": " << (value) << std::endl; \
86 int32_t size_data[2] = {OFM_H, OFM_W};
88 auto setup = [&](Interpreter &interp) {
89 // Comment from 'context.h'
91 // Parameters for asymmetric quantization. Quantized values can be converted
92 // back to float using:
93 // real_value = scale * (quantized_value - zero_point);
95 // Q: Is this necessary?
96 // A: This may be necessary, because quantization values(scale, zero_point) of TENSOR_INT32 and
97 // TENSOR_QUANT8_ASYMM are passed on to the runtime.
98 TfLiteQuantizationParams quantization = make_default_quantization();
100 // On AddTensors(N) call, T/F Lite interpreter creates N tensors whose index is [0 ~ N)
101 interp.AddTensors(3);
104 interp.SetTensorParametersReadWrite(0, kTfLiteFloat32 /* type */, "output" /* name */,
105 {1 /*N*/, OFM_H, OFM_W, OFM_C} /* dims */, quantization);
108 interp.SetTensorParametersReadWrite(1, kTfLiteFloat32 /* type */, "input" /* name */,
109 {1 /*N*/, IFM_H, IFM_W, IFM_C} /* dims */, quantization);
112 interp.SetTensorParametersReadOnly(
113 2, kTfLiteInt32 /* type */, "size" /* name */, {2} /* dims */, quantization,
114 reinterpret_cast<const char *>(size_data), 2 * sizeof(int32_t));
116 // NOTE AddNodeWithParameters take the ownership of param, and deallocate it with free
117 // So, param should be allocated with malloc
118 auto param = make_alloc<TfLiteResizeBilinearParams>();
120 // NOTE What is this?
121 param->align_corners = false;
123 interp.AddNodeWithParameters({1, 2}, {0}, nullptr, 0, reinterpret_cast<void *>(param),
124 BuiltinOpResolver().FindOp(BuiltinOperator_RESIZE_BILINEAR, 1));
126 // Set Tensor #1 as Input #0, and Tensor #0 as Output #0
127 interp.SetInputs({1});
128 interp.SetOutputs({0});
131 const nnfw::tflite::FunctionBuilder builder(setup);
133 RandomTestParam param;
135 param.verbose = verbose;
136 param.tolerance = tolerance;
138 int res = RandomTestRunner{SEED, param}.run(builder);