pad::examples);
}
-namespace pad_quan8_nnfw {
+namespace pad_quant8_nnfw {
std::vector<MixedTypedExample> examples = {
-// Generated pad_quan8_nnfw test
-#include "generated/examples/pad_quan8_nnfw.example.cpp"
+// Generated pad_quant8_nnfw test
+#include "generated/examples/pad_quant8_nnfw.example.cpp"
};
// Generated model constructor
-#include "generated/models/pad_quan8_nnfw.model.cpp"
-} // namespace pad_quan8_nnfw
-TEST_F(GeneratedTests, pad_quan8_nnfw) {
- execute(pad_quan8_nnfw::CreateModel,
- pad_quan8_nnfw::is_ignored,
- pad_quan8_nnfw::examples);
+#include "generated/models/pad_quant8_nnfw.model.cpp"
+} // namespace pad_quant8_nnfw
+TEST_F(GeneratedTests, pad_quant8_nnfw) {
+ execute(pad_quant8_nnfw::CreateModel,
+ pad_quant8_nnfw::is_ignored,
+ pad_quant8_nnfw::examples);
}
namespace space_to_batch_float_1 {
-// Generated file (from: pad_quan8_nnfw.mod.py). Do not edit
+// Generated file (from: pad_quant8_nnfw.mod.py). Do not edit
// Begin of an example
{
//Input(s)
// int -> INT32 map
{},
// int -> QUANT8_ASYMM map
- {{0, {0, 0, 0, 0, 0, 1, 2, 0, 0, 3, 4, 0, 0, 0, 0, 0}}}
+ {{0, {2, 2, 2, 2, 2, 1, 2, 2, 2, 3, 4, 2, 2, 2, 2, 2}}}
}
}, // End of an example
-// Generated file (from: pad_quan8_nnfw.mod.py). Do not edit
+// Generated file (from: pad_quant8_nnfw.mod.py). Do not edit
void CreateModel(Model *model) {
OperandType type1(Type::TENSOR_INT32, {4, 2});
OperandType type0(Type::TENSOR_QUANT8_ASYMM, {1, 2, 2, 1}, 1.0, 2);