if (!args.getDumpFilename().empty())
{
const std::string &dump_filename = args.getDumpFilename();
- TFLiteRun::TensorDumper output_dumper(*interpreter);
- output_dumper.dump(dump_filename);
+ TFLiteRun::TensorDumper tensor_dumper(*interpreter);
+ tensor_dumper.dump(dump_filename);
std::cout << "Output tensors have been dumped to file \"" << dump_filename << "\"."
<< std::endl;
}
std::cout << "Comparing the results with \"" << compare_filename << "\"." << std::endl;
std::cout << "========================================" << std::endl;
- TFLiteRun::TensorLoader output_loader(*interpreter);
- output_loader.load(compare_filename);
+ TFLiteRun::TensorLoader tensor_loader(*interpreter);
+ tensor_loader.load(compare_filename);
// TODO Code duplication (copied from RandomTestRunner)
for (const auto &o : interpreter->outputs())
{
- auto expected = output_loader.get(o);
+ auto expected = tensor_loader.get(o);
auto obtained = nnfw::support::tflite::TensorView<float>::make(*interpreter, o);
res = res && app.compareSingleTensorView(expected, obtained, o);