From 06304114286efd0773e0a6d702f94f322feca5e4 Mon Sep 17 00:00:00 2001 From: Bruno Goncalves Date: Thu, 27 Dec 2018 16:13:58 -0200 Subject: Added TfLiteParser test for MobilenetV2Quantized Change-Id: I0896c5f74d4e432a2ea8eff3cbe2db0702aa9ed9 --- tests/CMakeLists.txt | 6 +++ .../TfLiteMobilenetV2Quantized-Armnn.cpp | 63 ++++++++++++++++++++++ 2 files changed, 69 insertions(+) create mode 100644 tests/TfLiteMobilenetV2Quantized-Armnn/TfLiteMobilenetV2Quantized-Armnn.cpp (limited to 'tests') diff --git a/tests/CMakeLists.txt b/tests/CMakeLists.txt index e8f72eb4ee..90fd5ec783 100644 --- a/tests/CMakeLists.txt +++ b/tests/CMakeLists.txt @@ -171,6 +171,12 @@ if (BUILD_TF_LITE_PARSER) ObjectDetectionCommon.hpp) TfLiteParserTest(TfLiteMobileNetSsd-Armnn "${TfLiteMobileNetSsd-Armnn_sources}") + set(TfLiteMobilenetV2Quantized-Armnn_sources + TfLiteMobilenetV2Quantized-Armnn/TfLiteMobilenetV2Quantized-Armnn.cpp + ImagePreprocessor.hpp + ImagePreprocessor.cpp) + TfLiteParserTest(TfLiteMobilenetV2Quantized-Armnn "${TfLiteMobilenetV2Quantized-Armnn_sources}") + set(TfLiteVGG16Quantized-Armnn_sources TfLiteVGG16Quantized-Armnn/TfLiteVGG16Quantized-Armnn.cpp ImagePreprocessor.hpp diff --git a/tests/TfLiteMobilenetV2Quantized-Armnn/TfLiteMobilenetV2Quantized-Armnn.cpp b/tests/TfLiteMobilenetV2Quantized-Armnn/TfLiteMobilenetV2Quantized-Armnn.cpp new file mode 100644 index 0000000000..5db5c243b1 --- /dev/null +++ b/tests/TfLiteMobilenetV2Quantized-Armnn/TfLiteMobilenetV2Quantized-Armnn.cpp @@ -0,0 +1,63 @@ +// +// Copyright © 2017 Arm Ltd. All rights reserved. +// SPDX-License-Identifier: MIT +// +#include "../InferenceTest.hpp" +#include "../ImagePreprocessor.hpp" +#include "armnnTfLiteParser/ITfLiteParser.hpp" + +using namespace armnnTfLiteParser; + +int main(int argc, char* argv[]) +{ + int retVal = EXIT_FAILURE; + try + { + // Coverity fix: The following code may throw an exception of type std::length_error. + std::vector imageSet = + { + {"Dog.jpg", 209}, + {"Cat.jpg", 283}, + {"shark.jpg", 3}, + }; + + armnn::TensorShape inputTensorShape({ 1, 224, 224, 3 }); + + using DataType = uint8_t; + using DatabaseType = ImagePreprocessor; + using ParserType = armnnTfLiteParser::ITfLiteParser; + using ModelType = InferenceModel; + + // Coverity fix: ClassifierInferenceTestMain() may throw uncaught exceptions. + retVal = armnn::test::ClassifierInferenceTestMain( + argc, argv, + "mobilenet_v2_1.0_224_quant.tflite", // model name + true, // model is binary + "input", // input tensor name + "output", // output tensor name + { 0, 1, 2 }, // test images to test with as above + [&imageSet](const char* dataDir, const ModelType & model) { + // we need to get the input quantization parameters from + // the parsed model + auto inputBinding = model.GetInputBindingInfo(); + return DatabaseType( + dataDir, + 224, + 224, + imageSet, + inputBinding.second.GetQuantizationScale(), + inputBinding.second.GetQuantizationOffset()); + }, + &inputTensorShape); + } + catch (const std::exception& e) + { + // Coverity fix: BOOST_LOG_TRIVIAL (typically used to report errors) may throw an + // exception of type std::length_error. + // Using stderr instead in this context as there is no point in nesting try-catch blocks here. + std::cerr << "WARNING: " << *argv << ": An error has occurred when running " + "the classifier inference tests: " << e.what() << std::endl; + } + return retVal; +} -- cgit v1.2.1