diff options
author | Nattapat Chaimanowong <nattapat.chaimanowong@arm.com> | 2018-10-26 10:24:14 +0100 |
---|---|---|
committer | nattapat.chaimanowong <nattapat.chaimanowong@arm.com> | 2018-10-26 12:38:34 +0000 |
commit | d8eee59735526ead6b87343c3ed9069e682b6e8c (patch) | |
tree | 44b994c98857b28674fb4911cb20489aaaecc437 /tests/TfLiteVGG16Quantized-Armnn/TfLiteVGG16Quantized-Armnn.cpp | |
parent | d4dfa684941a21314b70593d01b0fc2167eebad4 (diff) | |
download | armnn-d8eee59735526ead6b87343c3ed9069e682b6e8c.tar.gz |
IVGCVSW-2029 Fix fully connected layer support in TfLite Parser and implement test for TfLite VGG16 quantized
Change-Id: I2061f62f62684b963fa0f090718f1dcffe5c93ce
Diffstat (limited to 'tests/TfLiteVGG16Quantized-Armnn/TfLiteVGG16Quantized-Armnn.cpp')
-rw-r--r-- | tests/TfLiteVGG16Quantized-Armnn/TfLiteVGG16Quantized-Armnn.cpp | 68 |
1 files changed, 68 insertions, 0 deletions
diff --git a/tests/TfLiteVGG16Quantized-Armnn/TfLiteVGG16Quantized-Armnn.cpp b/tests/TfLiteVGG16Quantized-Armnn/TfLiteVGG16Quantized-Armnn.cpp new file mode 100644 index 0000000000..1313d2d01a --- /dev/null +++ b/tests/TfLiteVGG16Quantized-Armnn/TfLiteVGG16Quantized-Armnn.cpp @@ -0,0 +1,68 @@ +// +// Copyright © 2017 Arm Ltd. All rights reserved. +// SPDX-License-Identifier: MIT +// +#include "../InferenceTest.hpp" +#include "../ImagePreprocessor.hpp" +#include "armnnTfLiteParser/ITfLiteParser.hpp" + +using namespace armnnTfLiteParser; + +int main(int argc, char* argv[]) +{ + int retVal = EXIT_FAILURE; + try + { + // Coverity fix: The following code may throw an exception of type std::length_error. + std::vector<ImageSet> imageSet = + { + // Class number in probability print out offset by 1000 due to batch size fix + {"Dog.jpg", 669}, + {"Cat.jpg", 669}, + {"shark.jpg", 669}, + }; + + armnn::TensorShape inputTensorShape({ 2, 224, 224, 3 }); + + using DataType = uint8_t; + using DatabaseType = ImagePreprocessor<DataType>; + using ParserType = armnnTfLiteParser::ITfLiteParser; + using ModelType = InferenceModel<ParserType, DataType>; + + // Coverity fix: ClassifierInferenceTestMain() may throw uncaught exceptions. + retVal = armnn::test::ClassifierInferenceTestMain<DatabaseType, + ParserType>( + argc, argv, + "vgg_16_u8.tflite", // model name + true, // model is binary + "content_vgg/concat", // input tensor name + "content_vgg/prob", // output tensor name + { 0, 1, 2 }, // test images to test with as above + [&imageSet](const char* dataDir, const ModelType & model) { + // we need to get the input quantization parameters from + // the parsed model + auto inputBinding = model.GetInputBindingInfo(); + return DatabaseType( + dataDir, + 224, + 224, + imageSet, + inputBinding.second.GetQuantizationScale(), + inputBinding.second.GetQuantizationOffset(), + {{0, 0, 0}}, + {{1, 1, 1}}, + DatabaseType::DataFormat::NCHW, + 2); + }, + &inputTensorShape); + } + catch (const std::exception& e) + { + // Coverity fix: BOOST_LOG_TRIVIAL (typically used to report errors) may throw an + // exception of type std::length_error. + // Using stderr instead in this context as there is no point in nesting try-catch blocks here. + std::cerr << "WARNING: " << *argv << ": An error has occurred when running " + "the classifier inference tests: " << e.what() << std::endl; + } + return retVal; +} |