diff options
author | SiCong Li <sicong.li@arm.com> | 2019-06-21 12:00:04 +0100 |
---|---|---|
committer | Nikhil Raj Arm <nikhil.raj@arm.com> | 2019-07-09 12:05:05 +0000 |
commit | 39f4639a79625c8f37c6ca547dadc7925378ee3e (patch) | |
tree | 36bc9d1fccff2a25a40003f2a2b7ecf4c77de3f6 /tests/ImageTensorGenerator | |
parent | d01a83c8de77c44a938a618918d17385da3baa88 (diff) | |
download | armnn-39f4639a79625c8f37c6ca547dadc7925378ee3e.tar.gz |
MLCE-103 Use ImageTensorGenerator in ModelAccuracyTool
* Refactor ImageTensorGenerator into a header so that it can be used
inside ModelAccuracyTool. ModelAccuracyTool now can accept image files
instead of tensor text files as input. ImageTensorGenerator remains a
standalone tool for converting images into tensors text files.
* Inside the ImageTensorGenerator, use the existing image preprocessing
method InferenceTestImage::Resize which has the correct image normalization
techniques and other resize utilities.
Change-Id: Ia662fed4752fb81c5cfa6d15967c6aae4aaf1155
Signed-off-by: SiCong Li <sicong.li@arm.com>
Diffstat (limited to 'tests/ImageTensorGenerator')
-rw-r--r-- | tests/ImageTensorGenerator/ImageTensorGenerator.cpp | 123 | ||||
-rw-r--r-- | tests/ImageTensorGenerator/ImageTensorGenerator.hpp | 156 |
2 files changed, 261 insertions, 18 deletions
diff --git a/tests/ImageTensorGenerator/ImageTensorGenerator.cpp b/tests/ImageTensorGenerator/ImageTensorGenerator.cpp index 1f537745b4..f391a27a4d 100644 --- a/tests/ImageTensorGenerator/ImageTensorGenerator.cpp +++ b/tests/ImageTensorGenerator/ImageTensorGenerator.cpp @@ -3,13 +3,16 @@ // SPDX-License-Identifier: MIT // +#include "ImageTensorGenerator.hpp" #include "../InferenceTestImage.hpp" +#include <armnn/TypesUtils.hpp> #include <boost/filesystem.hpp> #include <boost/filesystem/operations.hpp> #include <boost/filesystem/path.hpp> #include <boost/log/trivial.hpp> #include <boost/program_options.hpp> +#include <boost/variant.hpp> #include <algorithm> #include <fstream> @@ -57,10 +60,7 @@ public: return false; } - std::vector<std::string> supportedLayouts = { - "NHWC", - "NCHW" - }; + std::vector<std::string> supportedLayouts = { "NHWC", "NCHW" }; auto iterator = std::find(supportedLayouts.begin(), supportedLayouts.end(), layout); if (iterator == supportedLayouts.end()) @@ -113,10 +113,20 @@ public: ("help,h", "Display help messages") ("infile,i", po::value<std::string>(&m_InputFileName)->required(), "Input image file to generate tensor from") - ("layout,l", po::value<std::string>(&m_Layout)->default_value("NHWC"), - "Output data layout, \"NHWC\" or \"NCHW\", default value NHWC") + ("model-format,f", po::value<std::string>(&m_ModelFormat)->required(), + "Format of the model file, Accepted values (caffe, tensorflow, tflite)") ("outfile,o", po::value<std::string>(&m_OutputFileName)->required(), - "Output raw tensor file path"); + "Output raw tensor file path") + ("output-type,z", po::value<std::string>(&m_OutputType)->default_value("float"), + "The data type of the output tensors." + "If unset, defaults to \"float\" for all defined inputs. " + "Accepted values (float, int or qasymm8)") + ("new-width,w", po::value<std::string>(&m_NewWidth)->default_value("0"), + "Resize image to new width. Keep original width if unspecified") + ("new-height,h", po::value<std::string>(&m_NewHeight)->default_value("0"), + "Resize image to new height. Keep original height if unspecified") + ("layout,l", po::value<std::string>(&m_Layout)->default_value("NHWC"), + "Output data layout, \"NHWC\" or \"NCHW\", default value NHWC"); } catch (const std::exception& e) { @@ -164,13 +174,71 @@ public: } std::string GetInputFileName() {return m_InputFileName;} - std::string GetLayout() {return m_Layout;} + armnn::DataLayout GetLayout() + { + if (m_Layout == "NHWC") + { + return armnn::DataLayout::NHWC; + } + else if (m_Layout == "NCHW") + { + return armnn::DataLayout::NCHW; + } + else + { + throw armnn::Exception("Unsupported data layout: " + m_Layout); + } + } std::string GetOutputFileName() {return m_OutputFileName;} + unsigned int GetNewWidth() {return static_cast<unsigned int>(std::stoi(m_NewWidth));} + unsigned int GetNewHeight() {return static_cast<unsigned int>(std::stoi(m_NewHeight));} + SupportedFrontend GetModelFormat() + { + if (m_ModelFormat == "caffe") + { + return SupportedFrontend::Caffe; + } + else if (m_ModelFormat == "tensorflow") + { + return SupportedFrontend::TensorFlow; + } + else if (m_ModelFormat == "tflite") + { + return SupportedFrontend::TFLite; + } + else + { + throw armnn::Exception("Unsupported model format" + m_ModelFormat); + } + } + armnn::DataType GetOutputType() + { + if (m_OutputType == "float") + { + return armnn::DataType::Float32; + } + else if (m_OutputType == "int") + { + return armnn::DataType::Signed32; + } + else if (m_OutputType == "qasymm8") + { + return armnn::DataType::QuantisedAsymm8; + } + else + { + throw armnn::Exception("Unsupported input type" + m_OutputType); + } + } private: std::string m_InputFileName; std::string m_Layout; std::string m_OutputFileName; + std::string m_NewWidth; + std::string m_NewHeight; + std::string m_ModelFormat; + std::string m_OutputType; }; } // namespace anonymous @@ -182,18 +250,36 @@ int main(int argc, char* argv[]) { return -1; } - const std::string imagePath(cmdline.GetInputFileName()); const std::string outputPath(cmdline.GetOutputFileName()); - - // generate image tensor - std::vector<float> imageData; + const SupportedFrontend& modelFormat(cmdline.GetModelFormat()); + const armnn::DataType outputType(cmdline.GetOutputType()); + const unsigned int newWidth = cmdline.GetNewWidth(); + const unsigned int newHeight = cmdline.GetNewHeight(); + const unsigned int batchSize = 1; + const armnn::DataLayout outputLayout(cmdline.GetLayout()); + + using TContainer = boost::variant<std::vector<float>, std::vector<int>, std::vector<uint8_t>>; + std::vector<TContainer> imageDataContainers; + const NormalizationParameters& normParams = GetNormalizationParameters(modelFormat, outputType); try { - InferenceTestImage testImage(imagePath.c_str()); - imageData = cmdline.GetLayout() == "NHWC" - ? GetImageDataAsNormalizedFloats(ImageChannelLayout::Rgb, testImage) - : GetImageDataInArmNnLayoutAsNormalizedFloats(ImageChannelLayout::Rgb, testImage); + switch (outputType) + { + case armnn::DataType::Signed32: + imageDataContainers.push_back(PrepareImageTensor<int>( + imagePath, newWidth, newHeight, normParams, batchSize, outputLayout)); + break; + case armnn::DataType::QuantisedAsymm8: + imageDataContainers.push_back(PrepareImageTensor<uint8_t>( + imagePath, newWidth, newHeight, normParams, batchSize, outputLayout)); + break; + case armnn::DataType::Float32: + default: + imageDataContainers.push_back(PrepareImageTensor<float>( + imagePath, newWidth, newHeight, normParams, batchSize, outputLayout)); + break; + } } catch (const InferenceTestImageException& e) { @@ -205,7 +291,8 @@ int main(int argc, char* argv[]) imageTensorFile.open(outputPath, std::ofstream::out); if (imageTensorFile.is_open()) { - std::copy(imageData.begin(), imageData.end(), std::ostream_iterator<float>(imageTensorFile, " ")); + boost::apply_visitor([&imageTensorFile](auto&& imageData) { WriteImageTensorImpl(imageData, imageTensorFile); }, + imageDataContainers[0]); if (!imageTensorFile) { BOOST_LOG_TRIVIAL(fatal) << "Failed to write to output file" << outputPath; @@ -221,4 +308,4 @@ int main(int argc, char* argv[]) } return 0; -}
\ No newline at end of file +} diff --git a/tests/ImageTensorGenerator/ImageTensorGenerator.hpp b/tests/ImageTensorGenerator/ImageTensorGenerator.hpp new file mode 100644 index 0000000000..987487d7bb --- /dev/null +++ b/tests/ImageTensorGenerator/ImageTensorGenerator.hpp @@ -0,0 +1,156 @@ +// +// Copyright © 2017 Arm Ltd. All rights reserved. +// SPDX-License-Identifier: MIT +// + +#include "../InferenceTestImage.hpp" +#include "Permute.hpp" +#include <armnn/TypesUtils.hpp> + +#include <algorithm> +#include <fstream> +#include <iterator> +#include <string> + +struct NormalizationParameters +{ + float scale{ 1.0 }; + std::array<float, 3> mean{ 0.0, 0.0, 0.0 }; + std::array<float, 3> stddev{ 1.0, 1.0, 1.0 }; +}; + +enum class SupportedFrontend +{ + Caffe = 0, + TensorFlow = 1, + TFLite = 2, +}; + +// Get normalization parameters. +// Note that different flavours of models have different normalization methods. +// This tool currently only supports Caffe, TF and TFLite models +NormalizationParameters GetNormalizationParameters(const SupportedFrontend& modelFormat, + const armnn::DataType& outputType) +{ + NormalizationParameters normParams; + // Explicitly set default parameters + normParams.scale = 1.0; + normParams.mean = { 0.0, 0.0, 0.0 }; + normParams.stddev = { 1.0, 1.0, 1.0 }; + switch (modelFormat) + { + case SupportedFrontend::Caffe: + break; + case SupportedFrontend::TensorFlow: + case SupportedFrontend::TFLite: + default: + switch (outputType) + { + case armnn::DataType::Float32: + normParams.scale = 127.5; + normParams.mean = { 1.0, 1.0, 1.0 }; + break; + case armnn::DataType::Signed32: + normParams.mean = { 128.0, 128.0, 128.0 }; + break; + case armnn::DataType::QuantisedAsymm8: + default: + break; + } + break; + } + return normParams; +} + +// Prepare raw image tensor data by loading the image from imagePath and preprocessing it. +template <typename ElemType> +std::vector<ElemType> PrepareImageTensor(const std::string& imagePath, + unsigned int newWidth, + unsigned int newHeight, + const NormalizationParameters& normParams, + unsigned int batchSize = 1, + const armnn::DataLayout& outputLayout = armnn::DataLayout::NHWC); + +// Prepare float32 image tensor +template <> +std::vector<float> PrepareImageTensor<float>(const std::string& imagePath, + unsigned int newWidth, + unsigned int newHeight, + const NormalizationParameters& normParams, + unsigned int batchSize, + const armnn::DataLayout& outputLayout) +{ + // Generate image tensor + std::vector<float> imageData; + InferenceTestImage testImage(imagePath.c_str()); + if (newWidth == 0) + { + newWidth = testImage.GetWidth(); + } + if (newHeight == 0) + { + newHeight = testImage.GetHeight(); + } + // Resize the image to new width and height or keep at original dimensions if the new width and height are specified + // as 0 Centre/Normalise the image. + imageData = testImage.Resize(newWidth, newHeight, CHECK_LOCATION(), + InferenceTestImage::ResizingMethods::BilinearAndNormalized, normParams.mean, + normParams.stddev, normParams.scale); + if (outputLayout == armnn::DataLayout::NCHW) + { + // Convert to NCHW format + const armnn::PermutationVector NHWCToArmNN = { 0, 2, 3, 1 }; + armnn::TensorShape dstShape({ batchSize, 3, newHeight, newWidth }); + std::vector<float> tempImage(imageData.size()); + armnnUtils::Permute(dstShape, NHWCToArmNN, imageData.data(), tempImage.data(), sizeof(float)); + imageData.swap(tempImage); + } + return imageData; +} + +// Prepare int32 image tensor +template <> +std::vector<int> PrepareImageTensor<int>(const std::string& imagePath, + unsigned int newWidth, + unsigned int newHeight, + const NormalizationParameters& normParams, + unsigned int batchSize, + const armnn::DataLayout& outputLayout) +{ + // Get float32 image tensor + std::vector<float> imageDataFloat = + PrepareImageTensor<float>(imagePath, newWidth, newHeight, normParams, batchSize, outputLayout); + // Convert to int32 image tensor with static cast + std::vector<int> imageDataInt; + imageDataInt.reserve(imageDataFloat.size()); + std::transform(imageDataFloat.begin(), imageDataFloat.end(), std::back_inserter(imageDataInt), + [](float val) { return static_cast<int>(val); }); + return imageDataInt; +} + +// Prepare qasymm8 image tensor +template <> +std::vector<uint8_t> PrepareImageTensor<uint8_t>(const std::string& imagePath, + unsigned int newWidth, + unsigned int newHeight, + const NormalizationParameters& normParams, + unsigned int batchSize, + const armnn::DataLayout& outputLayout) +{ + // Get float32 image tensor + std::vector<float> imageDataFloat = + PrepareImageTensor<float>(imagePath, newWidth, newHeight, normParams, batchSize, outputLayout); + std::vector<uint8_t> imageDataQasymm8; + imageDataQasymm8.reserve(imageDataFloat.size()); + // Convert to uint8 image tensor with static cast + std::transform(imageDataFloat.begin(), imageDataFloat.end(), std::back_inserter(imageDataQasymm8), + [](float val) { return static_cast<uint8_t>(val); }); + return imageDataQasymm8; +} + +// Write image tensor to ofstream +template <typename ElemType> +void WriteImageTensorImpl(const std::vector<ElemType>& imageData, std::ofstream& imageTensorFile) +{ + std::copy(imageData.begin(), imageData.end(), std::ostream_iterator<ElemType>(imageTensorFile, " ")); +}
\ No newline at end of file |