aboutsummaryrefslogtreecommitdiff
path: root/tests/CaffeYolo-Armnn/CaffeYolo-Armnn.cpp
blob: d563faaab20e3d0d1c5092bc7eb07aded3c457d6 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
//
// Copyright © 2017 Arm Ltd. All rights reserved.
// SPDX-License-Identifier: MIT
//
#include "../YoloInferenceTest.hpp"
#include "armnnCaffeParser/ICaffeParser.hpp"
#include "armnn/TypesUtils.hpp"

int main(int argc, char* argv[])
{
    armnn::TensorShape inputTensorShape{ { 1, 3, YoloImageHeight, YoloImageWidth } };

    using YoloInferenceModel = InferenceModel<armnnCaffeParser::ICaffeParser,
        float>;

    int retVal = EXIT_FAILURE;
    try
    {
        // Coverity fix: InferenceTestMain() may throw uncaught exceptions.
        retVal = InferenceTestMain(argc, argv, { 0 },
            [&inputTensorShape]()
            {
                return make_unique<YoloTestCaseProvider<YoloInferenceModel>>(
                    [&]
                    (const InferenceTestOptions &commonOptions,
                     typename YoloInferenceModel::CommandLineOptions modelOptions)
                    {
                        if (!ValidateDirectory(modelOptions.m_ModelDir))
                        {
                            return std::unique_ptr<YoloInferenceModel>();
                        }

                        typename YoloInferenceModel::Params modelParams;
                        modelParams.m_ModelPath = modelOptions.m_ModelDir + "yolov1_tiny_voc2007_model.caffemodel";
                        modelParams.m_InputBindings = { "data" };
                        modelParams.m_OutputBindings = { "fc12" };
                        modelParams.m_InputShapes = { inputTensorShape };
                        modelParams.m_IsModelBinary = true;
                        modelParams.m_ComputeDevices = modelOptions.GetComputeDevicesAsBackendIds();
                        modelParams.m_VisualizePostOptimizationModel = modelOptions.m_VisualizePostOptimizationModel;
                        modelParams.m_EnableFp16TurboMode = modelOptions.m_EnableFp16TurboMode;

                        return std::make_unique<YoloInferenceModel>(modelParams,
                                                                    commonOptions.m_EnableProfiling,
                                                                    commonOptions.m_DynamicBackendsPath);
                });
            });
    }
    catch (const std::exception& e)
    {
        // Coverity fix: BOOST_LOG_TRIVIAL (typically used to report errors) may throw an
        // exception of type std::length_error.
        // Using stderr instead in this context as there is no point in nesting try-catch blocks here.
        std::cerr << "WARNING: CaffeYolo-Armnn: An error has occurred when running "
                     "the classifier inference tests: " << e.what() << std::endl;
    }
    return retVal;
}