5 #include "../YoloInferenceTest.hpp" 9 int main(
int argc,
char* argv[])
16 int retVal = EXIT_FAILURE;
23 return make_unique<YoloTestCaseProvider<YoloInferenceModel>>(
26 typename YoloInferenceModel::CommandLineOptions modelOptions)
30 return std::unique_ptr<YoloInferenceModel>();
33 typename YoloInferenceModel::Params modelParams;
34 modelParams.m_ModelPath = modelOptions.m_ModelDir +
"yolov1_tiny_voc2007_model.caffemodel";
35 modelParams.m_InputBindings = {
"data" };
36 modelParams.m_OutputBindings = {
"fc12" };
37 modelParams.m_InputShapes = { inputTensorShape };
38 modelParams.m_IsModelBinary =
true;
39 modelParams.m_ComputeDevices = modelOptions.GetComputeDevicesAsBackendIds();
40 modelParams.m_VisualizePostOptimizationModel = modelOptions.m_VisualizePostOptimizationModel;
41 modelParams.m_EnableFp16TurboMode = modelOptions.m_EnableFp16TurboMode;
43 return std::make_unique<YoloInferenceModel>(modelParams,
44 commonOptions.m_EnableProfiling,
45 commonOptions.m_DynamicBackendsPath);
49 catch (
const std::exception& e)
54 std::cerr <<
"WARNING: CaffeYolo-Armnn: An error has occurred when running " 55 "the classifier inference tests: " << e.what() << std::endl;
constexpr unsigned int YoloImageHeight
constexpr unsigned int YoloImageWidth
bool ValidateDirectory(std::string &dir)
int main(int argc, char *argv[])
int InferenceTestMain(int argc, char *argv[], const std::vector< unsigned int > &defaultTestCaseIds, TConstructTestCaseProvider constructTestCaseProvider)