ArmNN
NotReleased
|
#include <InferenceModel.hpp>
Public Member Functions | |
Params () | |
Public Attributes | |
std::string | m_ModelPath |
std::vector< std::string > | m_InputBindings |
std::vector< armnn::TensorShape > | m_InputShapes |
std::vector< std::string > | m_OutputBindings |
std::vector< armnn::BackendId > | m_ComputeDevices |
std::string | m_DynamicBackendsPath |
size_t | m_SubgraphId |
bool | m_IsModelBinary |
bool | m_VisualizePostOptimizationModel |
bool | m_EnableFp16TurboMode |
bool | m_PrintIntermediateLayers |
bool | m_ParseUnsupported |
Definition at line 82 of file InferenceModel.hpp.
|
inline |
Definition at line 97 of file InferenceModel.hpp.
References Params::m_EnableFp16TurboMode, Params::m_IsModelBinary, Params::m_ParseUnsupported, Params::m_PrintIntermediateLayers, Params::m_SubgraphId, and Params::m_VisualizePostOptimizationModel.
std::vector<armnn::BackendId> m_ComputeDevices |
Definition at line 88 of file InferenceModel.hpp.
Referenced by InferenceModel< IParser, TDataType >::InferenceModel(), main(), and MainImpl().
std::string m_DynamicBackendsPath |
Definition at line 89 of file InferenceModel.hpp.
Referenced by InferenceModel< IParser, TDataType >::GetAllQuantizationParams(), InferenceModel< IParser, TDataType >::InferenceModel(), and MainImpl().
bool m_EnableFp16TurboMode |
Definition at line 93 of file InferenceModel.hpp.
Referenced by InferenceModel< IParser, TDataType >::InferenceModel(), MainImpl(), and Params::Params().
std::vector<std::string> m_InputBindings |
Definition at line 85 of file InferenceModel.hpp.
Referenced by InferenceModel< IParser, TDataType >::CheckInputIndexIsValid(), CreateNetworkImpl< IParser >::Create(), InferenceModel< IParser, TDataType >::GetAllQuantizationParams(), InferenceModel< IParser, TDataType >::GetInputBindingInfo(), InferenceModel< IParser, TDataType >::GetInputBindingInfos(), InferenceModel< IParser, TDataType >::GetInputQuantizationParams(), InferenceModel< IParser, TDataType >::GetInputSize(), InferenceModel< IParser, TDataType >::InferenceModel(), main(), and MainImpl().
std::vector<armnn::TensorShape> m_InputShapes |
Definition at line 86 of file InferenceModel.hpp.
Referenced by CreateNetworkImpl< IParser >::Create(), and MainImpl().
bool m_IsModelBinary |
Definition at line 91 of file InferenceModel.hpp.
Referenced by CreateNetworkImpl< IParser >::Create(), main(), MainImpl(), and Params::Params().
std::string m_ModelPath |
Definition at line 84 of file InferenceModel.hpp.
Referenced by armnn::test::ClassifierInferenceTestMain(), CreateNetworkImpl< IParser >::Create(), InferenceModel< IParser, TDataType >::InferenceModel(), main(), and MainImpl().
std::vector<std::string> m_OutputBindings |
Definition at line 87 of file InferenceModel.hpp.
Referenced by InferenceModel< IParser, TDataType >::CheckOutputIndexIsValid(), CreateNetworkImpl< IParser >::Create(), InferenceModel< IParser, TDataType >::GetAllQuantizationParams(), InferenceModel< IParser, TDataType >::GetOutputBindingInfo(), InferenceModel< IParser, TDataType >::GetOutputBindingInfos(), InferenceModel< IParser, TDataType >::GetOutputSize(), InferenceModel< IParser, TDataType >::GetQuantizationParams(), InferenceModel< IParser, TDataType >::InferenceModel(), main(), and MainImpl().
bool m_ParseUnsupported |
Definition at line 95 of file InferenceModel.hpp.
Referenced by CreateNetworkImpl< IParser >::Create(), MainImpl(), and Params::Params().
bool m_PrintIntermediateLayers |
Definition at line 94 of file InferenceModel.hpp.
Referenced by InferenceModel< IParser, TDataType >::InferenceModel(), MainImpl(), and Params::Params().
size_t m_SubgraphId |
Definition at line 90 of file InferenceModel.hpp.
Referenced by CreateNetworkImpl< IParser >::Create(), MainImpl(), and Params::Params().
bool m_VisualizePostOptimizationModel |
Definition at line 92 of file InferenceModel.hpp.
Referenced by InferenceModel< IParser, TDataType >::InferenceModel(), MainImpl(), and Params::Params().