diff options
author | Jan Eilers <jan.eilers@arm.com> | 2021-02-25 17:44:00 +0000 |
---|---|---|
committer | Jan Eilers <jan.eilers@arm.com> | 2021-02-25 18:27:49 +0000 |
commit | fd627ffaec8fd8801d980b4c91ee7c0607ab6aaf (patch) | |
tree | eb4bc8f9b411f30c7655616142b5a4bdd3a1acd0 /21.02/struct_inference_model_internal_1_1_params.xhtml | |
parent | fb14ebbd68e04876809145296af96f6f41857418 (diff) | |
download | armnn-fd627ffaec8fd8801d980b4c91ee7c0607ab6aaf.tar.gz |
IVGCVSW-5687 Update Doxygen Docu
* Update Doxygen Documentation for 21.02 release
Signed-off-by: Jan Eilers <jan.eilers@arm.com>
Change-Id: I9ed2f9caab038836ea99d7b378d7899fe431a4e5
Diffstat (limited to '21.02/struct_inference_model_internal_1_1_params.xhtml')
-rw-r--r-- | 21.02/struct_inference_model_internal_1_1_params.xhtml | 559 |
1 files changed, 559 insertions, 0 deletions
diff --git a/21.02/struct_inference_model_internal_1_1_params.xhtml b/21.02/struct_inference_model_internal_1_1_params.xhtml new file mode 100644 index 0000000000..03798bca2b --- /dev/null +++ b/21.02/struct_inference_model_internal_1_1_params.xhtml @@ -0,0 +1,559 @@ +<!-- Copyright (c) 2020 ARM Limited. --> +<!-- --> +<!-- SPDX-License-Identifier: MIT --> +<!-- --> +<!-- HTML header for doxygen 1.8.13--> +<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd"> +<html xmlns="http://www.w3.org/1999/xhtml"> +<head> +<meta http-equiv="Content-Type" content="text/xhtml;charset=UTF-8"/> +<meta http-equiv="X-UA-Compatible" content="IE=9"/> +<meta name="generator" content="Doxygen 1.8.13"/> +<meta name="robots" content="NOINDEX, NOFOLLOW" /> +<meta name="viewport" content="width=device-width, initial-scale=1"/> +<title>ArmNN: Params Struct Reference</title> +<link href="tabs.css" rel="stylesheet" type="text/css"/> +<script type="text/javascript" src="jquery.js"></script> +<script type="text/javascript" src="dynsections.js"></script> +<link href="navtree.css" rel="stylesheet" type="text/css"/> +<script type="text/javascript" src="resize.js"></script> +<script type="text/javascript" src="navtreedata.js"></script> +<script type="text/javascript" src="navtree.js"></script> +<script type="text/javascript"> + $(document).ready(initResizable); +</script> +<link href="search/search.css" rel="stylesheet" type="text/css"/> +<script type="text/javascript" src="search/searchdata.js"></script> +<script type="text/javascript" src="search/search.js"></script> +<script type="text/x-mathjax-config"> + MathJax.Hub.Config({ + extensions: ["tex2jax.js"], + jax: ["input/TeX","output/HTML-CSS"], +}); +</script><script type="text/javascript" src="http://cdn.mathjax.org/mathjax/latest/MathJax.js"></script> +<link href="doxygen.css" rel="stylesheet" type="text/css" /> +<link href="stylesheet.css" rel="stylesheet" type="text/css"/> +</head> +<body> +<div id="top"><!-- do not remove this div, it is closed by doxygen! --> +<div id="titlearea"> +<table cellspacing="0" cellpadding="0"> + <tbody> + <tr style="height: 56px;"> + <img alt="ArmNN" src="Arm_NN_horizontal_blue.png" style="max-width: 10rem; margin-top: .5rem; margin-left 10px"/> + <td style="padding-left: 0.5em;"> + <div id="projectname"> +  <span id="projectnumber">21.02</span> + </div> + </td> + </tr> + </tbody> +</table> +</div> +<!-- end header part --> +<!-- Generated by Doxygen 1.8.13 --> +<script type="text/javascript"> +var searchBox = new SearchBox("searchBox", "search",false,'Search'); +</script> +<script type="text/javascript" src="menudata.js"></script> +<script type="text/javascript" src="menu.js"></script> +<script type="text/javascript"> +$(function() { + initMenu('',true,false,'search.php','Search'); + $(document).ready(function() { init_search(); }); +}); +</script> +<div id="main-nav"></div> +</div><!-- top --> +<div id="side-nav" class="ui-resizable side-nav-resizable"> + <div id="nav-tree"> + <div id="nav-tree-contents"> + <div id="nav-sync" class="sync"></div> + </div> + </div> + <div id="splitbar" style="-moz-user-select:none;" + class="ui-resizable-handle"> + </div> +</div> +<script type="text/javascript"> +$(document).ready(function(){initNavTree('struct_inference_model_internal_1_1_params.xhtml','');}); +</script> +<div id="doc-content"> +<!-- window showing the filter options --> +<div id="MSearchSelectWindow" + onmouseover="return searchBox.OnSearchSelectShow()" + onmouseout="return searchBox.OnSearchSelectHide()" + onkeydown="return searchBox.OnSearchSelectKey(event)"> +</div> + +<!-- iframe showing the search results (closed by default) --> +<div id="MSearchResultsWindow"> +<iframe src="javascript:void(0)" frameborder="0" + name="MSearchResults" id="MSearchResults"> +</iframe> +</div> + +<div class="header"> + <div class="summary"> +<a href="#pub-methods">Public Member Functions</a> | +<a href="#pub-attribs">Public Attributes</a> | +<a href="struct_inference_model_internal_1_1_params-members.xhtml">List of all members</a> </div> + <div class="headertitle"> +<div class="title">Params Struct Reference</div> </div> +</div><!--header--> +<div class="contents"> + +<p><code>#include <<a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>></code></p> +<table class="memberdecls"> +<tr class="heading"><td colspan="2"><h2 class="groupheader"><a name="pub-methods"></a> +Public Member Functions</h2></td></tr> +<tr class="memitem:a7974597e9d3c848fd265d9445f7cf8cb"><td class="memItemLeft" align="right" valign="top"> </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a7974597e9d3c848fd265d9445f7cf8cb">Params</a> ()</td></tr> +<tr class="separator:a7974597e9d3c848fd265d9445f7cf8cb"><td class="memSeparator" colspan="2"> </td></tr> +</table><table class="memberdecls"> +<tr class="heading"><td colspan="2"><h2 class="groupheader"><a name="pub-attribs"></a> +Public Attributes</h2></td></tr> +<tr class="memitem:a4fa312cf0d60fbd3988a7c76ab8e2980"><td class="memItemLeft" align="right" valign="top">std::string </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a4fa312cf0d60fbd3988a7c76ab8e2980">m_ModelPath</a></td></tr> +<tr class="separator:a4fa312cf0d60fbd3988a7c76ab8e2980"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:aad2ac35d4cb83ee4da9fad5fbcb907e0"><td class="memItemLeft" align="right" valign="top">std::vector< std::string > </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#aad2ac35d4cb83ee4da9fad5fbcb907e0">m_InputBindings</a></td></tr> +<tr class="separator:aad2ac35d4cb83ee4da9fad5fbcb907e0"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:ad69aa6b4967ce55ee4a915c52c71bf2e"><td class="memItemLeft" align="right" valign="top">std::vector< <a class="el" href="classarmnn_1_1_tensor_shape.xhtml">armnn::TensorShape</a> > </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#ad69aa6b4967ce55ee4a915c52c71bf2e">m_InputShapes</a></td></tr> +<tr class="separator:ad69aa6b4967ce55ee4a915c52c71bf2e"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:abeacb4ed1ca9256ee0e8aea73185a0cc"><td class="memItemLeft" align="right" valign="top">std::vector< std::string > </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#abeacb4ed1ca9256ee0e8aea73185a0cc">m_OutputBindings</a></td></tr> +<tr class="separator:abeacb4ed1ca9256ee0e8aea73185a0cc"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a2d54e6252c1c9a0e29f7706ba03b2b74"><td class="memItemLeft" align="right" valign="top">std::vector< <a class="el" href="classarmnn_1_1_backend_id.xhtml">armnn::BackendId</a> > </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a2d54e6252c1c9a0e29f7706ba03b2b74">m_ComputeDevices</a></td></tr> +<tr class="separator:a2d54e6252c1c9a0e29f7706ba03b2b74"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:ae43cf4b5df0068ee6a9151c98947248b"><td class="memItemLeft" align="right" valign="top">std::string </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#ae43cf4b5df0068ee6a9151c98947248b">m_DynamicBackendsPath</a></td></tr> +<tr class="separator:ae43cf4b5df0068ee6a9151c98947248b"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a2d4582aa74998c397bd064ae73745b62"><td class="memItemLeft" align="right" valign="top">size_t </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a2d4582aa74998c397bd064ae73745b62">m_SubgraphId</a></td></tr> +<tr class="separator:a2d4582aa74998c397bd064ae73745b62"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a6bf2f586c403977d31c7d32d371918cf"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a6bf2f586c403977d31c7d32d371918cf">m_IsModelBinary</a></td></tr> +<tr class="separator:a6bf2f586c403977d31c7d32d371918cf"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:aaae50a6c0f73e4c210c2e4331c439482"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#aaae50a6c0f73e4c210c2e4331c439482">m_VisualizePostOptimizationModel</a></td></tr> +<tr class="separator:aaae50a6c0f73e4c210c2e4331c439482"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a513151623e8d448951a0b94ad1946fbe"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a513151623e8d448951a0b94ad1946fbe">m_EnableFp16TurboMode</a></td></tr> +<tr class="separator:a513151623e8d448951a0b94ad1946fbe"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a0b99050baebe1d169392461b3a9be08d"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a0b99050baebe1d169392461b3a9be08d">m_EnableBf16TurboMode</a></td></tr> +<tr class="separator:a0b99050baebe1d169392461b3a9be08d"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:acde2af8cbbd224a9f94e509ca538a775"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#acde2af8cbbd224a9f94e509ca538a775">m_PrintIntermediateLayers</a></td></tr> +<tr class="separator:acde2af8cbbd224a9f94e509ca538a775"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a5c7f0c083da98e7b6e9ba79d2fcd985d"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a5c7f0c083da98e7b6e9ba79d2fcd985d">m_ParseUnsupported</a></td></tr> +<tr class="separator:a5c7f0c083da98e7b6e9ba79d2fcd985d"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a2a245a63e87f363df491ad8c35be54c5"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a2a245a63e87f363df491ad8c35be54c5">m_InferOutputShape</a></td></tr> +<tr class="separator:a2a245a63e87f363df491ad8c35be54c5"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a66f9597b152215daba3211379dad63d3"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a66f9597b152215daba3211379dad63d3">m_EnableFastMath</a></td></tr> +<tr class="separator:a66f9597b152215daba3211379dad63d3"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a9a04b8c7f19a84f63125edec509b6d53"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a9a04b8c7f19a84f63125edec509b6d53">m_SaveCachedNetwork</a></td></tr> +<tr class="separator:a9a04b8c7f19a84f63125edec509b6d53"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a9f8881646a38f828f92d3354930c4165"><td class="memItemLeft" align="right" valign="top">std::string </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a9f8881646a38f828f92d3354930c4165">m_CachedNetworkFilePath</a></td></tr> +<tr class="separator:a9f8881646a38f828f92d3354930c4165"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:adc650e032e7fce99f15e2bf903e7167b"><td class="memItemLeft" align="right" valign="top">unsigned int </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#adc650e032e7fce99f15e2bf903e7167b">m_NumberOfThreads</a></td></tr> +<tr class="separator:adc650e032e7fce99f15e2bf903e7167b"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:ac609a217f4de4c647422dbb358a5f2ef"><td class="memItemLeft" align="right" valign="top">std::string </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#ac609a217f4de4c647422dbb358a5f2ef">m_MLGOTuningFilePath</a></td></tr> +<tr class="separator:ac609a217f4de4c647422dbb358a5f2ef"><td class="memSeparator" colspan="2"> </td></tr> +</table> +<a name="details" id="details"></a><h2 class="groupheader">Detailed Description</h2> +<div class="textblock"> +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00083">83</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> +</div><h2 class="groupheader">Constructor & Destructor Documentation</h2> +<a id="a7974597e9d3c848fd265d9445f7cf8cb"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a7974597e9d3c848fd265d9445f7cf8cb">◆ </a></span>Params()</h2> + +<div class="memitem"> +<div class="memproto"> +<table class="mlabels"> + <tr> + <td class="mlabels-left"> + <table class="memname"> + <tr> + <td class="memname"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml">Params</a> </td> + <td>(</td> + <td class="paramname"></td><td>)</td> + <td></td> + </tr> + </table> + </td> + <td class="mlabels-right"> +<span class="mlabels"><span class="mlabel">inline</span></span> </td> + </tr> +</table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00106">106</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">References <a class="el" href="_inference_model_8hpp_source.xhtml#l00101">Params::m_CachedNetworkFilePath</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00095">Params::m_EnableBf16TurboMode</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00099">Params::m_EnableFastMath</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00094">Params::m_EnableFp16TurboMode</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00098">Params::m_InferOutputShape</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00092">Params::m_IsModelBinary</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00103">Params::m_MLGOTuningFilePath</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00102">Params::m_NumberOfThreads</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00097">Params::m_ParseUnsupported</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00096">Params::m_PrintIntermediateLayers</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00100">Params::m_SaveCachedNetwork</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00091">Params::m_SubgraphId</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00093">Params::m_VisualizePostOptimizationModel</a>.</p> +<div class="fragment"><div class="line"><a name="l00107"></a><span class="lineno"> 107</span>  : <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a2d54e6252c1c9a0e29f7706ba03b2b74">m_ComputeDevices</a>{}</div><div class="line"><a name="l00108"></a><span class="lineno"> 108</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a2d4582aa74998c397bd064ae73745b62">m_SubgraphId</a>(0)</div><div class="line"><a name="l00109"></a><span class="lineno"> 109</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a6bf2f586c403977d31c7d32d371918cf">m_IsModelBinary</a>(<span class="keyword">true</span>)</div><div class="line"><a name="l00110"></a><span class="lineno"> 110</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#aaae50a6c0f73e4c210c2e4331c439482">m_VisualizePostOptimizationModel</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00111"></a><span class="lineno"> 111</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a513151623e8d448951a0b94ad1946fbe">m_EnableFp16TurboMode</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00112"></a><span class="lineno"> 112</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a0b99050baebe1d169392461b3a9be08d">m_EnableBf16TurboMode</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00113"></a><span class="lineno"> 113</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#acde2af8cbbd224a9f94e509ca538a775">m_PrintIntermediateLayers</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00114"></a><span class="lineno"> 114</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a5c7f0c083da98e7b6e9ba79d2fcd985d">m_ParseUnsupported</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00115"></a><span class="lineno"> 115</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a2a245a63e87f363df491ad8c35be54c5">m_InferOutputShape</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00116"></a><span class="lineno"> 116</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a66f9597b152215daba3211379dad63d3">m_EnableFastMath</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00117"></a><span class="lineno"> 117</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a9a04b8c7f19a84f63125edec509b6d53">m_SaveCachedNetwork</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00118"></a><span class="lineno"> 118</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a9f8881646a38f828f92d3354930c4165">m_CachedNetworkFilePath</a>(<span class="stringliteral">""</span>)</div><div class="line"><a name="l00119"></a><span class="lineno"> 119</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#adc650e032e7fce99f15e2bf903e7167b">m_NumberOfThreads</a>(0)</div><div class="line"><a name="l00120"></a><span class="lineno"> 120</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#ac609a217f4de4c647422dbb358a5f2ef">m_MLGOTuningFilePath</a>(<span class="stringliteral">""</span>)</div><div class="line"><a name="l00121"></a><span class="lineno"> 121</span>  {}</div><div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a513151623e8d448951a0b94ad1946fbe"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a513151623e8d448951a0b94ad1946fbe">InferenceModelInternal::Params::m_EnableFp16TurboMode</a></div><div class="ttdeci">bool m_EnableFp16TurboMode</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00094">InferenceModel.hpp:94</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_aaae50a6c0f73e4c210c2e4331c439482"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#aaae50a6c0f73e4c210c2e4331c439482">InferenceModelInternal::Params::m_VisualizePostOptimizationModel</a></div><div class="ttdeci">bool m_VisualizePostOptimizationModel</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00093">InferenceModel.hpp:93</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a66f9597b152215daba3211379dad63d3"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a66f9597b152215daba3211379dad63d3">InferenceModelInternal::Params::m_EnableFastMath</a></div><div class="ttdeci">bool m_EnableFastMath</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00099">InferenceModel.hpp:99</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a0b99050baebe1d169392461b3a9be08d"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a0b99050baebe1d169392461b3a9be08d">InferenceModelInternal::Params::m_EnableBf16TurboMode</a></div><div class="ttdeci">bool m_EnableBf16TurboMode</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00095">InferenceModel.hpp:95</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a9a04b8c7f19a84f63125edec509b6d53"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a9a04b8c7f19a84f63125edec509b6d53">InferenceModelInternal::Params::m_SaveCachedNetwork</a></div><div class="ttdeci">bool m_SaveCachedNetwork</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00100">InferenceModel.hpp:100</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a2d54e6252c1c9a0e29f7706ba03b2b74"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a2d54e6252c1c9a0e29f7706ba03b2b74">InferenceModelInternal::Params::m_ComputeDevices</a></div><div class="ttdeci">std::vector< armnn::BackendId > m_ComputeDevices</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00089">InferenceModel.hpp:89</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a5c7f0c083da98e7b6e9ba79d2fcd985d"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a5c7f0c083da98e7b6e9ba79d2fcd985d">InferenceModelInternal::Params::m_ParseUnsupported</a></div><div class="ttdeci">bool m_ParseUnsupported</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00097">InferenceModel.hpp:97</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a2a245a63e87f363df491ad8c35be54c5"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a2a245a63e87f363df491ad8c35be54c5">InferenceModelInternal::Params::m_InferOutputShape</a></div><div class="ttdeci">bool m_InferOutputShape</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00098">InferenceModel.hpp:98</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_acde2af8cbbd224a9f94e509ca538a775"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#acde2af8cbbd224a9f94e509ca538a775">InferenceModelInternal::Params::m_PrintIntermediateLayers</a></div><div class="ttdeci">bool m_PrintIntermediateLayers</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00096">InferenceModel.hpp:96</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a9f8881646a38f828f92d3354930c4165"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a9f8881646a38f828f92d3354930c4165">InferenceModelInternal::Params::m_CachedNetworkFilePath</a></div><div class="ttdeci">std::string m_CachedNetworkFilePath</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00101">InferenceModel.hpp:101</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_ac609a217f4de4c647422dbb358a5f2ef"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#ac609a217f4de4c647422dbb358a5f2ef">InferenceModelInternal::Params::m_MLGOTuningFilePath</a></div><div class="ttdeci">std::string m_MLGOTuningFilePath</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00103">InferenceModel.hpp:103</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_adc650e032e7fce99f15e2bf903e7167b"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#adc650e032e7fce99f15e2bf903e7167b">InferenceModelInternal::Params::m_NumberOfThreads</a></div><div class="ttdeci">unsigned int m_NumberOfThreads</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00102">InferenceModel.hpp:102</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a6bf2f586c403977d31c7d32d371918cf"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a6bf2f586c403977d31c7d32d371918cf">InferenceModelInternal::Params::m_IsModelBinary</a></div><div class="ttdeci">bool m_IsModelBinary</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00092">InferenceModel.hpp:92</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a2d4582aa74998c397bd064ae73745b62"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a2d4582aa74998c397bd064ae73745b62">InferenceModelInternal::Params::m_SubgraphId</a></div><div class="ttdeci">size_t m_SubgraphId</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00091">InferenceModel.hpp:91</a></div></div> +</div><!-- fragment --> +</div> +</div> +<h2 class="groupheader">Member Data Documentation</h2> +<a id="a9f8881646a38f828f92d3354930c4165"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a9f8881646a38f828f92d3354930c4165">◆ </a></span>m_CachedNetworkFilePath</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">std::string m_CachedNetworkFilePath</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00101">101</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00396">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00106">Params::Params()</a>.</p> + +</div> +</div> +<a id="a2d54e6252c1c9a0e29f7706ba03b2b74"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a2d54e6252c1c9a0e29f7706ba03b2b74">◆ </a></span>m_ComputeDevices</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">std::vector<<a class="el" href="classarmnn_1_1_backend_id.xhtml">armnn::BackendId</a>> m_ComputeDevices</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00089">89</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00396">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_model_accuracy_tool-_armnn_8cpp_source.xhtml#l00047">main()</a>, and <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>.</p> + +</div> +</div> +<a id="ae43cf4b5df0068ee6a9151c98947248b"></a> +<h2 class="memtitle"><span class="permalink"><a href="#ae43cf4b5df0068ee6a9151c98947248b">◆ </a></span>m_DynamicBackendsPath</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">std::string m_DynamicBackendsPath</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00090">90</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00606">InferenceModel< IParser, TDataType >::GetAllQuantizationParams()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00396">InferenceModel< IParser, TDataType >::InferenceModel()</a>, and <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>.</p> + +</div> +</div> +<a id="a0b99050baebe1d169392461b3a9be08d"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a0b99050baebe1d169392461b3a9be08d">◆ </a></span>m_EnableBf16TurboMode</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_EnableBf16TurboMode</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00095">95</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00396">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00106">Params::Params()</a>.</p> + +</div> +</div> +<a id="a66f9597b152215daba3211379dad63d3"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a66f9597b152215daba3211379dad63d3">◆ </a></span>m_EnableFastMath</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_EnableFastMath</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00099">99</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00396">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00106">Params::Params()</a>.</p> + +</div> +</div> +<a id="a513151623e8d448951a0b94ad1946fbe"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a513151623e8d448951a0b94ad1946fbe">◆ </a></span>m_EnableFp16TurboMode</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_EnableFp16TurboMode</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00094">94</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00396">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00106">Params::Params()</a>.</p> + +</div> +</div> +<a id="a2a245a63e87f363df491ad8c35be54c5"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a2a245a63e87f363df491ad8c35be54c5">◆ </a></span>m_InferOutputShape</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_InferOutputShape</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00098">98</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00132">CreateNetworkImpl< IParser >::Create()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00106">Params::Params()</a>.</p> + +</div> +</div> +<a id="aad2ac35d4cb83ee4da9fad5fbcb907e0"></a> +<h2 class="memtitle"><span class="permalink"><a href="#aad2ac35d4cb83ee4da9fad5fbcb907e0">◆ </a></span>m_InputBindings</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">std::vector<std::string> m_InputBindings</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00086">86</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00489">InferenceModel< IParser, TDataType >::CheckInputIndexIsValid()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00132">CreateNetworkImpl< IParser >::Create()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00606">InferenceModel< IParser, TDataType >::GetAllQuantizationParams()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00570">InferenceModel< IParser, TDataType >::GetInputBindingInfo()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00576">InferenceModel< IParser, TDataType >::GetInputBindingInfos()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00599">InferenceModel< IParser, TDataType >::GetInputQuantizationParams()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00505">InferenceModel< IParser, TDataType >::GetInputSize()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00396">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_model_accuracy_tool-_armnn_8cpp_source.xhtml#l00047">main()</a>, and <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>.</p> + +</div> +</div> +<a id="ad69aa6b4967ce55ee4a915c52c71bf2e"></a> +<h2 class="memtitle"><span class="permalink"><a href="#ad69aa6b4967ce55ee4a915c52c71bf2e">◆ </a></span>m_InputShapes</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">std::vector<<a class="el" href="classarmnn_1_1_tensor_shape.xhtml">armnn::TensorShape</a>> m_InputShapes</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00087">87</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00132">CreateNetworkImpl< IParser >::Create()</a>, and <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>.</p> + +</div> +</div> +<a id="a6bf2f586c403977d31c7d32d371918cf"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a6bf2f586c403977d31c7d32d371918cf">◆ </a></span>m_IsModelBinary</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_IsModelBinary</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00092">92</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00132">CreateNetworkImpl< IParser >::Create()</a>, <a class="el" href="_model_accuracy_tool-_armnn_8cpp_source.xhtml#l00047">main()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00106">Params::Params()</a>.</p> + +</div> +</div> +<a id="ac609a217f4de4c647422dbb358a5f2ef"></a> +<h2 class="memtitle"><span class="permalink"><a href="#ac609a217f4de4c647422dbb358a5f2ef">◆ </a></span>m_MLGOTuningFilePath</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">std::string m_MLGOTuningFilePath</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00103">103</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00396">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00106">Params::Params()</a>.</p> + +</div> +</div> +<a id="a4fa312cf0d60fbd3988a7c76ab8e2980"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a4fa312cf0d60fbd3988a7c76ab8e2980">◆ </a></span>m_ModelPath</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">std::string m_ModelPath</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00085">85</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_test_8inl_source.xhtml#l00353">armnn::test::ClassifierInferenceTestMain()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00132">CreateNetworkImpl< IParser >::Create()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00396">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_model_accuracy_tool-_armnn_8cpp_source.xhtml#l00047">main()</a>, and <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>.</p> + +</div> +</div> +<a id="adc650e032e7fce99f15e2bf903e7167b"></a> +<h2 class="memtitle"><span class="permalink"><a href="#adc650e032e7fce99f15e2bf903e7167b">◆ </a></span>m_NumberOfThreads</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">unsigned int m_NumberOfThreads</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00102">102</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00396">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00106">Params::Params()</a>.</p> + +</div> +</div> +<a id="abeacb4ed1ca9256ee0e8aea73185a0cc"></a> +<h2 class="memtitle"><span class="permalink"><a href="#abeacb4ed1ca9256ee0e8aea73185a0cc">◆ </a></span>m_OutputBindings</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">std::vector<std::string> m_OutputBindings</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00088">88</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00497">InferenceModel< IParser, TDataType >::CheckOutputIndexIsValid()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00132">CreateNetworkImpl< IParser >::Create()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00606">InferenceModel< IParser, TDataType >::GetAllQuantizationParams()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00581">InferenceModel< IParser, TDataType >::GetOutputBindingInfo()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00587">InferenceModel< IParser, TDataType >::GetOutputBindingInfos()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00511">InferenceModel< IParser, TDataType >::GetOutputSize()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00592">InferenceModel< IParser, TDataType >::GetQuantizationParams()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00396">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_model_accuracy_tool-_armnn_8cpp_source.xhtml#l00047">main()</a>, and <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>.</p> + +</div> +</div> +<a id="a5c7f0c083da98e7b6e9ba79d2fcd985d"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a5c7f0c083da98e7b6e9ba79d2fcd985d">◆ </a></span>m_ParseUnsupported</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_ParseUnsupported</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00097">97</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00132">CreateNetworkImpl< IParser >::Create()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00106">Params::Params()</a>.</p> + +</div> +</div> +<a id="acde2af8cbbd224a9f94e509ca538a775"></a> +<h2 class="memtitle"><span class="permalink"><a href="#acde2af8cbbd224a9f94e509ca538a775">◆ </a></span>m_PrintIntermediateLayers</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_PrintIntermediateLayers</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00096">96</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00396">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00106">Params::Params()</a>.</p> + +</div> +</div> +<a id="a9a04b8c7f19a84f63125edec509b6d53"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a9a04b8c7f19a84f63125edec509b6d53">◆ </a></span>m_SaveCachedNetwork</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_SaveCachedNetwork</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00100">100</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00396">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00106">Params::Params()</a>.</p> + +</div> +</div> +<a id="a2d4582aa74998c397bd064ae73745b62"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a2d4582aa74998c397bd064ae73745b62">◆ </a></span>m_SubgraphId</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">size_t m_SubgraphId</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00091">91</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00132">CreateNetworkImpl< IParser >::Create()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00106">Params::Params()</a>.</p> + +</div> +</div> +<a id="aaae50a6c0f73e4c210c2e4331c439482"></a> +<h2 class="memtitle"><span class="permalink"><a href="#aaae50a6c0f73e4c210c2e4331c439482">◆ </a></span>m_VisualizePostOptimizationModel</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_VisualizePostOptimizationModel</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00093">93</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00396">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00289">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00106">Params::Params()</a>.</p> + +</div> +</div> +<hr/>The documentation for this struct was generated from the following file:<ul> +<li>tests/<a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a></li> +</ul> +</div><!-- contents --> +</div><!-- doc-content --> +<!-- start footer part --> +<div id="nav-path" class="navpath"><!-- id is needed for treeview function! --> + <ul> + <li class="navelem"><a class="el" href="namespace_inference_model_internal.xhtml">InferenceModelInternal</a></li><li class="navelem"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml">Params</a></li> + <li class="footer">Generated on Thu Feb 25 2021 17:28:03 for ArmNN by + <a href="http://www.doxygen.org/index.html"> + <img class="footer" src="doxygen.png" alt="doxygen"/></a> 1.8.13 </li> + </ul> +</div> +</body> +</html> |