diff options
author | Nikhil Raj <nikhil.raj@arm.com> | 2022-06-17 13:24:58 +0100 |
---|---|---|
committer | Nikhil Raj <nikhil.raj@arm.com> | 2022-06-17 13:24:58 +0100 |
commit | d5d43d82c0137e08553e44345c609cdd1a7931c7 (patch) | |
tree | f1509f7fa94db0373a2c127682dd3d0ccc1915bd /22.05.01/struct_inference_model_internal_1_1_params.xhtml | |
parent | 549b9600a6eaf0727fa084465a75f173edf8f381 (diff) | |
download | armnn-d5d43d82c0137e08553e44345c609cdd1a7931c7.tar.gz |
Update Doxygen for 22.05 patch release
* Pooling3D added to tfLite delegate
* Available in tag 22.05.01
Signed-off-by: Nikhil Raj <nikhil.raj@arm.com>
Change-Id: I8d605bba4e87d30baa2c6d7b338c78a4400dc021
Diffstat (limited to '22.05.01/struct_inference_model_internal_1_1_params.xhtml')
-rw-r--r-- | 22.05.01/struct_inference_model_internal_1_1_params.xhtml | 685 |
1 files changed, 685 insertions, 0 deletions
diff --git a/22.05.01/struct_inference_model_internal_1_1_params.xhtml b/22.05.01/struct_inference_model_internal_1_1_params.xhtml new file mode 100644 index 0000000000..be343950ae --- /dev/null +++ b/22.05.01/struct_inference_model_internal_1_1_params.xhtml @@ -0,0 +1,685 @@ +<!-- Copyright (c) 2020 ARM Limited. --> +<!-- --> +<!-- SPDX-License-Identifier: MIT --> +<!-- --> +<!-- HTML header for doxygen 1.8.13--> +<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd"> +<html xmlns="http://www.w3.org/1999/xhtml"> +<head> +<meta http-equiv="Content-Type" content="text/xhtml;charset=UTF-8"/> +<meta http-equiv="X-UA-Compatible" content="IE=9"/> +<meta name="generator" content="Doxygen 1.8.13"/> +<meta name="robots" content="NOINDEX, NOFOLLOW" /> +<meta name="viewport" content="width=device-width, initial-scale=1"/> +<title>ArmNN: Params Struct Reference</title> +<link href="tabs.css" rel="stylesheet" type="text/css"/> +<script type="text/javascript" src="jquery.js"></script> +<script type="text/javascript" src="dynsections.js"></script> +<link href="navtree.css" rel="stylesheet" type="text/css"/> +<script type="text/javascript" src="resize.js"></script> +<script type="text/javascript" src="navtreedata.js"></script> +<script type="text/javascript" src="navtree.js"></script> +<script type="text/javascript"> + $(document).ready(initResizable); +</script> +<link href="search/search.css" rel="stylesheet" type="text/css"/> +<script type="text/javascript" src="search/searchdata.js"></script> +<script type="text/javascript" src="search/search.js"></script> +<script type="text/x-mathjax-config"> + MathJax.Hub.Config({ + extensions: ["tex2jax.js"], + jax: ["input/TeX","output/HTML-CSS"], +}); +</script><script type="text/javascript" src="http://cdn.mathjax.org/mathjax/latest/MathJax.js"></script> +<link href="doxygen.css" rel="stylesheet" type="text/css" /> +<link href="stylesheet.css" rel="stylesheet" type="text/css"/> +</head> +<body> +<div id="top"><!-- do not remove this div, it is closed by doxygen! --> +<div id="titlearea"> +<table cellspacing="0" cellpadding="0"> + <tbody> + <tr style="height: 56px;"> + <img alt="ArmNN" src="Arm_NN_horizontal_blue.png" style="max-width: 10rem; margin-top: .5rem; margin-left 10px"/> + <td style="padding-left: 0.5em;"> + <div id="projectname"> +  <span id="projectnumber">22.05.01</span> + </div> + </td> + </tr> + </tbody> +</table> +</div> +<!-- end header part --> +<!-- Generated by Doxygen 1.8.13 --> +<script type="text/javascript"> +var searchBox = new SearchBox("searchBox", "search",false,'Search'); +</script> +<script type="text/javascript" src="menudata.js"></script> +<script type="text/javascript" src="menu.js"></script> +<script type="text/javascript"> +$(function() { + initMenu('',true,false,'search.php','Search'); + $(document).ready(function() { init_search(); }); +}); +</script> +<div id="main-nav"></div> +</div><!-- top --> +<div id="side-nav" class="ui-resizable side-nav-resizable"> + <div id="nav-tree"> + <div id="nav-tree-contents"> + <div id="nav-sync" class="sync"></div> + </div> + </div> + <div id="splitbar" style="-moz-user-select:none;" + class="ui-resizable-handle"> + </div> +</div> +<script type="text/javascript"> +$(document).ready(function(){initNavTree('struct_inference_model_internal_1_1_params.xhtml','');}); +</script> +<div id="doc-content"> +<!-- window showing the filter options --> +<div id="MSearchSelectWindow" + onmouseover="return searchBox.OnSearchSelectShow()" + onmouseout="return searchBox.OnSearchSelectHide()" + onkeydown="return searchBox.OnSearchSelectKey(event)"> +</div> + +<!-- iframe showing the search results (closed by default) --> +<div id="MSearchResultsWindow"> +<iframe src="javascript:void(0)" frameborder="0" + name="MSearchResults" id="MSearchResults"> +</iframe> +</div> + +<div class="header"> + <div class="summary"> +<a href="#pub-methods">Public Member Functions</a> | +<a href="#pub-attribs">Public Attributes</a> | +<a href="struct_inference_model_internal_1_1_params-members.xhtml">List of all members</a> </div> + <div class="headertitle"> +<div class="title">Params Struct Reference</div> </div> +</div><!--header--> +<div class="contents"> + +<p><code>#include <<a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>></code></p> +<table class="memberdecls"> +<tr class="heading"><td colspan="2"><h2 class="groupheader"><a name="pub-methods"></a> +Public Member Functions</h2></td></tr> +<tr class="memitem:a7974597e9d3c848fd265d9445f7cf8cb"><td class="memItemLeft" align="right" valign="top"> </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a7974597e9d3c848fd265d9445f7cf8cb">Params</a> ()</td></tr> +<tr class="separator:a7974597e9d3c848fd265d9445f7cf8cb"><td class="memSeparator" colspan="2"> </td></tr> +</table><table class="memberdecls"> +<tr class="heading"><td colspan="2"><h2 class="groupheader"><a name="pub-attribs"></a> +Public Attributes</h2></td></tr> +<tr class="memitem:a4fa312cf0d60fbd3988a7c76ab8e2980"><td class="memItemLeft" align="right" valign="top">std::string </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a4fa312cf0d60fbd3988a7c76ab8e2980">m_ModelPath</a></td></tr> +<tr class="separator:a4fa312cf0d60fbd3988a7c76ab8e2980"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:aad2ac35d4cb83ee4da9fad5fbcb907e0"><td class="memItemLeft" align="right" valign="top">std::vector< std::string > </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#aad2ac35d4cb83ee4da9fad5fbcb907e0">m_InputBindings</a></td></tr> +<tr class="separator:aad2ac35d4cb83ee4da9fad5fbcb907e0"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:ad69aa6b4967ce55ee4a915c52c71bf2e"><td class="memItemLeft" align="right" valign="top">std::vector< <a class="el" href="classarmnn_1_1_tensor_shape.xhtml">armnn::TensorShape</a> > </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#ad69aa6b4967ce55ee4a915c52c71bf2e">m_InputShapes</a></td></tr> +<tr class="separator:ad69aa6b4967ce55ee4a915c52c71bf2e"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:abeacb4ed1ca9256ee0e8aea73185a0cc"><td class="memItemLeft" align="right" valign="top">std::vector< std::string > </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#abeacb4ed1ca9256ee0e8aea73185a0cc">m_OutputBindings</a></td></tr> +<tr class="separator:abeacb4ed1ca9256ee0e8aea73185a0cc"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a2d54e6252c1c9a0e29f7706ba03b2b74"><td class="memItemLeft" align="right" valign="top">std::vector< <a class="el" href="classarmnn_1_1_backend_id.xhtml">armnn::BackendId</a> > </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a2d54e6252c1c9a0e29f7706ba03b2b74">m_ComputeDevices</a></td></tr> +<tr class="separator:a2d54e6252c1c9a0e29f7706ba03b2b74"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:ae43cf4b5df0068ee6a9151c98947248b"><td class="memItemLeft" align="right" valign="top">std::string </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#ae43cf4b5df0068ee6a9151c98947248b">m_DynamicBackendsPath</a></td></tr> +<tr class="separator:ae43cf4b5df0068ee6a9151c98947248b"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a2d4582aa74998c397bd064ae73745b62"><td class="memItemLeft" align="right" valign="top">size_t </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a2d4582aa74998c397bd064ae73745b62">m_SubgraphId</a></td></tr> +<tr class="separator:a2d4582aa74998c397bd064ae73745b62"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a9ef8ddbeee4b869e4c68eb2ed278b8d9"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a9ef8ddbeee4b869e4c68eb2ed278b8d9">m_AllowExpandedDims</a></td></tr> +<tr class="separator:a9ef8ddbeee4b869e4c68eb2ed278b8d9"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a6bf2f586c403977d31c7d32d371918cf"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a6bf2f586c403977d31c7d32d371918cf">m_IsModelBinary</a></td></tr> +<tr class="separator:a6bf2f586c403977d31c7d32d371918cf"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:aaae50a6c0f73e4c210c2e4331c439482"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#aaae50a6c0f73e4c210c2e4331c439482">m_VisualizePostOptimizationModel</a></td></tr> +<tr class="separator:aaae50a6c0f73e4c210c2e4331c439482"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a513151623e8d448951a0b94ad1946fbe"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a513151623e8d448951a0b94ad1946fbe">m_EnableFp16TurboMode</a></td></tr> +<tr class="separator:a513151623e8d448951a0b94ad1946fbe"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a0b99050baebe1d169392461b3a9be08d"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a0b99050baebe1d169392461b3a9be08d">m_EnableBf16TurboMode</a></td></tr> +<tr class="separator:a0b99050baebe1d169392461b3a9be08d"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:acde2af8cbbd224a9f94e509ca538a775"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#acde2af8cbbd224a9f94e509ca538a775">m_PrintIntermediateLayers</a></td></tr> +<tr class="separator:acde2af8cbbd224a9f94e509ca538a775"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a5c7f0c083da98e7b6e9ba79d2fcd985d"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a5c7f0c083da98e7b6e9ba79d2fcd985d">m_ParseUnsupported</a></td></tr> +<tr class="separator:a5c7f0c083da98e7b6e9ba79d2fcd985d"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a2a245a63e87f363df491ad8c35be54c5"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a2a245a63e87f363df491ad8c35be54c5">m_InferOutputShape</a></td></tr> +<tr class="separator:a2a245a63e87f363df491ad8c35be54c5"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a66f9597b152215daba3211379dad63d3"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a66f9597b152215daba3211379dad63d3">m_EnableFastMath</a></td></tr> +<tr class="separator:a66f9597b152215daba3211379dad63d3"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a9a04b8c7f19a84f63125edec509b6d53"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a9a04b8c7f19a84f63125edec509b6d53">m_SaveCachedNetwork</a></td></tr> +<tr class="separator:a9a04b8c7f19a84f63125edec509b6d53"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:ae96fc745917a3a0c0de7a818c9a05012"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#ae96fc745917a3a0c0de7a818c9a05012">m_OutputDetailsToStdOut</a></td></tr> +<tr class="separator:ae96fc745917a3a0c0de7a818c9a05012"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:ab17deb382179697b4702cc4f909e71f8"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#ab17deb382179697b4702cc4f909e71f8">m_OutputDetailsOnlyToStdOut</a></td></tr> +<tr class="separator:ab17deb382179697b4702cc4f909e71f8"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a9f8881646a38f828f92d3354930c4165"><td class="memItemLeft" align="right" valign="top">std::string </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a9f8881646a38f828f92d3354930c4165">m_CachedNetworkFilePath</a></td></tr> +<tr class="separator:a9f8881646a38f828f92d3354930c4165"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:adc650e032e7fce99f15e2bf903e7167b"><td class="memItemLeft" align="right" valign="top">unsigned int </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#adc650e032e7fce99f15e2bf903e7167b">m_NumberOfThreads</a></td></tr> +<tr class="separator:adc650e032e7fce99f15e2bf903e7167b"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:ac609a217f4de4c647422dbb358a5f2ef"><td class="memItemLeft" align="right" valign="top">std::string </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#ac609a217f4de4c647422dbb358a5f2ef">m_MLGOTuningFilePath</a></td></tr> +<tr class="separator:ac609a217f4de4c647422dbb358a5f2ef"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:af5e88d9bb91dff85ad5a6f7e190aff4e"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#af5e88d9bb91dff85ad5a6f7e190aff4e">m_AsyncEnabled</a></td></tr> +<tr class="separator:af5e88d9bb91dff85ad5a6f7e190aff4e"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a7adc5dcfe3d76ac489f253c4d5f439c8"><td class="memItemLeft" align="right" valign="top">size_t </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a7adc5dcfe3d76ac489f253c4d5f439c8">m_ThreadPoolSize</a></td></tr> +<tr class="separator:a7adc5dcfe3d76ac489f253c4d5f439c8"><td class="memSeparator" colspan="2"> </td></tr> +<tr class="memitem:a1697b9751b4ba381f89b8c81cd9dfc3c"><td class="memItemLeft" align="right" valign="top">bool </td><td class="memItemRight" valign="bottom"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml#a1697b9751b4ba381f89b8c81cd9dfc3c">m_ImportInputsIfAligned</a></td></tr> +<tr class="separator:a1697b9751b4ba381f89b8c81cd9dfc3c"><td class="memSeparator" colspan="2"> </td></tr> +</table> +<a name="details" id="details"></a><h2 class="groupheader">Detailed Description</h2> +<div class="textblock"> +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00089">89</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> +</div><h2 class="groupheader">Constructor & Destructor Documentation</h2> +<a id="a7974597e9d3c848fd265d9445f7cf8cb"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a7974597e9d3c848fd265d9445f7cf8cb">◆ </a></span>Params()</h2> + +<div class="memitem"> +<div class="memproto"> +<table class="mlabels"> + <tr> + <td class="mlabels-left"> + <table class="memname"> + <tr> + <td class="memname"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml">Params</a> </td> + <td>(</td> + <td class="paramname"></td><td>)</td> + <td></td> + </tr> + </table> + </td> + <td class="mlabels-right"> +<span class="mlabels"><span class="mlabel">inline</span></span> </td> + </tr> +</table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">118</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">References <a class="el" href="_inference_model_8hpp_source.xhtml#l00098">Params::m_AllowExpandedDims</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00113">Params::m_AsyncEnabled</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00110">Params::m_CachedNetworkFilePath</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00102">Params::m_EnableBf16TurboMode</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00106">Params::m_EnableFastMath</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00101">Params::m_EnableFp16TurboMode</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00115">Params::m_ImportInputsIfAligned</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00105">Params::m_InferOutputShape</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00099">Params::m_IsModelBinary</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00112">Params::m_MLGOTuningFilePath</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00111">Params::m_NumberOfThreads</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00109">Params::m_OutputDetailsOnlyToStdOut</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00108">Params::m_OutputDetailsToStdOut</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00104">Params::m_ParseUnsupported</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00103">Params::m_PrintIntermediateLayers</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00107">Params::m_SaveCachedNetwork</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00097">Params::m_SubgraphId</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00114">Params::m_ThreadPoolSize</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00100">Params::m_VisualizePostOptimizationModel</a>.</p> +<div class="fragment"><div class="line"><a name="l00119"></a><span class="lineno"> 119</span>  : <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a2d54e6252c1c9a0e29f7706ba03b2b74">m_ComputeDevices</a>{}</div><div class="line"><a name="l00120"></a><span class="lineno"> 120</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a2d4582aa74998c397bd064ae73745b62">m_SubgraphId</a>(0)</div><div class="line"><a name="l00121"></a><span class="lineno"> 121</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a9ef8ddbeee4b869e4c68eb2ed278b8d9">m_AllowExpandedDims</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00122"></a><span class="lineno"> 122</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a6bf2f586c403977d31c7d32d371918cf">m_IsModelBinary</a>(<span class="keyword">true</span>)</div><div class="line"><a name="l00123"></a><span class="lineno"> 123</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#aaae50a6c0f73e4c210c2e4331c439482">m_VisualizePostOptimizationModel</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00124"></a><span class="lineno"> 124</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a513151623e8d448951a0b94ad1946fbe">m_EnableFp16TurboMode</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00125"></a><span class="lineno"> 125</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a0b99050baebe1d169392461b3a9be08d">m_EnableBf16TurboMode</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00126"></a><span class="lineno"> 126</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#acde2af8cbbd224a9f94e509ca538a775">m_PrintIntermediateLayers</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00127"></a><span class="lineno"> 127</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a5c7f0c083da98e7b6e9ba79d2fcd985d">m_ParseUnsupported</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00128"></a><span class="lineno"> 128</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a2a245a63e87f363df491ad8c35be54c5">m_InferOutputShape</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00129"></a><span class="lineno"> 129</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a66f9597b152215daba3211379dad63d3">m_EnableFastMath</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00130"></a><span class="lineno"> 130</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a9a04b8c7f19a84f63125edec509b6d53">m_SaveCachedNetwork</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00131"></a><span class="lineno"> 131</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#ae96fc745917a3a0c0de7a818c9a05012">m_OutputDetailsToStdOut</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00132"></a><span class="lineno"> 132</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#ab17deb382179697b4702cc4f909e71f8">m_OutputDetailsOnlyToStdOut</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00133"></a><span class="lineno"> 133</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a9f8881646a38f828f92d3354930c4165">m_CachedNetworkFilePath</a>(<span class="stringliteral">""</span>)</div><div class="line"><a name="l00134"></a><span class="lineno"> 134</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#adc650e032e7fce99f15e2bf903e7167b">m_NumberOfThreads</a>(0)</div><div class="line"><a name="l00135"></a><span class="lineno"> 135</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#ac609a217f4de4c647422dbb358a5f2ef">m_MLGOTuningFilePath</a>(<span class="stringliteral">""</span>)</div><div class="line"><a name="l00136"></a><span class="lineno"> 136</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#af5e88d9bb91dff85ad5a6f7e190aff4e">m_AsyncEnabled</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00137"></a><span class="lineno"> 137</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a7adc5dcfe3d76ac489f253c4d5f439c8">m_ThreadPoolSize</a>(0)</div><div class="line"><a name="l00138"></a><span class="lineno"> 138</span>  , <a class="code" href="struct_inference_model_internal_1_1_params.xhtml#a1697b9751b4ba381f89b8c81cd9dfc3c">m_ImportInputsIfAligned</a>(<span class="keyword">false</span>)</div><div class="line"><a name="l00139"></a><span class="lineno"> 139</span>  {}</div><div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a513151623e8d448951a0b94ad1946fbe"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a513151623e8d448951a0b94ad1946fbe">InferenceModelInternal::Params::m_EnableFp16TurboMode</a></div><div class="ttdeci">bool m_EnableFp16TurboMode</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00101">InferenceModel.hpp:101</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_aaae50a6c0f73e4c210c2e4331c439482"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#aaae50a6c0f73e4c210c2e4331c439482">InferenceModelInternal::Params::m_VisualizePostOptimizationModel</a></div><div class="ttdeci">bool m_VisualizePostOptimizationModel</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00100">InferenceModel.hpp:100</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a66f9597b152215daba3211379dad63d3"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a66f9597b152215daba3211379dad63d3">InferenceModelInternal::Params::m_EnableFastMath</a></div><div class="ttdeci">bool m_EnableFastMath</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00106">InferenceModel.hpp:106</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a0b99050baebe1d169392461b3a9be08d"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a0b99050baebe1d169392461b3a9be08d">InferenceModelInternal::Params::m_EnableBf16TurboMode</a></div><div class="ttdeci">bool m_EnableBf16TurboMode</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00102">InferenceModel.hpp:102</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a1697b9751b4ba381f89b8c81cd9dfc3c"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a1697b9751b4ba381f89b8c81cd9dfc3c">InferenceModelInternal::Params::m_ImportInputsIfAligned</a></div><div class="ttdeci">bool m_ImportInputsIfAligned</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00115">InferenceModel.hpp:115</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a9a04b8c7f19a84f63125edec509b6d53"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a9a04b8c7f19a84f63125edec509b6d53">InferenceModelInternal::Params::m_SaveCachedNetwork</a></div><div class="ttdeci">bool m_SaveCachedNetwork</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00107">InferenceModel.hpp:107</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a7adc5dcfe3d76ac489f253c4d5f439c8"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a7adc5dcfe3d76ac489f253c4d5f439c8">InferenceModelInternal::Params::m_ThreadPoolSize</a></div><div class="ttdeci">size_t m_ThreadPoolSize</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00114">InferenceModel.hpp:114</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_af5e88d9bb91dff85ad5a6f7e190aff4e"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#af5e88d9bb91dff85ad5a6f7e190aff4e">InferenceModelInternal::Params::m_AsyncEnabled</a></div><div class="ttdeci">bool m_AsyncEnabled</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00113">InferenceModel.hpp:113</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a2d54e6252c1c9a0e29f7706ba03b2b74"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a2d54e6252c1c9a0e29f7706ba03b2b74">InferenceModelInternal::Params::m_ComputeDevices</a></div><div class="ttdeci">std::vector< armnn::BackendId > m_ComputeDevices</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00095">InferenceModel.hpp:95</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a9ef8ddbeee4b869e4c68eb2ed278b8d9"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a9ef8ddbeee4b869e4c68eb2ed278b8d9">InferenceModelInternal::Params::m_AllowExpandedDims</a></div><div class="ttdeci">bool m_AllowExpandedDims</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00098">InferenceModel.hpp:98</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a5c7f0c083da98e7b6e9ba79d2fcd985d"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a5c7f0c083da98e7b6e9ba79d2fcd985d">InferenceModelInternal::Params::m_ParseUnsupported</a></div><div class="ttdeci">bool m_ParseUnsupported</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00104">InferenceModel.hpp:104</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a2a245a63e87f363df491ad8c35be54c5"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a2a245a63e87f363df491ad8c35be54c5">InferenceModelInternal::Params::m_InferOutputShape</a></div><div class="ttdeci">bool m_InferOutputShape</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00105">InferenceModel.hpp:105</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_acde2af8cbbd224a9f94e509ca538a775"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#acde2af8cbbd224a9f94e509ca538a775">InferenceModelInternal::Params::m_PrintIntermediateLayers</a></div><div class="ttdeci">bool m_PrintIntermediateLayers</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00103">InferenceModel.hpp:103</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a9f8881646a38f828f92d3354930c4165"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a9f8881646a38f828f92d3354930c4165">InferenceModelInternal::Params::m_CachedNetworkFilePath</a></div><div class="ttdeci">std::string m_CachedNetworkFilePath</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00110">InferenceModel.hpp:110</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_ae96fc745917a3a0c0de7a818c9a05012"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#ae96fc745917a3a0c0de7a818c9a05012">InferenceModelInternal::Params::m_OutputDetailsToStdOut</a></div><div class="ttdeci">bool m_OutputDetailsToStdOut</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00108">InferenceModel.hpp:108</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_ac609a217f4de4c647422dbb358a5f2ef"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#ac609a217f4de4c647422dbb358a5f2ef">InferenceModelInternal::Params::m_MLGOTuningFilePath</a></div><div class="ttdeci">std::string m_MLGOTuningFilePath</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00112">InferenceModel.hpp:112</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_adc650e032e7fce99f15e2bf903e7167b"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#adc650e032e7fce99f15e2bf903e7167b">InferenceModelInternal::Params::m_NumberOfThreads</a></div><div class="ttdeci">unsigned int m_NumberOfThreads</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00111">InferenceModel.hpp:111</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_ab17deb382179697b4702cc4f909e71f8"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#ab17deb382179697b4702cc4f909e71f8">InferenceModelInternal::Params::m_OutputDetailsOnlyToStdOut</a></div><div class="ttdeci">bool m_OutputDetailsOnlyToStdOut</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00109">InferenceModel.hpp:109</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a6bf2f586c403977d31c7d32d371918cf"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a6bf2f586c403977d31c7d32d371918cf">InferenceModelInternal::Params::m_IsModelBinary</a></div><div class="ttdeci">bool m_IsModelBinary</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00099">InferenceModel.hpp:99</a></div></div> +<div class="ttc" id="struct_inference_model_internal_1_1_params_xhtml_a2d4582aa74998c397bd064ae73745b62"><div class="ttname"><a href="struct_inference_model_internal_1_1_params.xhtml#a2d4582aa74998c397bd064ae73745b62">InferenceModelInternal::Params::m_SubgraphId</a></div><div class="ttdeci">size_t m_SubgraphId</div><div class="ttdef"><b>Definition:</b> <a href="_inference_model_8hpp_source.xhtml#l00097">InferenceModel.hpp:97</a></div></div> +</div><!-- fragment --> +</div> +</div> +<h2 class="groupheader">Member Data Documentation</h2> +<a id="a9ef8ddbeee4b869e4c68eb2ed278b8d9"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a9ef8ddbeee4b869e4c68eb2ed278b8d9">◆ </a></span>m_AllowExpandedDims</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_AllowExpandedDims</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00098">98</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00150">CreateNetworkImpl< IParser >::Create()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>.</p> + +</div> +</div> +<a id="af5e88d9bb91dff85ad5a6f7e190aff4e"></a> +<h2 class="memtitle"><span class="permalink"><a href="#af5e88d9bb91dff85ad5a6f7e190aff4e">◆ </a></span>m_AsyncEnabled</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_AsyncEnabled</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00113">113</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>.</p> + +</div> +</div> +<a id="a9f8881646a38f828f92d3354930c4165"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a9f8881646a38f828f92d3354930c4165">◆ </a></span>m_CachedNetworkFilePath</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">std::string m_CachedNetworkFilePath</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00110">110</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>.</p> + +</div> +</div> +<a id="a2d54e6252c1c9a0e29f7706ba03b2b74"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a2d54e6252c1c9a0e29f7706ba03b2b74">◆ </a></span>m_ComputeDevices</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">std::vector<<a class="el" href="classarmnn_1_1_backend_id.xhtml">armnn::BackendId</a>> m_ComputeDevices</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00095">95</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_model_accuracy_tool-_armnn_8cpp_source.xhtml#l00049">main()</a>, and <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>.</p> + +</div> +</div> +<a id="ae43cf4b5df0068ee6a9151c98947248b"></a> +<h2 class="memtitle"><span class="permalink"><a href="#ae43cf4b5df0068ee6a9151c98947248b">◆ </a></span>m_DynamicBackendsPath</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">std::string m_DynamicBackendsPath</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00096">96</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00796">InferenceModel< IParser, TDataType >::CreateWorkingMemHandle()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, and <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>.</p> + +</div> +</div> +<a id="a0b99050baebe1d169392461b3a9be08d"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a0b99050baebe1d169392461b3a9be08d">◆ </a></span>m_EnableBf16TurboMode</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_EnableBf16TurboMode</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00102">102</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>.</p> + +</div> +</div> +<a id="a66f9597b152215daba3211379dad63d3"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a66f9597b152215daba3211379dad63d3">◆ </a></span>m_EnableFastMath</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_EnableFastMath</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00106">106</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>.</p> + +</div> +</div> +<a id="a513151623e8d448951a0b94ad1946fbe"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a513151623e8d448951a0b94ad1946fbe">◆ </a></span>m_EnableFp16TurboMode</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_EnableFp16TurboMode</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00101">101</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>.</p> + +</div> +</div> +<a id="a1697b9751b4ba381f89b8c81cd9dfc3c"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a1697b9751b4ba381f89b8c81cd9dfc3c">◆ </a></span>m_ImportInputsIfAligned</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_ImportInputsIfAligned</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00115">115</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00796">InferenceModel< IParser, TDataType >::CreateWorkingMemHandle()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00594">InferenceModel< IParser, TDataType >::Run()</a>.</p> + +</div> +</div> +<a id="a2a245a63e87f363df491ad8c35be54c5"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a2a245a63e87f363df491ad8c35be54c5">◆ </a></span>m_InferOutputShape</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_InferOutputShape</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00105">105</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00150">CreateNetworkImpl< IParser >::Create()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>.</p> + +</div> +</div> +<a id="aad2ac35d4cb83ee4da9fad5fbcb907e0"></a> +<h2 class="memtitle"><span class="permalink"><a href="#aad2ac35d4cb83ee4da9fad5fbcb907e0">◆ </a></span>m_InputBindings</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">std::vector<std::string> m_InputBindings</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00092">92</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00566">InferenceModel< IParser, TDataType >::CheckInputIndexIsValid()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00150">CreateNetworkImpl< IParser >::Create()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00796">InferenceModel< IParser, TDataType >::CreateWorkingMemHandle()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00750">InferenceModel< IParser, TDataType >::GetInputBindingInfo()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00756">InferenceModel< IParser, TDataType >::GetInputBindingInfos()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00779">InferenceModel< IParser, TDataType >::GetInputQuantizationParams()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00582">InferenceModel< IParser, TDataType >::GetInputSize()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_model_accuracy_tool-_armnn_8cpp_source.xhtml#l00049">main()</a>, and <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>.</p> + +</div> +</div> +<a id="ad69aa6b4967ce55ee4a915c52c71bf2e"></a> +<h2 class="memtitle"><span class="permalink"><a href="#ad69aa6b4967ce55ee4a915c52c71bf2e">◆ </a></span>m_InputShapes</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">std::vector<<a class="el" href="classarmnn_1_1_tensor_shape.xhtml">armnn::TensorShape</a>> m_InputShapes</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00093">93</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00150">CreateNetworkImpl< IParser >::Create()</a>, and <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>.</p> + +</div> +</div> +<a id="a6bf2f586c403977d31c7d32d371918cf"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a6bf2f586c403977d31c7d32d371918cf">◆ </a></span>m_IsModelBinary</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_IsModelBinary</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00099">99</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00150">CreateNetworkImpl< IParser >::Create()</a>, <a class="el" href="_model_accuracy_tool-_armnn_8cpp_source.xhtml#l00049">main()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>.</p> + +</div> +</div> +<a id="ac609a217f4de4c647422dbb358a5f2ef"></a> +<h2 class="memtitle"><span class="permalink"><a href="#ac609a217f4de4c647422dbb358a5f2ef">◆ </a></span>m_MLGOTuningFilePath</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">std::string m_MLGOTuningFilePath</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00112">112</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>.</p> + +</div> +</div> +<a id="a4fa312cf0d60fbd3988a7c76ab8e2980"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a4fa312cf0d60fbd3988a7c76ab8e2980">◆ </a></span>m_ModelPath</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">std::string m_ModelPath</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00091">91</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_test_8inl_source.xhtml#l00362">armnn::test::ClassifierInferenceTestMain()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00150">CreateNetworkImpl< IParser >::Create()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_model_accuracy_tool-_armnn_8cpp_source.xhtml#l00049">main()</a>, and <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>.</p> + +</div> +</div> +<a id="adc650e032e7fce99f15e2bf903e7167b"></a> +<h2 class="memtitle"><span class="permalink"><a href="#adc650e032e7fce99f15e2bf903e7167b">◆ </a></span>m_NumberOfThreads</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">unsigned int m_NumberOfThreads</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00111">111</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>.</p> + +</div> +</div> +<a id="abeacb4ed1ca9256ee0e8aea73185a0cc"></a> +<h2 class="memtitle"><span class="permalink"><a href="#abeacb4ed1ca9256ee0e8aea73185a0cc">◆ </a></span>m_OutputBindings</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">std::vector<std::string> m_OutputBindings</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00094">94</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00574">InferenceModel< IParser, TDataType >::CheckOutputIndexIsValid()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00150">CreateNetworkImpl< IParser >::Create()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00796">InferenceModel< IParser, TDataType >::CreateWorkingMemHandle()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00786">InferenceModel< IParser, TDataType >::GetAllQuantizationParams()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00761">InferenceModel< IParser, TDataType >::GetOutputBindingInfo()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00767">InferenceModel< IParser, TDataType >::GetOutputBindingInfos()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00588">InferenceModel< IParser, TDataType >::GetOutputSize()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00772">InferenceModel< IParser, TDataType >::GetQuantizationParams()</a>, <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_model_accuracy_tool-_armnn_8cpp_source.xhtml#l00049">main()</a>, and <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>.</p> + +</div> +</div> +<a id="ab17deb382179697b4702cc4f909e71f8"></a> +<h2 class="memtitle"><span class="permalink"><a href="#ab17deb382179697b4702cc4f909e71f8">◆ </a></span>m_OutputDetailsOnlyToStdOut</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_OutputDetailsOnlyToStdOut</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00109">109</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>.</p> + +</div> +</div> +<a id="ae96fc745917a3a0c0de7a818c9a05012"></a> +<h2 class="memtitle"><span class="permalink"><a href="#ae96fc745917a3a0c0de7a818c9a05012">◆ </a></span>m_OutputDetailsToStdOut</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_OutputDetailsToStdOut</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00108">108</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>.</p> + +</div> +</div> +<a id="a5c7f0c083da98e7b6e9ba79d2fcd985d"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a5c7f0c083da98e7b6e9ba79d2fcd985d">◆ </a></span>m_ParseUnsupported</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_ParseUnsupported</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00104">104</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00150">CreateNetworkImpl< IParser >::Create()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>.</p> + +</div> +</div> +<a id="acde2af8cbbd224a9f94e509ca538a775"></a> +<h2 class="memtitle"><span class="permalink"><a href="#acde2af8cbbd224a9f94e509ca538a775">◆ </a></span>m_PrintIntermediateLayers</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_PrintIntermediateLayers</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00103">103</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>.</p> + +</div> +</div> +<a id="a9a04b8c7f19a84f63125edec509b6d53"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a9a04b8c7f19a84f63125edec509b6d53">◆ </a></span>m_SaveCachedNetwork</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_SaveCachedNetwork</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00107">107</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>.</p> + +</div> +</div> +<a id="a2d4582aa74998c397bd064ae73745b62"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a2d4582aa74998c397bd064ae73745b62">◆ </a></span>m_SubgraphId</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">size_t m_SubgraphId</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00097">97</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00150">CreateNetworkImpl< IParser >::Create()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>.</p> + +</div> +</div> +<a id="a7adc5dcfe3d76ac489f253c4d5f439c8"></a> +<h2 class="memtitle"><span class="permalink"><a href="#a7adc5dcfe3d76ac489f253c4d5f439c8">◆ </a></span>m_ThreadPoolSize</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">size_t m_ThreadPoolSize</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00114">114</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>.</p> + +</div> +</div> +<a id="aaae50a6c0f73e4c210c2e4331c439482"></a> +<h2 class="memtitle"><span class="permalink"><a href="#aaae50a6c0f73e4c210c2e4331c439482">◆ </a></span>m_VisualizePostOptimizationModel</h2> + +<div class="memitem"> +<div class="memproto"> + <table class="memname"> + <tr> + <td class="memname">bool m_VisualizePostOptimizationModel</td> + </tr> + </table> +</div><div class="memdoc"> + +<p class="definition">Definition at line <a class="el" href="_inference_model_8hpp_source.xhtml#l00100">100</a> of file <a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a>.</p> + +<p class="reference">Referenced by <a class="el" href="_inference_model_8hpp_source.xhtml#l00441">InferenceModel< IParser, TDataType >::InferenceModel()</a>, <a class="el" href="_execute_network_8cpp_source.xhtml#l00379">MainImpl()</a>, and <a class="el" href="_inference_model_8hpp_source.xhtml#l00118">Params::Params()</a>.</p> + +</div> +</div> +<hr/>The documentation for this struct was generated from the following file:<ul> +<li>tests/<a class="el" href="_inference_model_8hpp_source.xhtml">InferenceModel.hpp</a></li> +</ul> +</div><!-- contents --> +</div><!-- doc-content --> +<!-- start footer part --> +<div id="nav-path" class="navpath"><!-- id is needed for treeview function! --> + <ul> + <li class="navelem"><a class="el" href="namespace_inference_model_internal.xhtml">InferenceModelInternal</a></li><li class="navelem"><a class="el" href="struct_inference_model_internal_1_1_params.xhtml">Params</a></li> + <li class="footer">Generated on Fri Jun 17 2022 13:20:38 for ArmNN by + <a href="http://www.doxygen.org/index.html"> + <img class="footer" src="doxygen.png" alt="doxygen"/></a> 1.8.13 </li> + </ul> +</div> +</body> +</html> |