diff options
author | Nikhil Raj <nikhil.raj@arm.com> | 2024-02-15 15:02:19 +0000 |
---|---|---|
committer | Nikhil Raj <nikhil.raj@arm.com> | 2024-02-15 15:02:19 +0000 |
commit | 38b600d8abb2c5f7a44511b5deddf441f975d51d (patch) | |
tree | 32a082a9e39f3af53d8678a5c18dee67a7f572ec /latest/classarmnn_1_1_network_impl.html | |
parent | 6f92c8e9f8bb38dcf5dccf8deeff5112ecd8e37c (diff) | |
download | armnn-38b600d8abb2c5f7a44511b5deddf441f975d51d.tar.gz |
IVGCVSW-7968 Update Doxygen docu for 24.02
Signed-off-by: Nikhil Raj <nikhil.raj@arm.com>
Change-Id: I8c1e45815c6cf78f80d6f2c0959a5bbba6cd11de
Diffstat (limited to 'latest/classarmnn_1_1_network_impl.html')
-rw-r--r-- | latest/classarmnn_1_1_network_impl.html | 1540 |
1 files changed, 770 insertions, 770 deletions
diff --git a/latest/classarmnn_1_1_network_impl.html b/latest/classarmnn_1_1_network_impl.html index 0a8a4df874..4cce1c939f 100644 --- a/latest/classarmnn_1_1_network_impl.html +++ b/latest/classarmnn_1_1_network_impl.html @@ -36,7 +36,7 @@ <img alt="ArmNN" src="Arm_NN_horizontal_blue.png" style="max-width: 15rem; margin-top: .5rem; margin-left 13px"/> <td id="projectalign" style="padding-left: 0.9em;"> <div id="projectname"> -  <span id="projectnumber">23.11</span> +  <span id="projectnumber">24.02</span> </div> </td> </tr> @@ -281,10 +281,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02162">2162</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02163"></a><span class="lineno"> 2163</span> : m_NetworkOptions(networkOptions),</div> -<div class="line"><a name="l02164"></a><span class="lineno"> 2164</span>  m_Graph(std::make_unique<Graph>(GetShapeInferenceMethod(), GetAllowExpandedDims()))</div> -<div class="line"><a name="l02165"></a><span class="lineno"> 2165</span> {}</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02173">2173</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02174"></a><span class="lineno"> 2174</span> : m_NetworkOptions(networkOptions),</div> +<div class="line"><a name="l02175"></a><span class="lineno"> 2175</span>  m_Graph(std::make_unique<Graph>(GetShapeInferenceMethod(), GetAllowExpandedDims()))</div> +<div class="line"><a name="l02176"></a><span class="lineno"> 2176</span> {}</div> </div><!-- fragment --> </div> </div> @@ -303,9 +303,9 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02167">2167</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02168"></a><span class="lineno"> 2168</span> {</div> -<div class="line"><a name="l02169"></a><span class="lineno"> 2169</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02178">2178</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02179"></a><span class="lineno"> 2179</span> {</div> +<div class="line"><a name="l02180"></a><span class="lineno"> 2180</span> }</div> </div><!-- fragment --> </div> </div> @@ -336,10 +336,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02303">2303</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02305"></a><span class="lineno"> 2305</span> {</div> -<div class="line"><a name="l02306"></a><span class="lineno"> 2306</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ActivationLayer>(activationDescriptor, name);</div> -<div class="line"><a name="l02307"></a><span class="lineno"> 2307</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02314">2314</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02316"></a><span class="lineno"> 2316</span> {</div> +<div class="line"><a name="l02317"></a><span class="lineno"> 2317</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ActivationLayer>(activationDescriptor, name);</div> +<div class="line"><a name="l02318"></a><span class="lineno"> 2318</span> }</div> </div><!-- fragment --> </div> </div> @@ -359,10 +359,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02349">2349</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02350"></a><span class="lineno"> 2350</span> {</div> -<div class="line"><a name="l02351"></a><span class="lineno"> 2351</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<AdditionLayer>(name);</div> -<div class="line"><a name="l02352"></a><span class="lineno"> 2352</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02360">2360</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02361"></a><span class="lineno"> 2361</span> {</div> +<div class="line"><a name="l02362"></a><span class="lineno"> 2362</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<AdditionLayer>(name);</div> +<div class="line"><a name="l02363"></a><span class="lineno"> 2363</span> }</div> </div><!-- fragment --> </div> </div> @@ -392,10 +392,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02309">2309</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02311"></a><span class="lineno"> 2311</span> {</div> -<div class="line"><a name="l02312"></a><span class="lineno"> 2312</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ArgMinMaxLayer>(argMinMaxDescriptor, name);</div> -<div class="line"><a name="l02313"></a><span class="lineno"> 2313</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02320">2320</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02322"></a><span class="lineno"> 2322</span> {</div> +<div class="line"><a name="l02323"></a><span class="lineno"> 2323</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ArgMinMaxLayer>(argMinMaxDescriptor, name);</div> +<div class="line"><a name="l02324"></a><span class="lineno"> 2324</span> }</div> </div><!-- fragment --> </div> </div> @@ -425,10 +425,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l03024">3024</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l03025"></a><span class="lineno"> 3025</span> {</div> -<div class="line"><a name="l03026"></a><span class="lineno"> 3026</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<BatchMatMulLayer>(desc, name);</div> -<div class="line"><a name="l03027"></a><span class="lineno"> 3027</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l03035">3035</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l03036"></a><span class="lineno"> 3036</span> {</div> +<div class="line"><a name="l03037"></a><span class="lineno"> 3037</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<BatchMatMulLayer>(desc, name);</div> +<div class="line"><a name="l03038"></a><span class="lineno"> 3038</span> }</div> </div><!-- fragment --> </div> </div> @@ -482,17 +482,17 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02364">2364</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02370"></a><span class="lineno"> 2370</span> {</div> -<div class="line"><a name="l02371"></a><span class="lineno"> 2371</span>  <span class="keyword">const</span> <span class="keyword">auto</span> layer = m_Graph->AddLayer<BatchNormalizationLayer>(desc, name);</div> -<div class="line"><a name="l02372"></a><span class="lineno"> 2372</span>  </div> -<div class="line"><a name="l02373"></a><span class="lineno"> 2373</span>  layer->m_Mean = std::make_shared<ScopedTensorHandle>(mean);</div> -<div class="line"><a name="l02374"></a><span class="lineno"> 2374</span>  layer->m_Variance = std::make_shared<ScopedTensorHandle>(variance);</div> -<div class="line"><a name="l02375"></a><span class="lineno"> 2375</span>  layer->m_Beta = std::make_shared<ScopedTensorHandle>(beta);</div> -<div class="line"><a name="l02376"></a><span class="lineno"> 2376</span>  layer->m_Gamma = std::make_shared<ScopedTensorHandle>(gamma);</div> -<div class="line"><a name="l02377"></a><span class="lineno"> 2377</span>  </div> -<div class="line"><a name="l02378"></a><span class="lineno"> 2378</span>  <span class="keywordflow">return</span> layer;</div> -<div class="line"><a name="l02379"></a><span class="lineno"> 2379</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02375">2375</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02381"></a><span class="lineno"> 2381</span> {</div> +<div class="line"><a name="l02382"></a><span class="lineno"> 2382</span>  <span class="keyword">const</span> <span class="keyword">auto</span> layer = m_Graph->AddLayer<BatchNormalizationLayer>(desc, name);</div> +<div class="line"><a name="l02383"></a><span class="lineno"> 2383</span>  </div> +<div class="line"><a name="l02384"></a><span class="lineno"> 2384</span>  layer->m_Mean = std::make_shared<ScopedTensorHandle>(mean);</div> +<div class="line"><a name="l02385"></a><span class="lineno"> 2385</span>  layer->m_Variance = std::make_shared<ScopedTensorHandle>(variance);</div> +<div class="line"><a name="l02386"></a><span class="lineno"> 2386</span>  layer->m_Beta = std::make_shared<ScopedTensorHandle>(beta);</div> +<div class="line"><a name="l02387"></a><span class="lineno"> 2387</span>  layer->m_Gamma = std::make_shared<ScopedTensorHandle>(gamma);</div> +<div class="line"><a name="l02388"></a><span class="lineno"> 2388</span>  </div> +<div class="line"><a name="l02389"></a><span class="lineno"> 2389</span>  <span class="keywordflow">return</span> layer;</div> +<div class="line"><a name="l02390"></a><span class="lineno"> 2390</span> }</div> </div><!-- fragment --> <p class="reference">References <a class="el" href="_batch_normalization_layer_8hpp_source.html#l00019">BatchNormalizationLayer::m_Mean</a>.</p> @@ -524,10 +524,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02182">2182</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02184"></a><span class="lineno"> 2184</span> {</div> -<div class="line"><a name="l02185"></a><span class="lineno"> 2185</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<BatchToSpaceNdLayer>(batchToSpaceNdDescriptor, name);</div> -<div class="line"><a name="l02186"></a><span class="lineno"> 2186</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02193">2193</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02195"></a><span class="lineno"> 2195</span> {</div> +<div class="line"><a name="l02196"></a><span class="lineno"> 2196</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<BatchToSpaceNdLayer>(batchToSpaceNdDescriptor, name);</div> +<div class="line"><a name="l02197"></a><span class="lineno"> 2197</span> }</div> </div><!-- fragment --> </div> </div> @@ -557,10 +557,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l03072">3072</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l03073"></a><span class="lineno"> 3073</span> {</div> -<div class="line"><a name="l03074"></a><span class="lineno"> 3074</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<BroadcastToLayer>(desc, name);</div> -<div class="line"><a name="l03075"></a><span class="lineno"> 3075</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l03083">3083</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l03084"></a><span class="lineno"> 3084</span> {</div> +<div class="line"><a name="l03085"></a><span class="lineno"> 3085</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<BroadcastToLayer>(desc, name);</div> +<div class="line"><a name="l03086"></a><span class="lineno"> 3086</span> }</div> </div><!-- fragment --> </div> </div> @@ -580,10 +580,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02188">2188</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02189"></a><span class="lineno"> 2189</span> {</div> -<div class="line"><a name="l02190"></a><span class="lineno"> 2190</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<CastLayer>(name);</div> -<div class="line"><a name="l02191"></a><span class="lineno"> 2191</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02199">2199</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02200"></a><span class="lineno"> 2200</span> {</div> +<div class="line"><a name="l02201"></a><span class="lineno"> 2201</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<CastLayer>(name);</div> +<div class="line"><a name="l02202"></a><span class="lineno"> 2202</span> }</div> </div><!-- fragment --> </div> </div> @@ -613,10 +613,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02192">2192</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02194"></a><span class="lineno"> 2194</span> {</div> -<div class="line"><a name="l02195"></a><span class="lineno"> 2195</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ChannelShuffleLayer>(channelShuffleDescriptor, name);</div> -<div class="line"><a name="l02196"></a><span class="lineno"> 2196</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02203">2203</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02205"></a><span class="lineno"> 2205</span> {</div> +<div class="line"><a name="l02206"></a><span class="lineno"> 2206</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ChannelShuffleLayer>(channelShuffleDescriptor, name);</div> +<div class="line"><a name="l02207"></a><span class="lineno"> 2207</span> }</div> </div><!-- fragment --> </div> </div> @@ -646,10 +646,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02198">2198</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02200"></a><span class="lineno"> 2200</span> {</div> -<div class="line"><a name="l02201"></a><span class="lineno"> 2201</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ComparisonLayer>(comparisonDescriptor, name);</div> -<div class="line"><a name="l02202"></a><span class="lineno"> 2202</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02209">2209</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02211"></a><span class="lineno"> 2211</span> {</div> +<div class="line"><a name="l02212"></a><span class="lineno"> 2212</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ComparisonLayer>(comparisonDescriptor, name);</div> +<div class="line"><a name="l02213"></a><span class="lineno"> 2213</span> }</div> </div><!-- fragment --> </div> </div> @@ -679,10 +679,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02234">2234</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02236"></a><span class="lineno"> 2236</span> {</div> -<div class="line"><a name="l02237"></a><span class="lineno"> 2237</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ConcatLayer>(concatDescriptor, name);</div> -<div class="line"><a name="l02238"></a><span class="lineno"> 2238</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02245">2245</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02247"></a><span class="lineno"> 2247</span> {</div> +<div class="line"><a name="l02248"></a><span class="lineno"> 2248</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ConcatLayer>(concatDescriptor, name);</div> +<div class="line"><a name="l02249"></a><span class="lineno"> 2249</span> }</div> </div><!-- fragment --> </div> </div> @@ -712,14 +712,14 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02420">2420</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02421"></a><span class="lineno"> 2421</span> {</div> -<div class="line"><a name="l02422"></a><span class="lineno"> 2422</span>  <span class="keyword">auto</span> layer = m_Graph->AddLayer<ConstantLayer>(name);</div> -<div class="line"><a name="l02423"></a><span class="lineno"> 2423</span>  </div> -<div class="line"><a name="l02424"></a><span class="lineno"> 2424</span>  layer->m_LayerOutput = std::make_shared<ScopedTensorHandle>(input);</div> -<div class="line"><a name="l02425"></a><span class="lineno"> 2425</span>  </div> -<div class="line"><a name="l02426"></a><span class="lineno"> 2426</span>  <span class="keywordflow">return</span> layer;</div> -<div class="line"><a name="l02427"></a><span class="lineno"> 2427</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02431">2431</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02432"></a><span class="lineno"> 2432</span> {</div> +<div class="line"><a name="l02433"></a><span class="lineno"> 2433</span>  <span class="keyword">auto</span> layer = m_Graph->AddLayer<ConstantLayer>(name);</div> +<div class="line"><a name="l02434"></a><span class="lineno"> 2434</span>  </div> +<div class="line"><a name="l02435"></a><span class="lineno"> 2435</span>  layer->m_LayerOutput = std::make_shared<ScopedTensorHandle>(input);</div> +<div class="line"><a name="l02436"></a><span class="lineno"> 2436</span>  </div> +<div class="line"><a name="l02437"></a><span class="lineno"> 2437</span>  <span class="keywordflow">return</span> layer;</div> +<div class="line"><a name="l02438"></a><span class="lineno"> 2438</span> }</div> </div><!-- fragment --> <p class="reference">References <a class="el" href="_constant_layer_8hpp_source.html#l00046">ConstantLayer::m_LayerOutput</a>.</p> @@ -741,10 +741,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02246">2246</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02247"></a><span class="lineno"> 2247</span> {</div> -<div class="line"><a name="l02248"></a><span class="lineno"> 2248</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ConvertFp16ToFp32Layer>(name);</div> -<div class="line"><a name="l02249"></a><span class="lineno"> 2249</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02257">2257</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02258"></a><span class="lineno"> 2258</span> {</div> +<div class="line"><a name="l02259"></a><span class="lineno"> 2259</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ConvertFp16ToFp32Layer>(name);</div> +<div class="line"><a name="l02260"></a><span class="lineno"> 2260</span> }</div> </div><!-- fragment --> </div> </div> @@ -764,10 +764,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02251">2251</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02252"></a><span class="lineno"> 2252</span> {</div> -<div class="line"><a name="l02253"></a><span class="lineno"> 2253</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ConvertFp32ToFp16Layer>(name);</div> -<div class="line"><a name="l02254"></a><span class="lineno"> 2254</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02262">2262</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02263"></a><span class="lineno"> 2263</span> {</div> +<div class="line"><a name="l02264"></a><span class="lineno"> 2264</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ConvertFp32ToFp16Layer>(name);</div> +<div class="line"><a name="l02265"></a><span class="lineno"> 2265</span> }</div> </div><!-- fragment --> </div> </div> @@ -797,10 +797,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02240">2240</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02242"></a><span class="lineno"> 2242</span> {</div> -<div class="line"><a name="l02243"></a><span class="lineno"> 2243</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<Convolution2dLayer>(convolution2dDescriptor, name);</div> -<div class="line"><a name="l02244"></a><span class="lineno"> 2244</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02251">2251</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02253"></a><span class="lineno"> 2253</span> {</div> +<div class="line"><a name="l02254"></a><span class="lineno"> 2254</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<Convolution2dLayer>(convolution2dDescriptor, name);</div> +<div class="line"><a name="l02255"></a><span class="lineno"> 2255</span> }</div> </div><!-- fragment --> </div> </div> @@ -830,10 +830,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02256">2256</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02258"></a><span class="lineno"> 2258</span> {</div> -<div class="line"><a name="l02259"></a><span class="lineno"> 2259</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<Convolution3dLayer>(convolution3dDescriptor, name);</div> -<div class="line"><a name="l02260"></a><span class="lineno"> 2260</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02267">2267</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02269"></a><span class="lineno"> 2269</span> {</div> +<div class="line"><a name="l02270"></a><span class="lineno"> 2270</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<Convolution3dLayer>(convolution3dDescriptor, name);</div> +<div class="line"><a name="l02271"></a><span class="lineno"> 2271</span> }</div> </div><!-- fragment --> </div> </div> @@ -863,10 +863,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02262">2262</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02264"></a><span class="lineno"> 2264</span> {</div> -<div class="line"><a name="l02265"></a><span class="lineno"> 2265</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<DepthToSpaceLayer>(depthToSpaceDescriptor, name);</div> -<div class="line"><a name="l02266"></a><span class="lineno"> 2266</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02273">2273</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02275"></a><span class="lineno"> 2275</span> {</div> +<div class="line"><a name="l02276"></a><span class="lineno"> 2276</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<DepthToSpaceLayer>(depthToSpaceDescriptor, name);</div> +<div class="line"><a name="l02277"></a><span class="lineno"> 2277</span> }</div> </div><!-- fragment --> </div> </div> @@ -896,10 +896,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02268">2268</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02271"></a><span class="lineno"> 2271</span> {</div> -<div class="line"><a name="l02272"></a><span class="lineno"> 2272</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<DepthwiseConvolution2dLayer>(convolution2dDescriptor, name);</div> -<div class="line"><a name="l02273"></a><span class="lineno"> 2273</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02279">2279</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02282"></a><span class="lineno"> 2282</span> {</div> +<div class="line"><a name="l02283"></a><span class="lineno"> 2283</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<DepthwiseConvolution2dLayer>(convolution2dDescriptor, name);</div> +<div class="line"><a name="l02284"></a><span class="lineno"> 2284</span> }</div> </div><!-- fragment --> </div> </div> @@ -919,10 +919,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02618">2618</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02619"></a><span class="lineno"> 2619</span> {</div> -<div class="line"><a name="l02620"></a><span class="lineno"> 2620</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<DequantizeLayer>(name);</div> -<div class="line"><a name="l02621"></a><span class="lineno"> 2621</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02629">2629</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02630"></a><span class="lineno"> 2630</span> {</div> +<div class="line"><a name="l02631"></a><span class="lineno"> 2631</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<DequantizeLayer>(name);</div> +<div class="line"><a name="l02632"></a><span class="lineno"> 2632</span> }</div> </div><!-- fragment --> </div> </div> @@ -958,14 +958,14 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02275">2275</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02277"></a><span class="lineno"> 2277</span> {</div> -<div class="line"><a name="l02278"></a><span class="lineno"> 2278</span>  <span class="keyword">const</span> <span class="keyword">auto</span> layer = m_Graph->AddLayer<DetectionPostProcessLayer>(descriptor, name);</div> -<div class="line"><a name="l02279"></a><span class="lineno"> 2279</span>  </div> -<div class="line"><a name="l02280"></a><span class="lineno"> 2280</span>  layer->m_Anchors = std::make_shared<ScopedTensorHandle>(anchors);</div> -<div class="line"><a name="l02281"></a><span class="lineno"> 2281</span>  </div> -<div class="line"><a name="l02282"></a><span class="lineno"> 2282</span>  <span class="keywordflow">return</span> layer;</div> -<div class="line"><a name="l02283"></a><span class="lineno"> 2283</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02286">2286</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02288"></a><span class="lineno"> 2288</span> {</div> +<div class="line"><a name="l02289"></a><span class="lineno"> 2289</span>  <span class="keyword">const</span> <span class="keyword">auto</span> layer = m_Graph->AddLayer<DetectionPostProcessLayer>(descriptor, name);</div> +<div class="line"><a name="l02290"></a><span class="lineno"> 2290</span>  </div> +<div class="line"><a name="l02291"></a><span class="lineno"> 2291</span>  layer->m_Anchors = std::make_shared<ScopedTensorHandle>(anchors);</div> +<div class="line"><a name="l02292"></a><span class="lineno"> 2292</span>  </div> +<div class="line"><a name="l02293"></a><span class="lineno"> 2293</span>  <span class="keywordflow">return</span> layer;</div> +<div class="line"><a name="l02294"></a><span class="lineno"> 2294</span> }</div> </div><!-- fragment --> <p class="reference">References <a class="el" href="_detection_post_process_layer_8hpp_source.html#l00020">DetectionPostProcessLayer::m_Anchors</a>.</p> @@ -987,10 +987,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02593">2593</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02594"></a><span class="lineno"> 2594</span> {</div> -<div class="line"><a name="l02595"></a><span class="lineno"> 2595</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<DivisionLayer>(name);</div> -<div class="line"><a name="l02596"></a><span class="lineno"> 2596</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02604">2604</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02605"></a><span class="lineno"> 2605</span> {</div> +<div class="line"><a name="l02606"></a><span class="lineno"> 2606</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<DivisionLayer>(name);</div> +<div class="line"><a name="l02607"></a><span class="lineno"> 2607</span> }</div> </div><!-- fragment --> </div> </div> @@ -1020,10 +1020,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02204">2204</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02206"></a><span class="lineno"> 2206</span> {</div> -<div class="line"><a name="l02207"></a><span class="lineno"> 2207</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ElementwiseBinaryLayer>(elementwiseBinaryDesc, name);</div> -<div class="line"><a name="l02208"></a><span class="lineno"> 2208</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02215">2215</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02217"></a><span class="lineno"> 2217</span> {</div> +<div class="line"><a name="l02218"></a><span class="lineno"> 2218</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ElementwiseBinaryLayer>(elementwiseBinaryDesc, name);</div> +<div class="line"><a name="l02219"></a><span class="lineno"> 2219</span> }</div> </div><!-- fragment --> </div> </div> @@ -1053,10 +1053,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02210">2210</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02212"></a><span class="lineno"> 2212</span> {</div> -<div class="line"><a name="l02213"></a><span class="lineno"> 2213</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ElementwiseUnaryLayer>(elementwiseUnaryDescriptor, name);</div> -<div class="line"><a name="l02214"></a><span class="lineno"> 2214</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02221">2221</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02223"></a><span class="lineno"> 2223</span> {</div> +<div class="line"><a name="l02224"></a><span class="lineno"> 2224</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ElementwiseUnaryLayer>(elementwiseUnaryDescriptor, name);</div> +<div class="line"><a name="l02225"></a><span class="lineno"> 2225</span> }</div> </div><!-- fragment --> </div> </div> @@ -1086,10 +1086,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02216">2216</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02218"></a><span class="lineno"> 2218</span> {</div> -<div class="line"><a name="l02219"></a><span class="lineno"> 2219</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<FillLayer>(fillDescriptor, name);</div> -<div class="line"><a name="l02220"></a><span class="lineno"> 2220</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02227">2227</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02229"></a><span class="lineno"> 2229</span> {</div> +<div class="line"><a name="l02230"></a><span class="lineno"> 2230</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<FillLayer>(fillDescriptor, name);</div> +<div class="line"><a name="l02231"></a><span class="lineno"> 2231</span> }</div> </div><!-- fragment --> </div> </div> @@ -1109,10 +1109,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02447">2447</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02448"></a><span class="lineno"> 2448</span> {</div> -<div class="line"><a name="l02449"></a><span class="lineno"> 2449</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<FloorLayer>(name);</div> -<div class="line"><a name="l02450"></a><span class="lineno"> 2450</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02458">2458</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02459"></a><span class="lineno"> 2459</span> {</div> +<div class="line"><a name="l02460"></a><span class="lineno"> 2460</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<FloorLayer>(name);</div> +<div class="line"><a name="l02461"></a><span class="lineno"> 2461</span> }</div> </div><!-- fragment --> </div> </div> @@ -1142,10 +1142,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02222">2222</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02224"></a><span class="lineno"> 2224</span> {</div> -<div class="line"><a name="l02225"></a><span class="lineno"> 2225</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<FullyConnectedLayer>(fullyConnectedDescriptor, name);</div> -<div class="line"><a name="l02226"></a><span class="lineno"> 2226</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02233">2233</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02235"></a><span class="lineno"> 2235</span> {</div> +<div class="line"><a name="l02236"></a><span class="lineno"> 2236</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<FullyConnectedLayer>(fullyConnectedDescriptor, name);</div> +<div class="line"><a name="l02237"></a><span class="lineno"> 2237</span> }</div> </div><!-- fragment --> </div> </div> @@ -1175,10 +1175,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02228">2228</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02230"></a><span class="lineno"> 2230</span> {</div> -<div class="line"><a name="l02231"></a><span class="lineno"> 2231</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<FusedLayer>(fusedDescriptor, name);</div> -<div class="line"><a name="l02232"></a><span class="lineno"> 2232</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02239">2239</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02241"></a><span class="lineno"> 2241</span> {</div> +<div class="line"><a name="l02242"></a><span class="lineno"> 2242</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<FusedLayer>(fusedDescriptor, name);</div> +<div class="line"><a name="l02243"></a><span class="lineno"> 2243</span> }</div> </div><!-- fragment --> </div> </div> @@ -1208,10 +1208,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02629">2629</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02631"></a><span class="lineno"> 2631</span> {</div> -<div class="line"><a name="l02632"></a><span class="lineno"> 2632</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<GatherLayer>(gatherDescriptor, name);</div> -<div class="line"><a name="l02633"></a><span class="lineno"> 2633</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02640">2640</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02642"></a><span class="lineno"> 2642</span> {</div> +<div class="line"><a name="l02643"></a><span class="lineno"> 2643</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<GatherLayer>(gatherDescriptor, name);</div> +<div class="line"><a name="l02644"></a><span class="lineno"> 2644</span> }</div> </div><!-- fragment --> </div> </div> @@ -1231,10 +1231,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02635">2635</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02636"></a><span class="lineno"> 2636</span> {</div> -<div class="line"><a name="l02637"></a><span class="lineno"> 2637</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<GatherNdLayer>(name);</div> -<div class="line"><a name="l02638"></a><span class="lineno"> 2638</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02646">2646</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02647"></a><span class="lineno"> 2647</span> {</div> +<div class="line"><a name="l02648"></a><span class="lineno"> 2648</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<GatherNdLayer>(name);</div> +<div class="line"><a name="l02649"></a><span class="lineno"> 2649</span> }</div> </div><!-- fragment --> </div> </div> @@ -1264,10 +1264,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02177">2177</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02178"></a><span class="lineno"> 2178</span> {</div> -<div class="line"><a name="l02179"></a><span class="lineno"> 2179</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<InputLayer>(id, name);</div> -<div class="line"><a name="l02180"></a><span class="lineno"> 2180</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02188">2188</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02189"></a><span class="lineno"> 2189</span> {</div> +<div class="line"><a name="l02190"></a><span class="lineno"> 2190</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<InputLayer>(id, name);</div> +<div class="line"><a name="l02191"></a><span class="lineno"> 2191</span> }</div> </div><!-- fragment --> </div> </div> @@ -1297,10 +1297,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02402">2402</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02404"></a><span class="lineno"> 2404</span> {</div> -<div class="line"><a name="l02405"></a><span class="lineno"> 2405</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<InstanceNormalizationLayer>(desc, name);</div> -<div class="line"><a name="l02406"></a><span class="lineno"> 2406</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02413">2413</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02415"></a><span class="lineno"> 2415</span> {</div> +<div class="line"><a name="l02416"></a><span class="lineno"> 2416</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<InstanceNormalizationLayer>(desc, name);</div> +<div class="line"><a name="l02417"></a><span class="lineno"> 2417</span> }</div> </div><!-- fragment --> </div> </div> @@ -1330,10 +1330,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02408">2408</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02410"></a><span class="lineno"> 2410</span> {</div> -<div class="line"><a name="l02411"></a><span class="lineno"> 2411</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<L2NormalizationLayer>(desc, name);</div> -<div class="line"><a name="l02412"></a><span class="lineno"> 2412</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02419">2419</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02421"></a><span class="lineno"> 2421</span> {</div> +<div class="line"><a name="l02422"></a><span class="lineno"> 2422</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<L2NormalizationLayer>(desc, name);</div> +<div class="line"><a name="l02423"></a><span class="lineno"> 2423</span> }</div> </div><!-- fragment --> </div> </div> @@ -1363,10 +1363,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02876">2876</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02878"></a><span class="lineno"> 2878</span> {</div> -<div class="line"><a name="l02879"></a><span class="lineno"> 2879</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<LogicalBinaryLayer>(logicalBinaryDescriptor, name);</div> -<div class="line"><a name="l02880"></a><span class="lineno"> 2880</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02887">2887</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02889"></a><span class="lineno"> 2889</span> {</div> +<div class="line"><a name="l02890"></a><span class="lineno"> 2890</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<LogicalBinaryLayer>(logicalBinaryDescriptor, name);</div> +<div class="line"><a name="l02891"></a><span class="lineno"> 2891</span> }</div> </div><!-- fragment --> </div> </div> @@ -1396,10 +1396,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02414">2414</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02416"></a><span class="lineno"> 2416</span> {</div> -<div class="line"><a name="l02417"></a><span class="lineno"> 2417</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<LogSoftmaxLayer>(desc, name);</div> -<div class="line"><a name="l02418"></a><span class="lineno"> 2418</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02425">2425</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02427"></a><span class="lineno"> 2427</span> {</div> +<div class="line"><a name="l02428"></a><span class="lineno"> 2428</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<LogSoftmaxLayer>(desc, name);</div> +<div class="line"><a name="l02429"></a><span class="lineno"> 2429</span> }</div> </div><!-- fragment --> </div> </div> @@ -1435,144 +1435,144 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02452">2452</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02455"></a><span class="lineno"> 2455</span> {</div> -<div class="line"><a name="l02456"></a><span class="lineno"> 2456</span>  <span class="keyword">const</span> <span class="keyword">auto</span> layer = m_Graph->AddLayer<LstmLayer>(descriptor, name);</div> -<div class="line"><a name="l02457"></a><span class="lineno"> 2457</span>  </div> -<div class="line"><a name="l02458"></a><span class="lineno"> 2458</span>  <span class="comment">//Lstm Basic Parameters</span></div> -<div class="line"><a name="l02459"></a><span class="lineno"> 2459</span>  layer->m_BasicParameters.m_InputToForgetWeights =</div> -<div class="line"><a name="l02460"></a><span class="lineno"> 2460</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToForgetWeights));</div> -<div class="line"><a name="l02461"></a><span class="lineno"> 2461</span>  layer->m_BasicParameters.m_InputToCellWeights =</div> -<div class="line"><a name="l02462"></a><span class="lineno"> 2462</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToCellWeights));</div> -<div class="line"><a name="l02463"></a><span class="lineno"> 2463</span>  layer->m_BasicParameters.m_InputToOutputWeights =</div> -<div class="line"><a name="l02464"></a><span class="lineno"> 2464</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToOutputWeights));</div> -<div class="line"><a name="l02465"></a><span class="lineno"> 2465</span>  layer->m_BasicParameters.m_RecurrentToForgetWeights =</div> -<div class="line"><a name="l02466"></a><span class="lineno"> 2466</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToForgetWeights));</div> -<div class="line"><a name="l02467"></a><span class="lineno"> 2467</span>  layer->m_BasicParameters.m_RecurrentToCellWeights =</div> -<div class="line"><a name="l02468"></a><span class="lineno"> 2468</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToCellWeights));</div> -<div class="line"><a name="l02469"></a><span class="lineno"> 2469</span>  layer->m_BasicParameters.m_RecurrentToOutputWeights =</div> -<div class="line"><a name="l02470"></a><span class="lineno"> 2470</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToOutputWeights));</div> -<div class="line"><a name="l02471"></a><span class="lineno"> 2471</span>  layer->m_BasicParameters.m_ForgetGateBias =</div> -<div class="line"><a name="l02472"></a><span class="lineno"> 2472</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ForgetGateBias));</div> -<div class="line"><a name="l02473"></a><span class="lineno"> 2473</span>  layer->m_BasicParameters.m_CellBias =</div> -<div class="line"><a name="l02474"></a><span class="lineno"> 2474</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellBias));</div> -<div class="line"><a name="l02475"></a><span class="lineno"> 2475</span>  layer->m_BasicParameters.m_OutputGateBias =</div> -<div class="line"><a name="l02476"></a><span class="lineno"> 2476</span>  std::make_shared<ScopedTensorHandle>(*(params.m_OutputGateBias));</div> -<div class="line"><a name="l02477"></a><span class="lineno"> 2477</span>  </div> -<div class="line"><a name="l02478"></a><span class="lineno"> 2478</span>  <span class="comment">//Lstm Cifg parameters</span></div> -<div class="line"><a name="l02479"></a><span class="lineno"> 2479</span>  <span class="keywordflow">if</span>(!descriptor.m_CifgEnabled)</div> -<div class="line"><a name="l02480"></a><span class="lineno"> 2480</span>  {</div> -<div class="line"><a name="l02481"></a><span class="lineno"> 2481</span>  <span class="keywordflow">if</span>(params.m_InputToInputWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02482"></a><span class="lineno"> 2482</span>  {</div> -<div class="line"><a name="l02483"></a><span class="lineno"> 2483</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Input To Input Weights cannot be NULL "</span></div> -<div class="line"><a name="l02484"></a><span class="lineno"> 2484</span>  <span class="stringliteral">"when CIFG is disabled."</span>);</div> -<div class="line"><a name="l02485"></a><span class="lineno"> 2485</span>  }</div> -<div class="line"><a name="l02486"></a><span class="lineno"> 2486</span>  <span class="keywordflow">if</span>(params.m_RecurrentToInputWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02487"></a><span class="lineno"> 2487</span>  {</div> -<div class="line"><a name="l02488"></a><span class="lineno"> 2488</span>  <span class="keywordflow">throw</span> InvalidArgumentException(</div> -<div class="line"><a name="l02489"></a><span class="lineno"> 2489</span>  <span class="stringliteral">"AddLstmLayer: Recurrent To Input Weights cannot be NULL "</span></div> -<div class="line"><a name="l02490"></a><span class="lineno"> 2490</span>  <span class="stringliteral">"when CIFG is disabled."</span>);</div> -<div class="line"><a name="l02491"></a><span class="lineno"> 2491</span>  }</div> -<div class="line"><a name="l02492"></a><span class="lineno"> 2492</span>  <span class="keywordflow">if</span>(params.m_InputGateBias == <span class="keyword">nullptr</span>)</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02463">2463</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02466"></a><span class="lineno"> 2466</span> {</div> +<div class="line"><a name="l02467"></a><span class="lineno"> 2467</span>  <span class="keyword">const</span> <span class="keyword">auto</span> layer = m_Graph->AddLayer<LstmLayer>(descriptor, name);</div> +<div class="line"><a name="l02468"></a><span class="lineno"> 2468</span>  </div> +<div class="line"><a name="l02469"></a><span class="lineno"> 2469</span>  <span class="comment">//Lstm Basic Parameters</span></div> +<div class="line"><a name="l02470"></a><span class="lineno"> 2470</span>  layer->m_BasicParameters.m_InputToForgetWeights =</div> +<div class="line"><a name="l02471"></a><span class="lineno"> 2471</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToForgetWeights));</div> +<div class="line"><a name="l02472"></a><span class="lineno"> 2472</span>  layer->m_BasicParameters.m_InputToCellWeights =</div> +<div class="line"><a name="l02473"></a><span class="lineno"> 2473</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToCellWeights));</div> +<div class="line"><a name="l02474"></a><span class="lineno"> 2474</span>  layer->m_BasicParameters.m_InputToOutputWeights =</div> +<div class="line"><a name="l02475"></a><span class="lineno"> 2475</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToOutputWeights));</div> +<div class="line"><a name="l02476"></a><span class="lineno"> 2476</span>  layer->m_BasicParameters.m_RecurrentToForgetWeights =</div> +<div class="line"><a name="l02477"></a><span class="lineno"> 2477</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToForgetWeights));</div> +<div class="line"><a name="l02478"></a><span class="lineno"> 2478</span>  layer->m_BasicParameters.m_RecurrentToCellWeights =</div> +<div class="line"><a name="l02479"></a><span class="lineno"> 2479</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToCellWeights));</div> +<div class="line"><a name="l02480"></a><span class="lineno"> 2480</span>  layer->m_BasicParameters.m_RecurrentToOutputWeights =</div> +<div class="line"><a name="l02481"></a><span class="lineno"> 2481</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToOutputWeights));</div> +<div class="line"><a name="l02482"></a><span class="lineno"> 2482</span>  layer->m_BasicParameters.m_ForgetGateBias =</div> +<div class="line"><a name="l02483"></a><span class="lineno"> 2483</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ForgetGateBias));</div> +<div class="line"><a name="l02484"></a><span class="lineno"> 2484</span>  layer->m_BasicParameters.m_CellBias =</div> +<div class="line"><a name="l02485"></a><span class="lineno"> 2485</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellBias));</div> +<div class="line"><a name="l02486"></a><span class="lineno"> 2486</span>  layer->m_BasicParameters.m_OutputGateBias =</div> +<div class="line"><a name="l02487"></a><span class="lineno"> 2487</span>  std::make_shared<ScopedTensorHandle>(*(params.m_OutputGateBias));</div> +<div class="line"><a name="l02488"></a><span class="lineno"> 2488</span>  </div> +<div class="line"><a name="l02489"></a><span class="lineno"> 2489</span>  <span class="comment">//Lstm Cifg parameters</span></div> +<div class="line"><a name="l02490"></a><span class="lineno"> 2490</span>  <span class="keywordflow">if</span>(!descriptor.m_CifgEnabled)</div> +<div class="line"><a name="l02491"></a><span class="lineno"> 2491</span>  {</div> +<div class="line"><a name="l02492"></a><span class="lineno"> 2492</span>  <span class="keywordflow">if</span>(params.m_InputToInputWeights == <span class="keyword">nullptr</span>)</div> <div class="line"><a name="l02493"></a><span class="lineno"> 2493</span>  {</div> -<div class="line"><a name="l02494"></a><span class="lineno"> 2494</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Input Gate Bias cannot be NULL "</span></div> +<div class="line"><a name="l02494"></a><span class="lineno"> 2494</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Input To Input Weights cannot be NULL "</span></div> <div class="line"><a name="l02495"></a><span class="lineno"> 2495</span>  <span class="stringliteral">"when CIFG is disabled."</span>);</div> <div class="line"><a name="l02496"></a><span class="lineno"> 2496</span>  }</div> -<div class="line"><a name="l02497"></a><span class="lineno"> 2497</span>  layer->m_CifgParameters.m_InputToInputWeights =</div> -<div class="line"><a name="l02498"></a><span class="lineno"> 2498</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToInputWeights));</div> -<div class="line"><a name="l02499"></a><span class="lineno"> 2499</span>  layer->m_CifgParameters.m_RecurrentToInputWeights =</div> -<div class="line"><a name="l02500"></a><span class="lineno"> 2500</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToInputWeights));</div> -<div class="line"><a name="l02501"></a><span class="lineno"> 2501</span>  layer->m_CifgParameters.m_InputGateBias =</div> -<div class="line"><a name="l02502"></a><span class="lineno"> 2502</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputGateBias));</div> -<div class="line"><a name="l02503"></a><span class="lineno"> 2503</span>  }</div> -<div class="line"><a name="l02504"></a><span class="lineno"> 2504</span>  </div> -<div class="line"><a name="l02505"></a><span class="lineno"> 2505</span>  <span class="comment">//Lstm projection parameters</span></div> -<div class="line"><a name="l02506"></a><span class="lineno"> 2506</span>  <span class="keywordflow">if</span>(descriptor.m_ProjectionEnabled)</div> -<div class="line"><a name="l02507"></a><span class="lineno"> 2507</span>  {</div> -<div class="line"><a name="l02508"></a><span class="lineno"> 2508</span>  <span class="keywordflow">if</span>(params.m_ProjectionWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02509"></a><span class="lineno"> 2509</span>  {</div> -<div class="line"><a name="l02510"></a><span class="lineno"> 2510</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Projection Weights cannot be NULL "</span></div> -<div class="line"><a name="l02511"></a><span class="lineno"> 2511</span>  <span class="stringliteral">"when projection is enabled."</span>);</div> -<div class="line"><a name="l02512"></a><span class="lineno"> 2512</span>  }</div> -<div class="line"><a name="l02513"></a><span class="lineno"> 2513</span>  layer->m_ProjectionParameters.m_ProjectionWeights =</div> -<div class="line"><a name="l02514"></a><span class="lineno"> 2514</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ProjectionWeights));</div> -<div class="line"><a name="l02515"></a><span class="lineno"> 2515</span>  <span class="keywordflow">if</span>(params.m_ProjectionBias != <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02516"></a><span class="lineno"> 2516</span>  {</div> -<div class="line"><a name="l02517"></a><span class="lineno"> 2517</span>  layer->m_ProjectionParameters.m_ProjectionBias =</div> -<div class="line"><a name="l02518"></a><span class="lineno"> 2518</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ProjectionBias));</div> -<div class="line"><a name="l02519"></a><span class="lineno"> 2519</span>  }</div> -<div class="line"><a name="l02520"></a><span class="lineno"> 2520</span>  }</div> -<div class="line"><a name="l02521"></a><span class="lineno"> 2521</span>  </div> -<div class="line"><a name="l02522"></a><span class="lineno"> 2522</span>  <span class="comment">//Lstm Peephole params</span></div> -<div class="line"><a name="l02523"></a><span class="lineno"> 2523</span>  <span class="keywordflow">if</span>(descriptor.m_PeepholeEnabled)</div> -<div class="line"><a name="l02524"></a><span class="lineno"> 2524</span>  {</div> -<div class="line"><a name="l02525"></a><span class="lineno"> 2525</span>  <span class="keywordflow">if</span>(!descriptor.m_CifgEnabled)</div> -<div class="line"><a name="l02526"></a><span class="lineno"> 2526</span>  {</div> -<div class="line"><a name="l02527"></a><span class="lineno"> 2527</span>  <span class="keywordflow">if</span>(params.m_CellToInputWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02528"></a><span class="lineno"> 2528</span>  {</div> -<div class="line"><a name="l02529"></a><span class="lineno"> 2529</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Cell To Input Weights cannot be NULL "</span></div> -<div class="line"><a name="l02530"></a><span class="lineno"> 2530</span>  <span class="stringliteral">"when Peephole is enabled and CIFG disabled."</span>);</div> -<div class="line"><a name="l02531"></a><span class="lineno"> 2531</span>  }</div> +<div class="line"><a name="l02497"></a><span class="lineno"> 2497</span>  <span class="keywordflow">if</span>(params.m_RecurrentToInputWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02498"></a><span class="lineno"> 2498</span>  {</div> +<div class="line"><a name="l02499"></a><span class="lineno"> 2499</span>  <span class="keywordflow">throw</span> InvalidArgumentException(</div> +<div class="line"><a name="l02500"></a><span class="lineno"> 2500</span>  <span class="stringliteral">"AddLstmLayer: Recurrent To Input Weights cannot be NULL "</span></div> +<div class="line"><a name="l02501"></a><span class="lineno"> 2501</span>  <span class="stringliteral">"when CIFG is disabled."</span>);</div> +<div class="line"><a name="l02502"></a><span class="lineno"> 2502</span>  }</div> +<div class="line"><a name="l02503"></a><span class="lineno"> 2503</span>  <span class="keywordflow">if</span>(params.m_InputGateBias == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02504"></a><span class="lineno"> 2504</span>  {</div> +<div class="line"><a name="l02505"></a><span class="lineno"> 2505</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Input Gate Bias cannot be NULL "</span></div> +<div class="line"><a name="l02506"></a><span class="lineno"> 2506</span>  <span class="stringliteral">"when CIFG is disabled."</span>);</div> +<div class="line"><a name="l02507"></a><span class="lineno"> 2507</span>  }</div> +<div class="line"><a name="l02508"></a><span class="lineno"> 2508</span>  layer->m_CifgParameters.m_InputToInputWeights =</div> +<div class="line"><a name="l02509"></a><span class="lineno"> 2509</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToInputWeights));</div> +<div class="line"><a name="l02510"></a><span class="lineno"> 2510</span>  layer->m_CifgParameters.m_RecurrentToInputWeights =</div> +<div class="line"><a name="l02511"></a><span class="lineno"> 2511</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToInputWeights));</div> +<div class="line"><a name="l02512"></a><span class="lineno"> 2512</span>  layer->m_CifgParameters.m_InputGateBias =</div> +<div class="line"><a name="l02513"></a><span class="lineno"> 2513</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputGateBias));</div> +<div class="line"><a name="l02514"></a><span class="lineno"> 2514</span>  }</div> +<div class="line"><a name="l02515"></a><span class="lineno"> 2515</span>  </div> +<div class="line"><a name="l02516"></a><span class="lineno"> 2516</span>  <span class="comment">//Lstm projection parameters</span></div> +<div class="line"><a name="l02517"></a><span class="lineno"> 2517</span>  <span class="keywordflow">if</span>(descriptor.m_ProjectionEnabled)</div> +<div class="line"><a name="l02518"></a><span class="lineno"> 2518</span>  {</div> +<div class="line"><a name="l02519"></a><span class="lineno"> 2519</span>  <span class="keywordflow">if</span>(params.m_ProjectionWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02520"></a><span class="lineno"> 2520</span>  {</div> +<div class="line"><a name="l02521"></a><span class="lineno"> 2521</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Projection Weights cannot be NULL "</span></div> +<div class="line"><a name="l02522"></a><span class="lineno"> 2522</span>  <span class="stringliteral">"when projection is enabled."</span>);</div> +<div class="line"><a name="l02523"></a><span class="lineno"> 2523</span>  }</div> +<div class="line"><a name="l02524"></a><span class="lineno"> 2524</span>  layer->m_ProjectionParameters.m_ProjectionWeights =</div> +<div class="line"><a name="l02525"></a><span class="lineno"> 2525</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ProjectionWeights));</div> +<div class="line"><a name="l02526"></a><span class="lineno"> 2526</span>  <span class="keywordflow">if</span>(params.m_ProjectionBias != <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02527"></a><span class="lineno"> 2527</span>  {</div> +<div class="line"><a name="l02528"></a><span class="lineno"> 2528</span>  layer->m_ProjectionParameters.m_ProjectionBias =</div> +<div class="line"><a name="l02529"></a><span class="lineno"> 2529</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ProjectionBias));</div> +<div class="line"><a name="l02530"></a><span class="lineno"> 2530</span>  }</div> +<div class="line"><a name="l02531"></a><span class="lineno"> 2531</span>  }</div> <div class="line"><a name="l02532"></a><span class="lineno"> 2532</span>  </div> -<div class="line"><a name="l02533"></a><span class="lineno"> 2533</span>  layer->m_PeepholeParameters.m_CellToInputWeights =</div> -<div class="line"><a name="l02534"></a><span class="lineno"> 2534</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellToInputWeights));</div> -<div class="line"><a name="l02535"></a><span class="lineno"> 2535</span>  }</div> -<div class="line"><a name="l02536"></a><span class="lineno"> 2536</span>  </div> -<div class="line"><a name="l02537"></a><span class="lineno"> 2537</span>  <span class="keywordflow">if</span>(params.m_CellToForgetWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02538"></a><span class="lineno"> 2538</span>  {</div> -<div class="line"><a name="l02539"></a><span class="lineno"> 2539</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Cell To Forget Weights cannot be NULL "</span></div> -<div class="line"><a name="l02540"></a><span class="lineno"> 2540</span>  <span class="stringliteral">"when Peephole is enabled."</span>);</div> -<div class="line"><a name="l02541"></a><span class="lineno"> 2541</span>  }</div> -<div class="line"><a name="l02542"></a><span class="lineno"> 2542</span>  <span class="keywordflow">if</span>(params.m_CellToOutputWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02543"></a><span class="lineno"> 2543</span>  {</div> -<div class="line"><a name="l02544"></a><span class="lineno"> 2544</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Cell To Output Weights cannot be NULL "</span></div> -<div class="line"><a name="l02545"></a><span class="lineno"> 2545</span>  <span class="stringliteral">"when Peephole is enabled."</span>);</div> +<div class="line"><a name="l02533"></a><span class="lineno"> 2533</span>  <span class="comment">//Lstm Peephole params</span></div> +<div class="line"><a name="l02534"></a><span class="lineno"> 2534</span>  <span class="keywordflow">if</span>(descriptor.m_PeepholeEnabled)</div> +<div class="line"><a name="l02535"></a><span class="lineno"> 2535</span>  {</div> +<div class="line"><a name="l02536"></a><span class="lineno"> 2536</span>  <span class="keywordflow">if</span>(!descriptor.m_CifgEnabled)</div> +<div class="line"><a name="l02537"></a><span class="lineno"> 2537</span>  {</div> +<div class="line"><a name="l02538"></a><span class="lineno"> 2538</span>  <span class="keywordflow">if</span>(params.m_CellToInputWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02539"></a><span class="lineno"> 2539</span>  {</div> +<div class="line"><a name="l02540"></a><span class="lineno"> 2540</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Cell To Input Weights cannot be NULL "</span></div> +<div class="line"><a name="l02541"></a><span class="lineno"> 2541</span>  <span class="stringliteral">"when Peephole is enabled and CIFG disabled."</span>);</div> +<div class="line"><a name="l02542"></a><span class="lineno"> 2542</span>  }</div> +<div class="line"><a name="l02543"></a><span class="lineno"> 2543</span>  </div> +<div class="line"><a name="l02544"></a><span class="lineno"> 2544</span>  layer->m_PeepholeParameters.m_CellToInputWeights =</div> +<div class="line"><a name="l02545"></a><span class="lineno"> 2545</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellToInputWeights));</div> <div class="line"><a name="l02546"></a><span class="lineno"> 2546</span>  }</div> <div class="line"><a name="l02547"></a><span class="lineno"> 2547</span>  </div> -<div class="line"><a name="l02548"></a><span class="lineno"> 2548</span>  layer->m_PeepholeParameters.m_CellToForgetWeights =</div> -<div class="line"><a name="l02549"></a><span class="lineno"> 2549</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellToForgetWeights));</div> -<div class="line"><a name="l02550"></a><span class="lineno"> 2550</span>  layer->m_PeepholeParameters.m_CellToOutputWeights =</div> -<div class="line"><a name="l02551"></a><span class="lineno"> 2551</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellToOutputWeights));</div> -<div class="line"><a name="l02552"></a><span class="lineno"> 2552</span>  }</div> -<div class="line"><a name="l02553"></a><span class="lineno"> 2553</span>  </div> -<div class="line"><a name="l02554"></a><span class="lineno"> 2554</span>  <span class="comment">//Lstm Layer Normalization params</span></div> -<div class="line"><a name="l02555"></a><span class="lineno"> 2555</span>  <span class="keywordflow">if</span>(descriptor.m_LayerNormEnabled)</div> -<div class="line"><a name="l02556"></a><span class="lineno"> 2556</span>  {</div> -<div class="line"><a name="l02557"></a><span class="lineno"> 2557</span>  <span class="keywordflow">if</span>(!descriptor.m_CifgEnabled)</div> -<div class="line"><a name="l02558"></a><span class="lineno"> 2558</span>  {</div> -<div class="line"><a name="l02559"></a><span class="lineno"> 2559</span>  <span class="keywordflow">if</span>(params.m_InputLayerNormWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02560"></a><span class="lineno"> 2560</span>  {</div> -<div class="line"><a name="l02561"></a><span class="lineno"> 2561</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Input layer normalization weights cannot be NULL "</span></div> -<div class="line"><a name="l02562"></a><span class="lineno"> 2562</span>  <span class="stringliteral">"when layer normalization is enabled and CIFG disabled."</span>);</div> -<div class="line"><a name="l02563"></a><span class="lineno"> 2563</span>  }</div> -<div class="line"><a name="l02564"></a><span class="lineno"> 2564</span>  layer->m_LayerNormParameters.m_InputLayerNormWeights =</div> -<div class="line"><a name="l02565"></a><span class="lineno"> 2565</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputLayerNormWeights));</div> -<div class="line"><a name="l02566"></a><span class="lineno"> 2566</span>  }</div> -<div class="line"><a name="l02567"></a><span class="lineno"> 2567</span>  </div> -<div class="line"><a name="l02568"></a><span class="lineno"> 2568</span>  <span class="keywordflow">if</span>(params.m_ForgetLayerNormWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02548"></a><span class="lineno"> 2548</span>  <span class="keywordflow">if</span>(params.m_CellToForgetWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02549"></a><span class="lineno"> 2549</span>  {</div> +<div class="line"><a name="l02550"></a><span class="lineno"> 2550</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Cell To Forget Weights cannot be NULL "</span></div> +<div class="line"><a name="l02551"></a><span class="lineno"> 2551</span>  <span class="stringliteral">"when Peephole is enabled."</span>);</div> +<div class="line"><a name="l02552"></a><span class="lineno"> 2552</span>  }</div> +<div class="line"><a name="l02553"></a><span class="lineno"> 2553</span>  <span class="keywordflow">if</span>(params.m_CellToOutputWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02554"></a><span class="lineno"> 2554</span>  {</div> +<div class="line"><a name="l02555"></a><span class="lineno"> 2555</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Cell To Output Weights cannot be NULL "</span></div> +<div class="line"><a name="l02556"></a><span class="lineno"> 2556</span>  <span class="stringliteral">"when Peephole is enabled."</span>);</div> +<div class="line"><a name="l02557"></a><span class="lineno"> 2557</span>  }</div> +<div class="line"><a name="l02558"></a><span class="lineno"> 2558</span>  </div> +<div class="line"><a name="l02559"></a><span class="lineno"> 2559</span>  layer->m_PeepholeParameters.m_CellToForgetWeights =</div> +<div class="line"><a name="l02560"></a><span class="lineno"> 2560</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellToForgetWeights));</div> +<div class="line"><a name="l02561"></a><span class="lineno"> 2561</span>  layer->m_PeepholeParameters.m_CellToOutputWeights =</div> +<div class="line"><a name="l02562"></a><span class="lineno"> 2562</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellToOutputWeights));</div> +<div class="line"><a name="l02563"></a><span class="lineno"> 2563</span>  }</div> +<div class="line"><a name="l02564"></a><span class="lineno"> 2564</span>  </div> +<div class="line"><a name="l02565"></a><span class="lineno"> 2565</span>  <span class="comment">//Lstm Layer Normalization params</span></div> +<div class="line"><a name="l02566"></a><span class="lineno"> 2566</span>  <span class="keywordflow">if</span>(descriptor.m_LayerNormEnabled)</div> +<div class="line"><a name="l02567"></a><span class="lineno"> 2567</span>  {</div> +<div class="line"><a name="l02568"></a><span class="lineno"> 2568</span>  <span class="keywordflow">if</span>(!descriptor.m_CifgEnabled)</div> <div class="line"><a name="l02569"></a><span class="lineno"> 2569</span>  {</div> -<div class="line"><a name="l02570"></a><span class="lineno"> 2570</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Forget layer normalization weights cannot be NULL "</span></div> -<div class="line"><a name="l02571"></a><span class="lineno"> 2571</span>  <span class="stringliteral">"when layer normalization is enabled."</span>);</div> -<div class="line"><a name="l02572"></a><span class="lineno"> 2572</span>  }</div> -<div class="line"><a name="l02573"></a><span class="lineno"> 2573</span>  <span class="keywordflow">if</span>(params.m_CellLayerNormWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02574"></a><span class="lineno"> 2574</span>  {</div> -<div class="line"><a name="l02575"></a><span class="lineno"> 2575</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Cell layer normalization weights cannot be NULL "</span></div> -<div class="line"><a name="l02576"></a><span class="lineno"> 2576</span>  <span class="stringliteral">"when layer normalization is enabled."</span>);</div> +<div class="line"><a name="l02570"></a><span class="lineno"> 2570</span>  <span class="keywordflow">if</span>(params.m_InputLayerNormWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02571"></a><span class="lineno"> 2571</span>  {</div> +<div class="line"><a name="l02572"></a><span class="lineno"> 2572</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Input layer normalization weights cannot be NULL "</span></div> +<div class="line"><a name="l02573"></a><span class="lineno"> 2573</span>  <span class="stringliteral">"when layer normalization is enabled and CIFG disabled."</span>);</div> +<div class="line"><a name="l02574"></a><span class="lineno"> 2574</span>  }</div> +<div class="line"><a name="l02575"></a><span class="lineno"> 2575</span>  layer->m_LayerNormParameters.m_InputLayerNormWeights =</div> +<div class="line"><a name="l02576"></a><span class="lineno"> 2576</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputLayerNormWeights));</div> <div class="line"><a name="l02577"></a><span class="lineno"> 2577</span>  }</div> -<div class="line"><a name="l02578"></a><span class="lineno"> 2578</span>  <span class="keywordflow">if</span>(params.m_OutputLayerNormWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02579"></a><span class="lineno"> 2579</span>  {</div> -<div class="line"><a name="l02580"></a><span class="lineno"> 2580</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Output layer normalization weights cannot be NULL "</span></div> -<div class="line"><a name="l02581"></a><span class="lineno"> 2581</span>  <span class="stringliteral">"when layer normalization is enabled."</span>);</div> -<div class="line"><a name="l02582"></a><span class="lineno"> 2582</span>  }</div> -<div class="line"><a name="l02583"></a><span class="lineno"> 2583</span>  layer->m_LayerNormParameters.m_ForgetLayerNormWeights =</div> -<div class="line"><a name="l02584"></a><span class="lineno"> 2584</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ForgetLayerNormWeights));</div> -<div class="line"><a name="l02585"></a><span class="lineno"> 2585</span>  layer->m_LayerNormParameters.m_CellLayerNormWeights =</div> -<div class="line"><a name="l02586"></a><span class="lineno"> 2586</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellLayerNormWeights));</div> -<div class="line"><a name="l02587"></a><span class="lineno"> 2587</span>  layer->m_LayerNormParameters.m_OutputLayerNormWeights =</div> -<div class="line"><a name="l02588"></a><span class="lineno"> 2588</span>  std::make_shared<ScopedTensorHandle>(*(params.m_OutputLayerNormWeights));</div> -<div class="line"><a name="l02589"></a><span class="lineno"> 2589</span>  }</div> -<div class="line"><a name="l02590"></a><span class="lineno"> 2590</span>  <span class="keywordflow">return</span> layer;</div> -<div class="line"><a name="l02591"></a><span class="lineno"> 2591</span> }</div> +<div class="line"><a name="l02578"></a><span class="lineno"> 2578</span>  </div> +<div class="line"><a name="l02579"></a><span class="lineno"> 2579</span>  <span class="keywordflow">if</span>(params.m_ForgetLayerNormWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02580"></a><span class="lineno"> 2580</span>  {</div> +<div class="line"><a name="l02581"></a><span class="lineno"> 2581</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Forget layer normalization weights cannot be NULL "</span></div> +<div class="line"><a name="l02582"></a><span class="lineno"> 2582</span>  <span class="stringliteral">"when layer normalization is enabled."</span>);</div> +<div class="line"><a name="l02583"></a><span class="lineno"> 2583</span>  }</div> +<div class="line"><a name="l02584"></a><span class="lineno"> 2584</span>  <span class="keywordflow">if</span>(params.m_CellLayerNormWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02585"></a><span class="lineno"> 2585</span>  {</div> +<div class="line"><a name="l02586"></a><span class="lineno"> 2586</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Cell layer normalization weights cannot be NULL "</span></div> +<div class="line"><a name="l02587"></a><span class="lineno"> 2587</span>  <span class="stringliteral">"when layer normalization is enabled."</span>);</div> +<div class="line"><a name="l02588"></a><span class="lineno"> 2588</span>  }</div> +<div class="line"><a name="l02589"></a><span class="lineno"> 2589</span>  <span class="keywordflow">if</span>(params.m_OutputLayerNormWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02590"></a><span class="lineno"> 2590</span>  {</div> +<div class="line"><a name="l02591"></a><span class="lineno"> 2591</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddLstmLayer: Output layer normalization weights cannot be NULL "</span></div> +<div class="line"><a name="l02592"></a><span class="lineno"> 2592</span>  <span class="stringliteral">"when layer normalization is enabled."</span>);</div> +<div class="line"><a name="l02593"></a><span class="lineno"> 2593</span>  }</div> +<div class="line"><a name="l02594"></a><span class="lineno"> 2594</span>  layer->m_LayerNormParameters.m_ForgetLayerNormWeights =</div> +<div class="line"><a name="l02595"></a><span class="lineno"> 2595</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ForgetLayerNormWeights));</div> +<div class="line"><a name="l02596"></a><span class="lineno"> 2596</span>  layer->m_LayerNormParameters.m_CellLayerNormWeights =</div> +<div class="line"><a name="l02597"></a><span class="lineno"> 2597</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellLayerNormWeights));</div> +<div class="line"><a name="l02598"></a><span class="lineno"> 2598</span>  layer->m_LayerNormParameters.m_OutputLayerNormWeights =</div> +<div class="line"><a name="l02599"></a><span class="lineno"> 2599</span>  std::make_shared<ScopedTensorHandle>(*(params.m_OutputLayerNormWeights));</div> +<div class="line"><a name="l02600"></a><span class="lineno"> 2600</span>  }</div> +<div class="line"><a name="l02601"></a><span class="lineno"> 2601</span>  <span class="keywordflow">return</span> layer;</div> +<div class="line"><a name="l02602"></a><span class="lineno"> 2602</span> }</div> </div><!-- fragment --> <p class="reference">References <a class="el" href="_lstm_layer_8hpp_source.html#l00020">LstmLayer::m_BasicParameters</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00053">LstmInputParams::m_CellBias</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00059">LstmInputParams::m_CellLayerNormWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00049">LstmInputParams::m_CellToForgetWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00048">LstmInputParams::m_CellToInputWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00050">LstmInputParams::m_CellToOutputWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00052">LstmInputParams::m_ForgetGateBias</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00058">LstmInputParams::m_ForgetLayerNormWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00051">LstmInputParams::m_InputGateBias</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00057">LstmInputParams::m_InputLayerNormWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00042">LstmInputParams::m_InputToCellWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00041">LstmInputParams::m_InputToForgetWeights</a>, <a class="el" href="_lstm_parameters_8hpp_source.html#l00057">LstmBasicParameters::m_InputToForgetWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00040">LstmInputParams::m_InputToInputWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00043">LstmInputParams::m_InputToOutputWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00054">LstmInputParams::m_OutputGateBias</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00060">LstmInputParams::m_OutputLayerNormWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00056">LstmInputParams::m_ProjectionBias</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00055">LstmInputParams::m_ProjectionWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00046">LstmInputParams::m_RecurrentToCellWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00045">LstmInputParams::m_RecurrentToForgetWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00044">LstmInputParams::m_RecurrentToInputWeights</a>, and <a class="el" href="_lstm_params_8hpp_source.html#l00047">LstmInputParams::m_RecurrentToOutputWeights</a>.</p> @@ -1594,10 +1594,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02339">2339</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02340"></a><span class="lineno"> 2340</span> {</div> -<div class="line"><a name="l02341"></a><span class="lineno"> 2341</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<MaximumLayer>(name);</div> -<div class="line"><a name="l02342"></a><span class="lineno"> 2342</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02350">2350</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02351"></a><span class="lineno"> 2351</span> {</div> +<div class="line"><a name="l02352"></a><span class="lineno"> 2352</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<MaximumLayer>(name);</div> +<div class="line"><a name="l02353"></a><span class="lineno"> 2353</span> }</div> </div><!-- fragment --> </div> </div> @@ -1627,10 +1627,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02603">2603</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02604"></a><span class="lineno"> 2604</span> {</div> -<div class="line"><a name="l02605"></a><span class="lineno"> 2605</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<MeanLayer>(meanDescriptor,name);</div> -<div class="line"><a name="l02606"></a><span class="lineno"> 2606</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02614">2614</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02615"></a><span class="lineno"> 2615</span> {</div> +<div class="line"><a name="l02616"></a><span class="lineno"> 2616</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<MeanLayer>(meanDescriptor,name);</div> +<div class="line"><a name="l02617"></a><span class="lineno"> 2617</span> }</div> </div><!-- fragment --> </div> </div> @@ -1650,10 +1650,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02640">2640</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02641"></a><span class="lineno"> 2641</span> {</div> -<div class="line"><a name="l02642"></a><span class="lineno"> 2642</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<MergeLayer>(name);</div> -<div class="line"><a name="l02643"></a><span class="lineno"> 2643</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02651">2651</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02652"></a><span class="lineno"> 2652</span> {</div> +<div class="line"><a name="l02653"></a><span class="lineno"> 2653</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<MergeLayer>(name);</div> +<div class="line"><a name="l02654"></a><span class="lineno"> 2654</span> }</div> </div><!-- fragment --> </div> </div> @@ -1673,10 +1673,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02344">2344</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02345"></a><span class="lineno"> 2345</span> {</div> -<div class="line"><a name="l02346"></a><span class="lineno"> 2346</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<MinimumLayer>(name);</div> -<div class="line"><a name="l02347"></a><span class="lineno"> 2347</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02355">2355</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02356"></a><span class="lineno"> 2356</span> {</div> +<div class="line"><a name="l02357"></a><span class="lineno"> 2357</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<MinimumLayer>(name);</div> +<div class="line"><a name="l02358"></a><span class="lineno"> 2358</span> }</div> </div><!-- fragment --> </div> </div> @@ -1696,10 +1696,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02354">2354</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02355"></a><span class="lineno"> 2355</span> {</div> -<div class="line"><a name="l02356"></a><span class="lineno"> 2356</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<MultiplicationLayer>(name);</div> -<div class="line"><a name="l02357"></a><span class="lineno"> 2357</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02365">2365</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02366"></a><span class="lineno"> 2366</span> {</div> +<div class="line"><a name="l02367"></a><span class="lineno"> 2367</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<MultiplicationLayer>(name);</div> +<div class="line"><a name="l02368"></a><span class="lineno"> 2368</span> }</div> </div><!-- fragment --> </div> </div> @@ -1729,10 +1729,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02315">2315</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02318"></a><span class="lineno"> 2318</span> {</div> -<div class="line"><a name="l02319"></a><span class="lineno"> 2319</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<NormalizationLayer>(normalizationDescriptor, name);</div> -<div class="line"><a name="l02320"></a><span class="lineno"> 2320</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02326">2326</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02329"></a><span class="lineno"> 2329</span> {</div> +<div class="line"><a name="l02330"></a><span class="lineno"> 2330</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<NormalizationLayer>(normalizationDescriptor, name);</div> +<div class="line"><a name="l02331"></a><span class="lineno"> 2331</span> }</div> </div><!-- fragment --> </div> </div> @@ -1762,10 +1762,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02359">2359</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02360"></a><span class="lineno"> 2360</span> {</div> -<div class="line"><a name="l02361"></a><span class="lineno"> 2361</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<OutputLayer>(id, name);</div> -<div class="line"><a name="l02362"></a><span class="lineno"> 2362</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02370">2370</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02371"></a><span class="lineno"> 2371</span> {</div> +<div class="line"><a name="l02372"></a><span class="lineno"> 2372</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<OutputLayer>(id, name);</div> +<div class="line"><a name="l02373"></a><span class="lineno"> 2373</span> }</div> </div><!-- fragment --> </div> </div> @@ -1795,10 +1795,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02608">2608</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02609"></a><span class="lineno"> 2609</span> {</div> -<div class="line"><a name="l02610"></a><span class="lineno"> 2610</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<PadLayer>(padDescriptor,name);</div> -<div class="line"><a name="l02611"></a><span class="lineno"> 2611</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02619">2619</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02620"></a><span class="lineno"> 2620</span> {</div> +<div class="line"><a name="l02621"></a><span class="lineno"> 2621</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<PadLayer>(padDescriptor,name);</div> +<div class="line"><a name="l02622"></a><span class="lineno"> 2622</span> }</div> </div><!-- fragment --> </div> </div> @@ -1828,10 +1828,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02285">2285</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02287"></a><span class="lineno"> 2287</span> {</div> -<div class="line"><a name="l02288"></a><span class="lineno"> 2288</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<PermuteLayer>(permuteDescriptor, name);</div> -<div class="line"><a name="l02289"></a><span class="lineno"> 2289</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02296">2296</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02298"></a><span class="lineno"> 2298</span> {</div> +<div class="line"><a name="l02299"></a><span class="lineno"> 2299</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<PermuteLayer>(permuteDescriptor, name);</div> +<div class="line"><a name="l02300"></a><span class="lineno"> 2300</span> }</div> </div><!-- fragment --> </div> </div> @@ -1861,10 +1861,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02291">2291</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02293"></a><span class="lineno"> 2293</span> {</div> -<div class="line"><a name="l02294"></a><span class="lineno"> 2294</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<Pooling2dLayer>(pooling2dDescriptor, name);</div> -<div class="line"><a name="l02295"></a><span class="lineno"> 2295</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02302">2302</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02304"></a><span class="lineno"> 2304</span> {</div> +<div class="line"><a name="l02305"></a><span class="lineno"> 2305</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<Pooling2dLayer>(pooling2dDescriptor, name);</div> +<div class="line"><a name="l02306"></a><span class="lineno"> 2306</span> }</div> </div><!-- fragment --> </div> </div> @@ -1894,10 +1894,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02297">2297</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02299"></a><span class="lineno"> 2299</span> {</div> -<div class="line"><a name="l02300"></a><span class="lineno"> 2300</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<Pooling3dLayer>(pooling3dDescriptor, name);</div> -<div class="line"><a name="l02301"></a><span class="lineno"> 2301</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02308">2308</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02310"></a><span class="lineno"> 2310</span> {</div> +<div class="line"><a name="l02311"></a><span class="lineno"> 2311</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<Pooling3dLayer>(pooling3dDescriptor, name);</div> +<div class="line"><a name="l02312"></a><span class="lineno"> 2312</span> }</div> </div><!-- fragment --> </div> </div> @@ -1939,35 +1939,35 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l03039">3039</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l03043"></a><span class="lineno"> 3043</span> {</div> -<div class="line"><a name="l03044"></a><span class="lineno"> 3044</span>  <span class="comment">// Method use is for backend users.</span></div> -<div class="line"><a name="l03045"></a><span class="lineno"> 3045</span>  PreCompiledLayer* layer;</div> -<div class="line"><a name="l03046"></a><span class="lineno"> 3046</span>  <span class="keywordflow">if</span> (name)</div> -<div class="line"><a name="l03047"></a><span class="lineno"> 3047</span>  {</div> -<div class="line"><a name="l03048"></a><span class="lineno"> 3048</span>  layer = m_Graph->AddLayer<PreCompiledLayer>(preCompiledDescriptor, name);</div> -<div class="line"><a name="l03049"></a><span class="lineno"> 3049</span>  }</div> -<div class="line"><a name="l03050"></a><span class="lineno"> 3050</span>  <span class="keywordflow">else</span></div> -<div class="line"><a name="l03051"></a><span class="lineno"> 3051</span>  {</div> -<div class="line"><a name="l03052"></a><span class="lineno"> 3052</span>  layer = m_Graph->AddLayer<PreCompiledLayer>(preCompiledDescriptor, <span class="stringliteral">"pre-compiled"</span>);</div> -<div class="line"><a name="l03053"></a><span class="lineno"> 3053</span>  }</div> -<div class="line"><a name="l03054"></a><span class="lineno"> 3054</span>  </div> -<div class="line"><a name="l03055"></a><span class="lineno"> 3055</span>  <span class="comment">// Assign the pre-compiled object to layer</span></div> -<div class="line"><a name="l03056"></a><span class="lineno"> 3056</span>  <span class="comment">// Pass only one compiled network, Arm NN does not handle multiple</span></div> -<div class="line"><a name="l03057"></a><span class="lineno"> 3057</span>  <span class="comment">// pre-compiled objects in a single pre-compiled layer currently</span></div> -<div class="line"><a name="l03058"></a><span class="lineno"> 3058</span>  layer->SetPreCompiledObject(std::move(compiledBlobPtr));</div> -<div class="line"><a name="l03059"></a><span class="lineno"> 3059</span>  </div> -<div class="line"><a name="l03060"></a><span class="lineno"> 3060</span>  <span class="keywordflow">if</span> (backend.has_value())</div> -<div class="line"><a name="l03061"></a><span class="lineno"> 3061</span>  {</div> -<div class="line"><a name="l03062"></a><span class="lineno"> 3062</span>  layer->SetBackendId(backend.value());</div> -<div class="line"><a name="l03063"></a><span class="lineno"> 3063</span>  }</div> -<div class="line"><a name="l03064"></a><span class="lineno"> 3064</span>  <span class="keywordflow">else</span> <span class="keywordflow">if</span> (layer->GetBackendHint().has_value())</div> -<div class="line"><a name="l03065"></a><span class="lineno"> 3065</span>  {</div> -<div class="line"><a name="l03066"></a><span class="lineno"> 3066</span>  layer->SetBackendId(layer->GetBackendHint().value());</div> -<div class="line"><a name="l03067"></a><span class="lineno"> 3067</span>  }</div> -<div class="line"><a name="l03068"></a><span class="lineno"> 3068</span>  </div> -<div class="line"><a name="l03069"></a><span class="lineno"> 3069</span>  <span class="keywordflow">return</span> layer;</div> -<div class="line"><a name="l03070"></a><span class="lineno"> 3070</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l03050">3050</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l03054"></a><span class="lineno"> 3054</span> {</div> +<div class="line"><a name="l03055"></a><span class="lineno"> 3055</span>  <span class="comment">// Method use is for backend users.</span></div> +<div class="line"><a name="l03056"></a><span class="lineno"> 3056</span>  PreCompiledLayer* layer;</div> +<div class="line"><a name="l03057"></a><span class="lineno"> 3057</span>  <span class="keywordflow">if</span> (name)</div> +<div class="line"><a name="l03058"></a><span class="lineno"> 3058</span>  {</div> +<div class="line"><a name="l03059"></a><span class="lineno"> 3059</span>  layer = m_Graph->AddLayer<PreCompiledLayer>(preCompiledDescriptor, name);</div> +<div class="line"><a name="l03060"></a><span class="lineno"> 3060</span>  }</div> +<div class="line"><a name="l03061"></a><span class="lineno"> 3061</span>  <span class="keywordflow">else</span></div> +<div class="line"><a name="l03062"></a><span class="lineno"> 3062</span>  {</div> +<div class="line"><a name="l03063"></a><span class="lineno"> 3063</span>  layer = m_Graph->AddLayer<PreCompiledLayer>(preCompiledDescriptor, <span class="stringliteral">"pre-compiled"</span>);</div> +<div class="line"><a name="l03064"></a><span class="lineno"> 3064</span>  }</div> +<div class="line"><a name="l03065"></a><span class="lineno"> 3065</span>  </div> +<div class="line"><a name="l03066"></a><span class="lineno"> 3066</span>  <span class="comment">// Assign the pre-compiled object to layer</span></div> +<div class="line"><a name="l03067"></a><span class="lineno"> 3067</span>  <span class="comment">// Pass only one compiled network, Arm NN does not handle multiple</span></div> +<div class="line"><a name="l03068"></a><span class="lineno"> 3068</span>  <span class="comment">// pre-compiled objects in a single pre-compiled layer currently</span></div> +<div class="line"><a name="l03069"></a><span class="lineno"> 3069</span>  layer->SetPreCompiledObject(std::move(compiledBlobPtr));</div> +<div class="line"><a name="l03070"></a><span class="lineno"> 3070</span>  </div> +<div class="line"><a name="l03071"></a><span class="lineno"> 3071</span>  <span class="keywordflow">if</span> (backend.has_value())</div> +<div class="line"><a name="l03072"></a><span class="lineno"> 3072</span>  {</div> +<div class="line"><a name="l03073"></a><span class="lineno"> 3073</span>  layer->SetBackendId(backend.value());</div> +<div class="line"><a name="l03074"></a><span class="lineno"> 3074</span>  }</div> +<div class="line"><a name="l03075"></a><span class="lineno"> 3075</span>  <span class="keywordflow">else</span> <span class="keywordflow">if</span> (layer->GetBackendHint().has_value())</div> +<div class="line"><a name="l03076"></a><span class="lineno"> 3076</span>  {</div> +<div class="line"><a name="l03077"></a><span class="lineno"> 3077</span>  layer->SetBackendId(layer->GetBackendHint().value());</div> +<div class="line"><a name="l03078"></a><span class="lineno"> 3078</span>  }</div> +<div class="line"><a name="l03079"></a><span class="lineno"> 3079</span>  </div> +<div class="line"><a name="l03080"></a><span class="lineno"> 3080</span>  <span class="keywordflow">return</span> layer;</div> +<div class="line"><a name="l03081"></a><span class="lineno"> 3081</span> }</div> </div><!-- fragment --> <p class="reference">References <a class="el" href="_layer_8hpp_source.html#l00355">Layer::GetBackendHint()</a>, <a class="el" href="_optional_8hpp_source.html#l00053">OptionalBase::has_value()</a>, <a class="el" href="_layer_8hpp_source.html#l00291">Layer::SetBackendId()</a>, <a class="el" href="_pre_compiled_layer_8cpp_source.html#l00047">PreCompiledLayer::SetPreCompiledObject()</a>, and <a class="el" href="_optional_8hpp_source.html#l00146">OptionalReferenceSwitch< std::is_reference< T >::value, T >::value()</a>.</p> @@ -1989,10 +1989,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02650">2650</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02651"></a><span class="lineno"> 2651</span> {</div> -<div class="line"><a name="l02652"></a><span class="lineno"> 2652</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<PreluLayer>(name);</div> -<div class="line"><a name="l02653"></a><span class="lineno"> 2653</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02661">2661</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02662"></a><span class="lineno"> 2662</span> {</div> +<div class="line"><a name="l02663"></a><span class="lineno"> 2663</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<PreluLayer>(name);</div> +<div class="line"><a name="l02664"></a><span class="lineno"> 2664</span> }</div> </div><!-- fragment --> </div> </div> @@ -2028,145 +2028,145 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02734">2734</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02737"></a><span class="lineno"> 2737</span> {</div> -<div class="line"><a name="l02738"></a><span class="lineno"> 2738</span>  <span class="keyword">const</span> <span class="keyword">auto</span> layer = m_Graph->AddLayer<QLstmLayer>(descriptor, name);</div> -<div class="line"><a name="l02739"></a><span class="lineno"> 2739</span>  </div> -<div class="line"><a name="l02740"></a><span class="lineno"> 2740</span>  <span class="comment">// QLstm Basic Parameters</span></div> -<div class="line"><a name="l02741"></a><span class="lineno"> 2741</span>  layer->m_BasicParameters.m_InputToForgetWeights =</div> -<div class="line"><a name="l02742"></a><span class="lineno"> 2742</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToForgetWeights));</div> -<div class="line"><a name="l02743"></a><span class="lineno"> 2743</span>  layer->m_BasicParameters.m_InputToCellWeights =</div> -<div class="line"><a name="l02744"></a><span class="lineno"> 2744</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToCellWeights));</div> -<div class="line"><a name="l02745"></a><span class="lineno"> 2745</span>  layer->m_BasicParameters.m_InputToOutputWeights =</div> -<div class="line"><a name="l02746"></a><span class="lineno"> 2746</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToOutputWeights));</div> -<div class="line"><a name="l02747"></a><span class="lineno"> 2747</span>  layer->m_BasicParameters.m_RecurrentToForgetWeights =</div> -<div class="line"><a name="l02748"></a><span class="lineno"> 2748</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToForgetWeights));</div> -<div class="line"><a name="l02749"></a><span class="lineno"> 2749</span>  layer->m_BasicParameters.m_RecurrentToCellWeights =</div> -<div class="line"><a name="l02750"></a><span class="lineno"> 2750</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToCellWeights));</div> -<div class="line"><a name="l02751"></a><span class="lineno"> 2751</span>  layer->m_BasicParameters.m_RecurrentToOutputWeights =</div> -<div class="line"><a name="l02752"></a><span class="lineno"> 2752</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToOutputWeights));</div> -<div class="line"><a name="l02753"></a><span class="lineno"> 2753</span>  layer->m_BasicParameters.m_ForgetGateBias =</div> -<div class="line"><a name="l02754"></a><span class="lineno"> 2754</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ForgetGateBias));</div> -<div class="line"><a name="l02755"></a><span class="lineno"> 2755</span>  layer->m_BasicParameters.m_CellBias =</div> -<div class="line"><a name="l02756"></a><span class="lineno"> 2756</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellBias));</div> -<div class="line"><a name="l02757"></a><span class="lineno"> 2757</span>  layer->m_BasicParameters.m_OutputGateBias =</div> -<div class="line"><a name="l02758"></a><span class="lineno"> 2758</span>  std::make_shared<ScopedTensorHandle>(*(params.m_OutputGateBias));</div> -<div class="line"><a name="l02759"></a><span class="lineno"> 2759</span>  </div> -<div class="line"><a name="l02760"></a><span class="lineno"> 2760</span>  <span class="comment">// QLstm Cifg parameters</span></div> -<div class="line"><a name="l02761"></a><span class="lineno"> 2761</span>  <span class="keywordflow">if</span>(!descriptor.m_CifgEnabled)</div> -<div class="line"><a name="l02762"></a><span class="lineno"> 2762</span>  {</div> -<div class="line"><a name="l02763"></a><span class="lineno"> 2763</span>  <span class="keywordflow">if</span>(params.m_InputToInputWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02764"></a><span class="lineno"> 2764</span>  {</div> -<div class="line"><a name="l02765"></a><span class="lineno"> 2765</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Input To Input Weights cannot be NULL"</span>);</div> -<div class="line"><a name="l02766"></a><span class="lineno"> 2766</span>  }</div> -<div class="line"><a name="l02767"></a><span class="lineno"> 2767</span>  </div> -<div class="line"><a name="l02768"></a><span class="lineno"> 2768</span>  <span class="keywordflow">if</span>(params.m_RecurrentToInputWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02769"></a><span class="lineno"> 2769</span>  {</div> -<div class="line"><a name="l02770"></a><span class="lineno"> 2770</span>  <span class="keywordflow">throw</span> InvalidArgumentException(</div> -<div class="line"><a name="l02771"></a><span class="lineno"> 2771</span>  <span class="stringliteral">"AddQLstmLayer: Recurrent To Input Weights cannot be NULL"</span>);</div> -<div class="line"><a name="l02772"></a><span class="lineno"> 2772</span>  }</div> -<div class="line"><a name="l02773"></a><span class="lineno"> 2773</span>  </div> -<div class="line"><a name="l02774"></a><span class="lineno"> 2774</span>  <span class="keywordflow">if</span>(params.m_InputGateBias == <span class="keyword">nullptr</span>)</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02745">2745</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02748"></a><span class="lineno"> 2748</span> {</div> +<div class="line"><a name="l02749"></a><span class="lineno"> 2749</span>  <span class="keyword">const</span> <span class="keyword">auto</span> layer = m_Graph->AddLayer<QLstmLayer>(descriptor, name);</div> +<div class="line"><a name="l02750"></a><span class="lineno"> 2750</span>  </div> +<div class="line"><a name="l02751"></a><span class="lineno"> 2751</span>  <span class="comment">// QLstm Basic Parameters</span></div> +<div class="line"><a name="l02752"></a><span class="lineno"> 2752</span>  layer->m_BasicParameters.m_InputToForgetWeights =</div> +<div class="line"><a name="l02753"></a><span class="lineno"> 2753</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToForgetWeights));</div> +<div class="line"><a name="l02754"></a><span class="lineno"> 2754</span>  layer->m_BasicParameters.m_InputToCellWeights =</div> +<div class="line"><a name="l02755"></a><span class="lineno"> 2755</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToCellWeights));</div> +<div class="line"><a name="l02756"></a><span class="lineno"> 2756</span>  layer->m_BasicParameters.m_InputToOutputWeights =</div> +<div class="line"><a name="l02757"></a><span class="lineno"> 2757</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToOutputWeights));</div> +<div class="line"><a name="l02758"></a><span class="lineno"> 2758</span>  layer->m_BasicParameters.m_RecurrentToForgetWeights =</div> +<div class="line"><a name="l02759"></a><span class="lineno"> 2759</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToForgetWeights));</div> +<div class="line"><a name="l02760"></a><span class="lineno"> 2760</span>  layer->m_BasicParameters.m_RecurrentToCellWeights =</div> +<div class="line"><a name="l02761"></a><span class="lineno"> 2761</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToCellWeights));</div> +<div class="line"><a name="l02762"></a><span class="lineno"> 2762</span>  layer->m_BasicParameters.m_RecurrentToOutputWeights =</div> +<div class="line"><a name="l02763"></a><span class="lineno"> 2763</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToOutputWeights));</div> +<div class="line"><a name="l02764"></a><span class="lineno"> 2764</span>  layer->m_BasicParameters.m_ForgetGateBias =</div> +<div class="line"><a name="l02765"></a><span class="lineno"> 2765</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ForgetGateBias));</div> +<div class="line"><a name="l02766"></a><span class="lineno"> 2766</span>  layer->m_BasicParameters.m_CellBias =</div> +<div class="line"><a name="l02767"></a><span class="lineno"> 2767</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellBias));</div> +<div class="line"><a name="l02768"></a><span class="lineno"> 2768</span>  layer->m_BasicParameters.m_OutputGateBias =</div> +<div class="line"><a name="l02769"></a><span class="lineno"> 2769</span>  std::make_shared<ScopedTensorHandle>(*(params.m_OutputGateBias));</div> +<div class="line"><a name="l02770"></a><span class="lineno"> 2770</span>  </div> +<div class="line"><a name="l02771"></a><span class="lineno"> 2771</span>  <span class="comment">// QLstm Cifg parameters</span></div> +<div class="line"><a name="l02772"></a><span class="lineno"> 2772</span>  <span class="keywordflow">if</span>(!descriptor.m_CifgEnabled)</div> +<div class="line"><a name="l02773"></a><span class="lineno"> 2773</span>  {</div> +<div class="line"><a name="l02774"></a><span class="lineno"> 2774</span>  <span class="keywordflow">if</span>(params.m_InputToInputWeights == <span class="keyword">nullptr</span>)</div> <div class="line"><a name="l02775"></a><span class="lineno"> 2775</span>  {</div> -<div class="line"><a name="l02776"></a><span class="lineno"> 2776</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Input Gate Bias cannot be NULL"</span>);</div> +<div class="line"><a name="l02776"></a><span class="lineno"> 2776</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Input To Input Weights cannot be NULL"</span>);</div> <div class="line"><a name="l02777"></a><span class="lineno"> 2777</span>  }</div> <div class="line"><a name="l02778"></a><span class="lineno"> 2778</span>  </div> -<div class="line"><a name="l02779"></a><span class="lineno"> 2779</span>  layer->m_CifgParameters.m_InputToInputWeights =</div> -<div class="line"><a name="l02780"></a><span class="lineno"> 2780</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToInputWeights));</div> -<div class="line"><a name="l02781"></a><span class="lineno"> 2781</span>  layer->m_CifgParameters.m_RecurrentToInputWeights =</div> -<div class="line"><a name="l02782"></a><span class="lineno"> 2782</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToInputWeights));</div> -<div class="line"><a name="l02783"></a><span class="lineno"> 2783</span>  layer->m_CifgParameters.m_InputGateBias =</div> -<div class="line"><a name="l02784"></a><span class="lineno"> 2784</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputGateBias));</div> -<div class="line"><a name="l02785"></a><span class="lineno"> 2785</span>  }</div> -<div class="line"><a name="l02786"></a><span class="lineno"> 2786</span>  </div> -<div class="line"><a name="l02787"></a><span class="lineno"> 2787</span>  <span class="comment">// QLstm Projection parameters</span></div> -<div class="line"><a name="l02788"></a><span class="lineno"> 2788</span>  <span class="keywordflow">if</span>(descriptor.m_ProjectionEnabled)</div> -<div class="line"><a name="l02789"></a><span class="lineno"> 2789</span>  {</div> -<div class="line"><a name="l02790"></a><span class="lineno"> 2790</span>  <span class="keywordflow">if</span>(params.m_ProjectionWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02791"></a><span class="lineno"> 2791</span>  {</div> -<div class="line"><a name="l02792"></a><span class="lineno"> 2792</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Projection Weights cannot be NULL"</span>);</div> -<div class="line"><a name="l02793"></a><span class="lineno"> 2793</span>  }</div> -<div class="line"><a name="l02794"></a><span class="lineno"> 2794</span>  </div> -<div class="line"><a name="l02795"></a><span class="lineno"> 2795</span>  layer->m_ProjectionParameters.m_ProjectionWeights =</div> -<div class="line"><a name="l02796"></a><span class="lineno"> 2796</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ProjectionWeights));</div> +<div class="line"><a name="l02779"></a><span class="lineno"> 2779</span>  <span class="keywordflow">if</span>(params.m_RecurrentToInputWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02780"></a><span class="lineno"> 2780</span>  {</div> +<div class="line"><a name="l02781"></a><span class="lineno"> 2781</span>  <span class="keywordflow">throw</span> InvalidArgumentException(</div> +<div class="line"><a name="l02782"></a><span class="lineno"> 2782</span>  <span class="stringliteral">"AddQLstmLayer: Recurrent To Input Weights cannot be NULL"</span>);</div> +<div class="line"><a name="l02783"></a><span class="lineno"> 2783</span>  }</div> +<div class="line"><a name="l02784"></a><span class="lineno"> 2784</span>  </div> +<div class="line"><a name="l02785"></a><span class="lineno"> 2785</span>  <span class="keywordflow">if</span>(params.m_InputGateBias == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02786"></a><span class="lineno"> 2786</span>  {</div> +<div class="line"><a name="l02787"></a><span class="lineno"> 2787</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Input Gate Bias cannot be NULL"</span>);</div> +<div class="line"><a name="l02788"></a><span class="lineno"> 2788</span>  }</div> +<div class="line"><a name="l02789"></a><span class="lineno"> 2789</span>  </div> +<div class="line"><a name="l02790"></a><span class="lineno"> 2790</span>  layer->m_CifgParameters.m_InputToInputWeights =</div> +<div class="line"><a name="l02791"></a><span class="lineno"> 2791</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToInputWeights));</div> +<div class="line"><a name="l02792"></a><span class="lineno"> 2792</span>  layer->m_CifgParameters.m_RecurrentToInputWeights =</div> +<div class="line"><a name="l02793"></a><span class="lineno"> 2793</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToInputWeights));</div> +<div class="line"><a name="l02794"></a><span class="lineno"> 2794</span>  layer->m_CifgParameters.m_InputGateBias =</div> +<div class="line"><a name="l02795"></a><span class="lineno"> 2795</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputGateBias));</div> +<div class="line"><a name="l02796"></a><span class="lineno"> 2796</span>  }</div> <div class="line"><a name="l02797"></a><span class="lineno"> 2797</span>  </div> -<div class="line"><a name="l02798"></a><span class="lineno"> 2798</span>  <span class="comment">// Projection bias is optional even if projection is enabled</span></div> -<div class="line"><a name="l02799"></a><span class="lineno"> 2799</span>  <span class="keywordflow">if</span>(params.m_ProjectionBias != <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02800"></a><span class="lineno"> 2800</span>  {</div> -<div class="line"><a name="l02801"></a><span class="lineno"> 2801</span>  layer->m_ProjectionParameters.m_ProjectionBias =</div> -<div class="line"><a name="l02802"></a><span class="lineno"> 2802</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ProjectionBias));</div> -<div class="line"><a name="l02803"></a><span class="lineno"> 2803</span>  }</div> -<div class="line"><a name="l02804"></a><span class="lineno"> 2804</span>  </div> -<div class="line"><a name="l02805"></a><span class="lineno"> 2805</span>  }</div> -<div class="line"><a name="l02806"></a><span class="lineno"> 2806</span>  </div> -<div class="line"><a name="l02807"></a><span class="lineno"> 2807</span>  <span class="comment">// QLstm Peephole params</span></div> -<div class="line"><a name="l02808"></a><span class="lineno"> 2808</span>  <span class="keywordflow">if</span>(descriptor.m_PeepholeEnabled)</div> -<div class="line"><a name="l02809"></a><span class="lineno"> 2809</span>  {</div> -<div class="line"><a name="l02810"></a><span class="lineno"> 2810</span>  <span class="keywordflow">if</span>(params.m_CellToForgetWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02798"></a><span class="lineno"> 2798</span>  <span class="comment">// QLstm Projection parameters</span></div> +<div class="line"><a name="l02799"></a><span class="lineno"> 2799</span>  <span class="keywordflow">if</span>(descriptor.m_ProjectionEnabled)</div> +<div class="line"><a name="l02800"></a><span class="lineno"> 2800</span>  {</div> +<div class="line"><a name="l02801"></a><span class="lineno"> 2801</span>  <span class="keywordflow">if</span>(params.m_ProjectionWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02802"></a><span class="lineno"> 2802</span>  {</div> +<div class="line"><a name="l02803"></a><span class="lineno"> 2803</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Projection Weights cannot be NULL"</span>);</div> +<div class="line"><a name="l02804"></a><span class="lineno"> 2804</span>  }</div> +<div class="line"><a name="l02805"></a><span class="lineno"> 2805</span>  </div> +<div class="line"><a name="l02806"></a><span class="lineno"> 2806</span>  layer->m_ProjectionParameters.m_ProjectionWeights =</div> +<div class="line"><a name="l02807"></a><span class="lineno"> 2807</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ProjectionWeights));</div> +<div class="line"><a name="l02808"></a><span class="lineno"> 2808</span>  </div> +<div class="line"><a name="l02809"></a><span class="lineno"> 2809</span>  <span class="comment">// Projection bias is optional even if projection is enabled</span></div> +<div class="line"><a name="l02810"></a><span class="lineno"> 2810</span>  <span class="keywordflow">if</span>(params.m_ProjectionBias != <span class="keyword">nullptr</span>)</div> <div class="line"><a name="l02811"></a><span class="lineno"> 2811</span>  {</div> -<div class="line"><a name="l02812"></a><span class="lineno"> 2812</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Cell To Forget Weights cannot be NULL"</span>);</div> -<div class="line"><a name="l02813"></a><span class="lineno"> 2813</span>  }</div> -<div class="line"><a name="l02814"></a><span class="lineno"> 2814</span>  </div> -<div class="line"><a name="l02815"></a><span class="lineno"> 2815</span>  <span class="keywordflow">if</span>(params.m_CellToOutputWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02816"></a><span class="lineno"> 2816</span>  {</div> -<div class="line"><a name="l02817"></a><span class="lineno"> 2817</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Cell To Output Weights cannot be NULL"</span>);</div> -<div class="line"><a name="l02818"></a><span class="lineno"> 2818</span>  }</div> -<div class="line"><a name="l02819"></a><span class="lineno"> 2819</span>  </div> -<div class="line"><a name="l02820"></a><span class="lineno"> 2820</span>  <span class="keywordflow">if</span>(!descriptor.m_CifgEnabled)</div> -<div class="line"><a name="l02821"></a><span class="lineno"> 2821</span>  {</div> -<div class="line"><a name="l02822"></a><span class="lineno"> 2822</span>  <span class="keywordflow">if</span>(params.m_CellToInputWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02823"></a><span class="lineno"> 2823</span>  {</div> -<div class="line"><a name="l02824"></a><span class="lineno"> 2824</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Cell To Input Weights cannot be NULL"</span>);</div> -<div class="line"><a name="l02825"></a><span class="lineno"> 2825</span>  }</div> -<div class="line"><a name="l02826"></a><span class="lineno"> 2826</span>  </div> -<div class="line"><a name="l02827"></a><span class="lineno"> 2827</span>  layer->m_PeepholeParameters.m_CellToInputWeights =</div> -<div class="line"><a name="l02828"></a><span class="lineno"> 2828</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellToInputWeights));</div> +<div class="line"><a name="l02812"></a><span class="lineno"> 2812</span>  layer->m_ProjectionParameters.m_ProjectionBias =</div> +<div class="line"><a name="l02813"></a><span class="lineno"> 2813</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ProjectionBias));</div> +<div class="line"><a name="l02814"></a><span class="lineno"> 2814</span>  }</div> +<div class="line"><a name="l02815"></a><span class="lineno"> 2815</span>  </div> +<div class="line"><a name="l02816"></a><span class="lineno"> 2816</span>  }</div> +<div class="line"><a name="l02817"></a><span class="lineno"> 2817</span>  </div> +<div class="line"><a name="l02818"></a><span class="lineno"> 2818</span>  <span class="comment">// QLstm Peephole params</span></div> +<div class="line"><a name="l02819"></a><span class="lineno"> 2819</span>  <span class="keywordflow">if</span>(descriptor.m_PeepholeEnabled)</div> +<div class="line"><a name="l02820"></a><span class="lineno"> 2820</span>  {</div> +<div class="line"><a name="l02821"></a><span class="lineno"> 2821</span>  <span class="keywordflow">if</span>(params.m_CellToForgetWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02822"></a><span class="lineno"> 2822</span>  {</div> +<div class="line"><a name="l02823"></a><span class="lineno"> 2823</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Cell To Forget Weights cannot be NULL"</span>);</div> +<div class="line"><a name="l02824"></a><span class="lineno"> 2824</span>  }</div> +<div class="line"><a name="l02825"></a><span class="lineno"> 2825</span>  </div> +<div class="line"><a name="l02826"></a><span class="lineno"> 2826</span>  <span class="keywordflow">if</span>(params.m_CellToOutputWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02827"></a><span class="lineno"> 2827</span>  {</div> +<div class="line"><a name="l02828"></a><span class="lineno"> 2828</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Cell To Output Weights cannot be NULL"</span>);</div> <div class="line"><a name="l02829"></a><span class="lineno"> 2829</span>  }</div> <div class="line"><a name="l02830"></a><span class="lineno"> 2830</span>  </div> -<div class="line"><a name="l02831"></a><span class="lineno"> 2831</span>  layer->m_PeepholeParameters.m_CellToForgetWeights =</div> -<div class="line"><a name="l02832"></a><span class="lineno"> 2832</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellToForgetWeights));</div> -<div class="line"><a name="l02833"></a><span class="lineno"> 2833</span>  layer->m_PeepholeParameters.m_CellToOutputWeights =</div> -<div class="line"><a name="l02834"></a><span class="lineno"> 2834</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellToOutputWeights));</div> -<div class="line"><a name="l02835"></a><span class="lineno"> 2835</span>  }</div> -<div class="line"><a name="l02836"></a><span class="lineno"> 2836</span>  </div> -<div class="line"><a name="l02837"></a><span class="lineno"> 2837</span>  <span class="comment">// QLstm Layer Normalization params</span></div> -<div class="line"><a name="l02838"></a><span class="lineno"> 2838</span>  <span class="keywordflow">if</span>(descriptor.m_LayerNormEnabled)</div> -<div class="line"><a name="l02839"></a><span class="lineno"> 2839</span>  {</div> -<div class="line"><a name="l02840"></a><span class="lineno"> 2840</span>  <span class="keywordflow">if</span>(params.m_ForgetLayerNormWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02841"></a><span class="lineno"> 2841</span>  {</div> -<div class="line"><a name="l02842"></a><span class="lineno"> 2842</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Forget layer normalization weights cannot be NULL"</span>);</div> -<div class="line"><a name="l02843"></a><span class="lineno"> 2843</span>  }</div> -<div class="line"><a name="l02844"></a><span class="lineno"> 2844</span>  </div> -<div class="line"><a name="l02845"></a><span class="lineno"> 2845</span>  <span class="keywordflow">if</span>(params.m_CellLayerNormWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02846"></a><span class="lineno"> 2846</span>  {</div> -<div class="line"><a name="l02847"></a><span class="lineno"> 2847</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Cell layer normalization weights cannot be NULL"</span>);</div> -<div class="line"><a name="l02848"></a><span class="lineno"> 2848</span>  }</div> -<div class="line"><a name="l02849"></a><span class="lineno"> 2849</span>  </div> -<div class="line"><a name="l02850"></a><span class="lineno"> 2850</span>  <span class="keywordflow">if</span>(params.m_OutputLayerNormWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02851"></a><span class="lineno"> 2851</span>  {</div> -<div class="line"><a name="l02852"></a><span class="lineno"> 2852</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Output layer normalization weights cannot be NULL"</span>);</div> -<div class="line"><a name="l02853"></a><span class="lineno"> 2853</span>  }</div> -<div class="line"><a name="l02854"></a><span class="lineno"> 2854</span>  </div> -<div class="line"><a name="l02855"></a><span class="lineno"> 2855</span>  <span class="keywordflow">if</span>(!descriptor.m_CifgEnabled)</div> -<div class="line"><a name="l02856"></a><span class="lineno"> 2856</span>  {</div> -<div class="line"><a name="l02857"></a><span class="lineno"> 2857</span>  <span class="keywordflow">if</span>(params.m_InputLayerNormWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02858"></a><span class="lineno"> 2858</span>  {</div> -<div class="line"><a name="l02859"></a><span class="lineno"> 2859</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Input layer normalization weights cannot be NULL"</span>);</div> -<div class="line"><a name="l02860"></a><span class="lineno"> 2860</span>  }</div> -<div class="line"><a name="l02861"></a><span class="lineno"> 2861</span>  </div> -<div class="line"><a name="l02862"></a><span class="lineno"> 2862</span>  layer->m_LayerNormParameters.m_InputLayerNormWeights =</div> -<div class="line"><a name="l02863"></a><span class="lineno"> 2863</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputLayerNormWeights));</div> +<div class="line"><a name="l02831"></a><span class="lineno"> 2831</span>  <span class="keywordflow">if</span>(!descriptor.m_CifgEnabled)</div> +<div class="line"><a name="l02832"></a><span class="lineno"> 2832</span>  {</div> +<div class="line"><a name="l02833"></a><span class="lineno"> 2833</span>  <span class="keywordflow">if</span>(params.m_CellToInputWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02834"></a><span class="lineno"> 2834</span>  {</div> +<div class="line"><a name="l02835"></a><span class="lineno"> 2835</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Cell To Input Weights cannot be NULL"</span>);</div> +<div class="line"><a name="l02836"></a><span class="lineno"> 2836</span>  }</div> +<div class="line"><a name="l02837"></a><span class="lineno"> 2837</span>  </div> +<div class="line"><a name="l02838"></a><span class="lineno"> 2838</span>  layer->m_PeepholeParameters.m_CellToInputWeights =</div> +<div class="line"><a name="l02839"></a><span class="lineno"> 2839</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellToInputWeights));</div> +<div class="line"><a name="l02840"></a><span class="lineno"> 2840</span>  }</div> +<div class="line"><a name="l02841"></a><span class="lineno"> 2841</span>  </div> +<div class="line"><a name="l02842"></a><span class="lineno"> 2842</span>  layer->m_PeepholeParameters.m_CellToForgetWeights =</div> +<div class="line"><a name="l02843"></a><span class="lineno"> 2843</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellToForgetWeights));</div> +<div class="line"><a name="l02844"></a><span class="lineno"> 2844</span>  layer->m_PeepholeParameters.m_CellToOutputWeights =</div> +<div class="line"><a name="l02845"></a><span class="lineno"> 2845</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellToOutputWeights));</div> +<div class="line"><a name="l02846"></a><span class="lineno"> 2846</span>  }</div> +<div class="line"><a name="l02847"></a><span class="lineno"> 2847</span>  </div> +<div class="line"><a name="l02848"></a><span class="lineno"> 2848</span>  <span class="comment">// QLstm Layer Normalization params</span></div> +<div class="line"><a name="l02849"></a><span class="lineno"> 2849</span>  <span class="keywordflow">if</span>(descriptor.m_LayerNormEnabled)</div> +<div class="line"><a name="l02850"></a><span class="lineno"> 2850</span>  {</div> +<div class="line"><a name="l02851"></a><span class="lineno"> 2851</span>  <span class="keywordflow">if</span>(params.m_ForgetLayerNormWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02852"></a><span class="lineno"> 2852</span>  {</div> +<div class="line"><a name="l02853"></a><span class="lineno"> 2853</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Forget layer normalization weights cannot be NULL"</span>);</div> +<div class="line"><a name="l02854"></a><span class="lineno"> 2854</span>  }</div> +<div class="line"><a name="l02855"></a><span class="lineno"> 2855</span>  </div> +<div class="line"><a name="l02856"></a><span class="lineno"> 2856</span>  <span class="keywordflow">if</span>(params.m_CellLayerNormWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02857"></a><span class="lineno"> 2857</span>  {</div> +<div class="line"><a name="l02858"></a><span class="lineno"> 2858</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Cell layer normalization weights cannot be NULL"</span>);</div> +<div class="line"><a name="l02859"></a><span class="lineno"> 2859</span>  }</div> +<div class="line"><a name="l02860"></a><span class="lineno"> 2860</span>  </div> +<div class="line"><a name="l02861"></a><span class="lineno"> 2861</span>  <span class="keywordflow">if</span>(params.m_OutputLayerNormWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02862"></a><span class="lineno"> 2862</span>  {</div> +<div class="line"><a name="l02863"></a><span class="lineno"> 2863</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Output layer normalization weights cannot be NULL"</span>);</div> <div class="line"><a name="l02864"></a><span class="lineno"> 2864</span>  }</div> <div class="line"><a name="l02865"></a><span class="lineno"> 2865</span>  </div> -<div class="line"><a name="l02866"></a><span class="lineno"> 2866</span>  layer->m_LayerNormParameters.m_ForgetLayerNormWeights =</div> -<div class="line"><a name="l02867"></a><span class="lineno"> 2867</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ForgetLayerNormWeights));</div> -<div class="line"><a name="l02868"></a><span class="lineno"> 2868</span>  layer->m_LayerNormParameters.m_CellLayerNormWeights =</div> -<div class="line"><a name="l02869"></a><span class="lineno"> 2869</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellLayerNormWeights));</div> -<div class="line"><a name="l02870"></a><span class="lineno"> 2870</span>  layer->m_LayerNormParameters.m_OutputLayerNormWeights =</div> -<div class="line"><a name="l02871"></a><span class="lineno"> 2871</span>  std::make_shared<ScopedTensorHandle>(*(params.m_OutputLayerNormWeights));</div> -<div class="line"><a name="l02872"></a><span class="lineno"> 2872</span>  }</div> -<div class="line"><a name="l02873"></a><span class="lineno"> 2873</span>  <span class="keywordflow">return</span> layer;</div> -<div class="line"><a name="l02874"></a><span class="lineno"> 2874</span> }</div> +<div class="line"><a name="l02866"></a><span class="lineno"> 2866</span>  <span class="keywordflow">if</span>(!descriptor.m_CifgEnabled)</div> +<div class="line"><a name="l02867"></a><span class="lineno"> 2867</span>  {</div> +<div class="line"><a name="l02868"></a><span class="lineno"> 2868</span>  <span class="keywordflow">if</span>(params.m_InputLayerNormWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02869"></a><span class="lineno"> 2869</span>  {</div> +<div class="line"><a name="l02870"></a><span class="lineno"> 2870</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddQLstmLayer: Input layer normalization weights cannot be NULL"</span>);</div> +<div class="line"><a name="l02871"></a><span class="lineno"> 2871</span>  }</div> +<div class="line"><a name="l02872"></a><span class="lineno"> 2872</span>  </div> +<div class="line"><a name="l02873"></a><span class="lineno"> 2873</span>  layer->m_LayerNormParameters.m_InputLayerNormWeights =</div> +<div class="line"><a name="l02874"></a><span class="lineno"> 2874</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputLayerNormWeights));</div> +<div class="line"><a name="l02875"></a><span class="lineno"> 2875</span>  }</div> +<div class="line"><a name="l02876"></a><span class="lineno"> 2876</span>  </div> +<div class="line"><a name="l02877"></a><span class="lineno"> 2877</span>  layer->m_LayerNormParameters.m_ForgetLayerNormWeights =</div> +<div class="line"><a name="l02878"></a><span class="lineno"> 2878</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ForgetLayerNormWeights));</div> +<div class="line"><a name="l02879"></a><span class="lineno"> 2879</span>  layer->m_LayerNormParameters.m_CellLayerNormWeights =</div> +<div class="line"><a name="l02880"></a><span class="lineno"> 2880</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellLayerNormWeights));</div> +<div class="line"><a name="l02881"></a><span class="lineno"> 2881</span>  layer->m_LayerNormParameters.m_OutputLayerNormWeights =</div> +<div class="line"><a name="l02882"></a><span class="lineno"> 2882</span>  std::make_shared<ScopedTensorHandle>(*(params.m_OutputLayerNormWeights));</div> +<div class="line"><a name="l02883"></a><span class="lineno"> 2883</span>  }</div> +<div class="line"><a name="l02884"></a><span class="lineno"> 2884</span>  <span class="keywordflow">return</span> layer;</div> +<div class="line"><a name="l02885"></a><span class="lineno"> 2885</span> }</div> </div><!-- fragment --> <p class="reference">References <a class="el" href="_q_lstm_layer_8hpp_source.html#l00083">QLstmLayer::m_BasicParameters</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00053">LstmInputParams::m_CellBias</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00059">LstmInputParams::m_CellLayerNormWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00049">LstmInputParams::m_CellToForgetWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00048">LstmInputParams::m_CellToInputWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00050">LstmInputParams::m_CellToOutputWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00052">LstmInputParams::m_ForgetGateBias</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00058">LstmInputParams::m_ForgetLayerNormWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00051">LstmInputParams::m_InputGateBias</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00057">LstmInputParams::m_InputLayerNormWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00042">LstmInputParams::m_InputToCellWeights</a>, <a class="el" href="_q_lstm_layer_8hpp_source.html#l00017">QLstmBasicParameters::m_InputToForgetWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00041">LstmInputParams::m_InputToForgetWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00040">LstmInputParams::m_InputToInputWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00043">LstmInputParams::m_InputToOutputWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00054">LstmInputParams::m_OutputGateBias</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00060">LstmInputParams::m_OutputLayerNormWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00056">LstmInputParams::m_ProjectionBias</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00055">LstmInputParams::m_ProjectionWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00046">LstmInputParams::m_RecurrentToCellWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00045">LstmInputParams::m_RecurrentToForgetWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00044">LstmInputParams::m_RecurrentToInputWeights</a>, and <a class="el" href="_lstm_params_8hpp_source.html#l00047">LstmInputParams::m_RecurrentToOutputWeights</a>.</p> @@ -2198,42 +2198,42 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02696">2696</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02698"></a><span class="lineno"> 2698</span> {</div> -<div class="line"><a name="l02699"></a><span class="lineno"> 2699</span>  <span class="keyword">const</span> <span class="keyword">auto</span> layer = m_Graph->AddLayer<QuantizedLstmLayer>(name);</div> -<div class="line"><a name="l02700"></a><span class="lineno"> 2700</span>  </div> -<div class="line"><a name="l02701"></a><span class="lineno"> 2701</span>  <span class="comment">// InputToX weights</span></div> -<div class="line"><a name="l02702"></a><span class="lineno"> 2702</span>  layer->m_QuantizedLstmParameters.m_InputToInputWeights =</div> -<div class="line"><a name="l02703"></a><span class="lineno"> 2703</span>  std::make_shared<ScopedTensorHandle>(params.GetInputToInputWeights());</div> -<div class="line"><a name="l02704"></a><span class="lineno"> 2704</span>  layer->m_QuantizedLstmParameters.m_InputToForgetWeights =</div> -<div class="line"><a name="l02705"></a><span class="lineno"> 2705</span>  std::make_shared<ScopedTensorHandle>(params.GetInputToForgetWeights());</div> -<div class="line"><a name="l02706"></a><span class="lineno"> 2706</span>  layer->m_QuantizedLstmParameters.m_InputToCellWeights =</div> -<div class="line"><a name="l02707"></a><span class="lineno"> 2707</span>  std::make_shared<ScopedTensorHandle>(params.GetInputToCellWeights());</div> -<div class="line"><a name="l02708"></a><span class="lineno"> 2708</span>  layer->m_QuantizedLstmParameters.m_InputToOutputWeights =</div> -<div class="line"><a name="l02709"></a><span class="lineno"> 2709</span>  std::make_shared<ScopedTensorHandle>(params.GetInputToOutputWeights());</div> -<div class="line"><a name="l02710"></a><span class="lineno"> 2710</span>  </div> -<div class="line"><a name="l02711"></a><span class="lineno"> 2711</span>  <span class="comment">// RecurrentToX weights</span></div> -<div class="line"><a name="l02712"></a><span class="lineno"> 2712</span>  layer->m_QuantizedLstmParameters.m_RecurrentToInputWeights =</div> -<div class="line"><a name="l02713"></a><span class="lineno"> 2713</span>  std::make_shared<ScopedTensorHandle>(params.GetRecurrentToInputWeights());</div> -<div class="line"><a name="l02714"></a><span class="lineno"> 2714</span>  layer->m_QuantizedLstmParameters.m_RecurrentToForgetWeights =</div> -<div class="line"><a name="l02715"></a><span class="lineno"> 2715</span>  std::make_shared<ScopedTensorHandle>(params.GetRecurrentToForgetWeights());</div> -<div class="line"><a name="l02716"></a><span class="lineno"> 2716</span>  layer->m_QuantizedLstmParameters.m_RecurrentToCellWeights =</div> -<div class="line"><a name="l02717"></a><span class="lineno"> 2717</span>  std::make_shared<ScopedTensorHandle>(params.GetRecurrentToCellWeights());</div> -<div class="line"><a name="l02718"></a><span class="lineno"> 2718</span>  layer->m_QuantizedLstmParameters.m_RecurrentToOutputWeights =</div> -<div class="line"><a name="l02719"></a><span class="lineno"> 2719</span>  std::make_shared<ScopedTensorHandle>(params.GetRecurrentToOutputWeights());</div> -<div class="line"><a name="l02720"></a><span class="lineno"> 2720</span>  </div> -<div class="line"><a name="l02721"></a><span class="lineno"> 2721</span>  <span class="comment">// Bias</span></div> -<div class="line"><a name="l02722"></a><span class="lineno"> 2722</span>  layer->m_QuantizedLstmParameters.m_InputGateBias =</div> -<div class="line"><a name="l02723"></a><span class="lineno"> 2723</span>  std::make_shared<ScopedTensorHandle>(params.GetInputGateBias());</div> -<div class="line"><a name="l02724"></a><span class="lineno"> 2724</span>  layer->m_QuantizedLstmParameters.m_ForgetGateBias =</div> -<div class="line"><a name="l02725"></a><span class="lineno"> 2725</span>  std::make_shared<ScopedTensorHandle>(params.GetForgetGateBias());</div> -<div class="line"><a name="l02726"></a><span class="lineno"> 2726</span>  layer->m_QuantizedLstmParameters.m_CellBias =</div> -<div class="line"><a name="l02727"></a><span class="lineno"> 2727</span>  std::make_shared<ScopedTensorHandle>(params.GetCellBias());</div> -<div class="line"><a name="l02728"></a><span class="lineno"> 2728</span>  layer->m_QuantizedLstmParameters.m_OutputGateBias =</div> -<div class="line"><a name="l02729"></a><span class="lineno"> 2729</span>  std::make_shared<ScopedTensorHandle>(params.GetOutputGateBias());</div> -<div class="line"><a name="l02730"></a><span class="lineno"> 2730</span>  </div> -<div class="line"><a name="l02731"></a><span class="lineno"> 2731</span>  <span class="keywordflow">return</span> layer;</div> -<div class="line"><a name="l02732"></a><span class="lineno"> 2732</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02707">2707</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02709"></a><span class="lineno"> 2709</span> {</div> +<div class="line"><a name="l02710"></a><span class="lineno"> 2710</span>  <span class="keyword">const</span> <span class="keyword">auto</span> layer = m_Graph->AddLayer<QuantizedLstmLayer>(name);</div> +<div class="line"><a name="l02711"></a><span class="lineno"> 2711</span>  </div> +<div class="line"><a name="l02712"></a><span class="lineno"> 2712</span>  <span class="comment">// InputToX weights</span></div> +<div class="line"><a name="l02713"></a><span class="lineno"> 2713</span>  layer->m_QuantizedLstmParameters.m_InputToInputWeights =</div> +<div class="line"><a name="l02714"></a><span class="lineno"> 2714</span>  std::make_shared<ScopedTensorHandle>(params.GetInputToInputWeights());</div> +<div class="line"><a name="l02715"></a><span class="lineno"> 2715</span>  layer->m_QuantizedLstmParameters.m_InputToForgetWeights =</div> +<div class="line"><a name="l02716"></a><span class="lineno"> 2716</span>  std::make_shared<ScopedTensorHandle>(params.GetInputToForgetWeights());</div> +<div class="line"><a name="l02717"></a><span class="lineno"> 2717</span>  layer->m_QuantizedLstmParameters.m_InputToCellWeights =</div> +<div class="line"><a name="l02718"></a><span class="lineno"> 2718</span>  std::make_shared<ScopedTensorHandle>(params.GetInputToCellWeights());</div> +<div class="line"><a name="l02719"></a><span class="lineno"> 2719</span>  layer->m_QuantizedLstmParameters.m_InputToOutputWeights =</div> +<div class="line"><a name="l02720"></a><span class="lineno"> 2720</span>  std::make_shared<ScopedTensorHandle>(params.GetInputToOutputWeights());</div> +<div class="line"><a name="l02721"></a><span class="lineno"> 2721</span>  </div> +<div class="line"><a name="l02722"></a><span class="lineno"> 2722</span>  <span class="comment">// RecurrentToX weights</span></div> +<div class="line"><a name="l02723"></a><span class="lineno"> 2723</span>  layer->m_QuantizedLstmParameters.m_RecurrentToInputWeights =</div> +<div class="line"><a name="l02724"></a><span class="lineno"> 2724</span>  std::make_shared<ScopedTensorHandle>(params.GetRecurrentToInputWeights());</div> +<div class="line"><a name="l02725"></a><span class="lineno"> 2725</span>  layer->m_QuantizedLstmParameters.m_RecurrentToForgetWeights =</div> +<div class="line"><a name="l02726"></a><span class="lineno"> 2726</span>  std::make_shared<ScopedTensorHandle>(params.GetRecurrentToForgetWeights());</div> +<div class="line"><a name="l02727"></a><span class="lineno"> 2727</span>  layer->m_QuantizedLstmParameters.m_RecurrentToCellWeights =</div> +<div class="line"><a name="l02728"></a><span class="lineno"> 2728</span>  std::make_shared<ScopedTensorHandle>(params.GetRecurrentToCellWeights());</div> +<div class="line"><a name="l02729"></a><span class="lineno"> 2729</span>  layer->m_QuantizedLstmParameters.m_RecurrentToOutputWeights =</div> +<div class="line"><a name="l02730"></a><span class="lineno"> 2730</span>  std::make_shared<ScopedTensorHandle>(params.GetRecurrentToOutputWeights());</div> +<div class="line"><a name="l02731"></a><span class="lineno"> 2731</span>  </div> +<div class="line"><a name="l02732"></a><span class="lineno"> 2732</span>  <span class="comment">// Bias</span></div> +<div class="line"><a name="l02733"></a><span class="lineno"> 2733</span>  layer->m_QuantizedLstmParameters.m_InputGateBias =</div> +<div class="line"><a name="l02734"></a><span class="lineno"> 2734</span>  std::make_shared<ScopedTensorHandle>(params.GetInputGateBias());</div> +<div class="line"><a name="l02735"></a><span class="lineno"> 2735</span>  layer->m_QuantizedLstmParameters.m_ForgetGateBias =</div> +<div class="line"><a name="l02736"></a><span class="lineno"> 2736</span>  std::make_shared<ScopedTensorHandle>(params.GetForgetGateBias());</div> +<div class="line"><a name="l02737"></a><span class="lineno"> 2737</span>  layer->m_QuantizedLstmParameters.m_CellBias =</div> +<div class="line"><a name="l02738"></a><span class="lineno"> 2738</span>  std::make_shared<ScopedTensorHandle>(params.GetCellBias());</div> +<div class="line"><a name="l02739"></a><span class="lineno"> 2739</span>  layer->m_QuantizedLstmParameters.m_OutputGateBias =</div> +<div class="line"><a name="l02740"></a><span class="lineno"> 2740</span>  std::make_shared<ScopedTensorHandle>(params.GetOutputGateBias());</div> +<div class="line"><a name="l02741"></a><span class="lineno"> 2741</span>  </div> +<div class="line"><a name="l02742"></a><span class="lineno"> 2742</span>  <span class="keywordflow">return</span> layer;</div> +<div class="line"><a name="l02743"></a><span class="lineno"> 2743</span> }</div> </div><!-- fragment --> <p class="reference">References <a class="el" href="_quantized_lstm_params_8hpp_source.html#l00108">QuantizedLstmInputParams::GetCellBias()</a>, <a class="el" href="_quantized_lstm_params_8hpp_source.html#l00103">QuantizedLstmInputParams::GetForgetGateBias()</a>, <a class="el" href="_quantized_lstm_params_8hpp_source.html#l00098">QuantizedLstmInputParams::GetInputGateBias()</a>, <a class="el" href="_quantized_lstm_params_8hpp_source.html#l00068">QuantizedLstmInputParams::GetInputToCellWeights()</a>, <a class="el" href="_quantized_lstm_params_8hpp_source.html#l00063">QuantizedLstmInputParams::GetInputToForgetWeights()</a>, <a class="el" href="_quantized_lstm_params_8hpp_source.html#l00058">QuantizedLstmInputParams::GetInputToInputWeights()</a>, <a class="el" href="_quantized_lstm_params_8hpp_source.html#l00073">QuantizedLstmInputParams::GetInputToOutputWeights()</a>, <a class="el" href="_quantized_lstm_params_8hpp_source.html#l00113">QuantizedLstmInputParams::GetOutputGateBias()</a>, <a class="el" href="_quantized_lstm_params_8hpp_source.html#l00088">QuantizedLstmInputParams::GetRecurrentToCellWeights()</a>, <a class="el" href="_quantized_lstm_params_8hpp_source.html#l00083">QuantizedLstmInputParams::GetRecurrentToForgetWeights()</a>, <a class="el" href="_quantized_lstm_params_8hpp_source.html#l00078">QuantizedLstmInputParams::GetRecurrentToInputWeights()</a>, <a class="el" href="_quantized_lstm_params_8hpp_source.html#l00093">QuantizedLstmInputParams::GetRecurrentToOutputWeights()</a>, <a class="el" href="_quantized_lstm_layer_8hpp_source.html#l00017">QuantizedLstmParameters::m_InputToInputWeights</a>, and <a class="el" href="_quantized_lstm_layer_8hpp_source.html#l00049">QuantizedLstmLayer::m_QuantizedLstmParameters</a>.</p> @@ -2255,10 +2255,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02613">2613</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02614"></a><span class="lineno"> 2614</span> {</div> -<div class="line"><a name="l02615"></a><span class="lineno"> 2615</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<QuantizeLayer>(name);</div> -<div class="line"><a name="l02616"></a><span class="lineno"> 2616</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02624">2624</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02625"></a><span class="lineno"> 2625</span> {</div> +<div class="line"><a name="l02626"></a><span class="lineno"> 2626</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<QuantizeLayer>(name);</div> +<div class="line"><a name="l02627"></a><span class="lineno"> 2627</span> }</div> </div><!-- fragment --> </div> </div> @@ -2278,10 +2278,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02381">2381</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02382"></a><span class="lineno"> 2382</span> {</div> -<div class="line"><a name="l02383"></a><span class="lineno"> 2383</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<RankLayer>(name);</div> -<div class="line"><a name="l02384"></a><span class="lineno"> 2384</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02392">2392</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02393"></a><span class="lineno"> 2393</span> {</div> +<div class="line"><a name="l02394"></a><span class="lineno"> 2394</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<RankLayer>(name);</div> +<div class="line"><a name="l02395"></a><span class="lineno"> 2395</span> }</div> </div><!-- fragment --> </div> </div> @@ -2311,10 +2311,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02386">2386</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02388"></a><span class="lineno"> 2388</span> {</div> -<div class="line"><a name="l02389"></a><span class="lineno"> 2389</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ReduceLayer>(reduceDescriptor, name);</div> -<div class="line"><a name="l02390"></a><span class="lineno"> 2390</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02397">2397</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02399"></a><span class="lineno"> 2399</span> {</div> +<div class="line"><a name="l02400"></a><span class="lineno"> 2400</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ReduceLayer>(reduceDescriptor, name);</div> +<div class="line"><a name="l02401"></a><span class="lineno"> 2401</span> }</div> </div><!-- fragment --> </div> </div> @@ -2344,10 +2344,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02429">2429</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02431"></a><span class="lineno"> 2431</span> {</div> -<div class="line"><a name="l02432"></a><span class="lineno"> 2432</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ReshapeLayer>(reshapeDescriptor, name);</div> -<div class="line"><a name="l02433"></a><span class="lineno"> 2433</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02440">2440</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02442"></a><span class="lineno"> 2442</span> {</div> +<div class="line"><a name="l02443"></a><span class="lineno"> 2443</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ReshapeLayer>(reshapeDescriptor, name);</div> +<div class="line"><a name="l02444"></a><span class="lineno"> 2444</span> }</div> </div><!-- fragment --> </div> </div> @@ -2377,10 +2377,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02392">2392</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02393"></a><span class="lineno"> 2393</span> {</div> -<div class="line"><a name="l02394"></a><span class="lineno"> 2394</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ResizeLayer>(resizeDescriptor, name);</div> -<div class="line"><a name="l02395"></a><span class="lineno"> 2395</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02403">2403</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02404"></a><span class="lineno"> 2404</span> {</div> +<div class="line"><a name="l02405"></a><span class="lineno"> 2405</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ResizeLayer>(resizeDescriptor, name);</div> +<div class="line"><a name="l02406"></a><span class="lineno"> 2406</span> }</div> </div><!-- fragment --> </div> </div> @@ -2400,10 +2400,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l03029">3029</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l03030"></a><span class="lineno"> 3030</span> {</div> -<div class="line"><a name="l03031"></a><span class="lineno"> 3031</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ReverseV2Layer>(name);</div> -<div class="line"><a name="l03032"></a><span class="lineno"> 3032</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l03040">3040</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l03041"></a><span class="lineno"> 3041</span> {</div> +<div class="line"><a name="l03042"></a><span class="lineno"> 3042</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ReverseV2Layer>(name);</div> +<div class="line"><a name="l03043"></a><span class="lineno"> 3043</span> }</div> </div><!-- fragment --> </div> </div> @@ -2423,10 +2423,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02397">2397</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02398"></a><span class="lineno"> 2398</span> {</div> -<div class="line"><a name="l02399"></a><span class="lineno"> 2399</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ShapeLayer>(name);</div> -<div class="line"><a name="l02400"></a><span class="lineno"> 2400</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02408">2408</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02409"></a><span class="lineno"> 2409</span> {</div> +<div class="line"><a name="l02410"></a><span class="lineno"> 2410</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<ShapeLayer>(name);</div> +<div class="line"><a name="l02411"></a><span class="lineno"> 2411</span> }</div> </div><!-- fragment --> </div> </div> @@ -2456,10 +2456,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02322">2322</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02323"></a><span class="lineno"> 2323</span> {</div> -<div class="line"><a name="l02324"></a><span class="lineno"> 2324</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<SliceLayer>(sliceDescriptor, name);</div> -<div class="line"><a name="l02325"></a><span class="lineno"> 2325</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02333">2333</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02334"></a><span class="lineno"> 2334</span> {</div> +<div class="line"><a name="l02335"></a><span class="lineno"> 2335</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<SliceLayer>(sliceDescriptor, name);</div> +<div class="line"><a name="l02336"></a><span class="lineno"> 2336</span> }</div> </div><!-- fragment --> </div> </div> @@ -2489,10 +2489,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02327">2327</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02329"></a><span class="lineno"> 2329</span> {</div> -<div class="line"><a name="l02330"></a><span class="lineno"> 2330</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<SoftmaxLayer>(softmaxDescriptor, name);</div> -<div class="line"><a name="l02331"></a><span class="lineno"> 2331</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02338">2338</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02340"></a><span class="lineno"> 2340</span> {</div> +<div class="line"><a name="l02341"></a><span class="lineno"> 2341</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<SoftmaxLayer>(softmaxDescriptor, name);</div> +<div class="line"><a name="l02342"></a><span class="lineno"> 2342</span> }</div> </div><!-- fragment --> </div> </div> @@ -2522,10 +2522,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02435">2435</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02437"></a><span class="lineno"> 2437</span> {</div> -<div class="line"><a name="l02438"></a><span class="lineno"> 2438</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<SpaceToBatchNdLayer>(spaceToBatchNdDescriptor, name);</div> -<div class="line"><a name="l02439"></a><span class="lineno"> 2439</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02446">2446</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02448"></a><span class="lineno"> 2448</span> {</div> +<div class="line"><a name="l02449"></a><span class="lineno"> 2449</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<SpaceToBatchNdLayer>(spaceToBatchNdDescriptor, name);</div> +<div class="line"><a name="l02450"></a><span class="lineno"> 2450</span> }</div> </div><!-- fragment --> </div> </div> @@ -2555,10 +2555,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02441">2441</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02443"></a><span class="lineno"> 2443</span> {</div> -<div class="line"><a name="l02444"></a><span class="lineno"> 2444</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<SpaceToDepthLayer>(spaceToDepthDescriptor, name);</div> -<div class="line"><a name="l02445"></a><span class="lineno"> 2445</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02452">2452</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02454"></a><span class="lineno"> 2454</span> {</div> +<div class="line"><a name="l02455"></a><span class="lineno"> 2455</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<SpaceToDepthLayer>(spaceToDepthDescriptor, name);</div> +<div class="line"><a name="l02456"></a><span class="lineno"> 2456</span> }</div> </div><!-- fragment --> </div> </div> @@ -2588,10 +2588,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02333">2333</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02335"></a><span class="lineno"> 2335</span> {</div> -<div class="line"><a name="l02336"></a><span class="lineno"> 2336</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<SplitterLayer>(splitterDescriptor, name);</div> -<div class="line"><a name="l02337"></a><span class="lineno"> 2337</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02344">2344</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02346"></a><span class="lineno"> 2346</span> {</div> +<div class="line"><a name="l02347"></a><span class="lineno"> 2347</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<SplitterLayer>(splitterDescriptor, name);</div> +<div class="line"><a name="l02348"></a><span class="lineno"> 2348</span> }</div> </div><!-- fragment --> </div> </div> @@ -2621,10 +2621,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02683">2683</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02685"></a><span class="lineno"> 2685</span> {</div> -<div class="line"><a name="l02686"></a><span class="lineno"> 2686</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<StackLayer>(stackDescriptor, name);</div> -<div class="line"><a name="l02687"></a><span class="lineno"> 2687</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02694">2694</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02696"></a><span class="lineno"> 2696</span> {</div> +<div class="line"><a name="l02697"></a><span class="lineno"> 2697</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<StackLayer>(stackDescriptor, name);</div> +<div class="line"><a name="l02698"></a><span class="lineno"> 2698</span> }</div> </div><!-- fragment --> </div> </div> @@ -2654,10 +2654,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02690">2690</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02692"></a><span class="lineno"> 2692</span> {</div> -<div class="line"><a name="l02693"></a><span class="lineno"> 2693</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<StandInLayer>(desc, name);</div> -<div class="line"><a name="l02694"></a><span class="lineno"> 2694</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02701">2701</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02703"></a><span class="lineno"> 2703</span> {</div> +<div class="line"><a name="l02704"></a><span class="lineno"> 2704</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<StandInLayer>(desc, name);</div> +<div class="line"><a name="l02705"></a><span class="lineno"> 2705</span> }</div> </div><!-- fragment --> </div> </div> @@ -2687,10 +2687,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02623">2623</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02625"></a><span class="lineno"> 2625</span> {</div> -<div class="line"><a name="l02626"></a><span class="lineno"> 2626</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<StridedSliceLayer>(stridedSliceDescriptor, name);</div> -<div class="line"><a name="l02627"></a><span class="lineno"> 2627</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02634">2634</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02636"></a><span class="lineno"> 2636</span> {</div> +<div class="line"><a name="l02637"></a><span class="lineno"> 2637</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<StridedSliceLayer>(stridedSliceDescriptor, name);</div> +<div class="line"><a name="l02638"></a><span class="lineno"> 2638</span> }</div> </div><!-- fragment --> </div> </div> @@ -2710,10 +2710,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02598">2598</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02599"></a><span class="lineno"> 2599</span> {</div> -<div class="line"><a name="l02600"></a><span class="lineno"> 2600</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<SubtractionLayer>(name);</div> -<div class="line"><a name="l02601"></a><span class="lineno"> 2601</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02609">2609</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02610"></a><span class="lineno"> 2610</span> {</div> +<div class="line"><a name="l02611"></a><span class="lineno"> 2611</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<SubtractionLayer>(name);</div> +<div class="line"><a name="l02612"></a><span class="lineno"> 2612</span> }</div> </div><!-- fragment --> </div> </div> @@ -2733,10 +2733,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02645">2645</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02646"></a><span class="lineno"> 2646</span> {</div> -<div class="line"><a name="l02647"></a><span class="lineno"> 2647</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<SwitchLayer>(name);</div> -<div class="line"><a name="l02648"></a><span class="lineno"> 2648</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02656">2656</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02657"></a><span class="lineno"> 2657</span> {</div> +<div class="line"><a name="l02658"></a><span class="lineno"> 2658</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<SwitchLayer>(name);</div> +<div class="line"><a name="l02659"></a><span class="lineno"> 2659</span> }</div> </div><!-- fragment --> </div> </div> @@ -2766,10 +2766,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l03034">3034</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l03035"></a><span class="lineno"> 3035</span> {</div> -<div class="line"><a name="l03036"></a><span class="lineno"> 3036</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<TileLayer>(desc, name);</div> -<div class="line"><a name="l03037"></a><span class="lineno"> 3037</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l03045">3045</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l03046"></a><span class="lineno"> 3046</span> {</div> +<div class="line"><a name="l03047"></a><span class="lineno"> 3047</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<TileLayer>(desc, name);</div> +<div class="line"><a name="l03048"></a><span class="lineno"> 3048</span> }</div> </div><!-- fragment --> </div> </div> @@ -2811,24 +2811,24 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02655">2655</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02659"></a><span class="lineno"> 2659</span> {</div> -<div class="line"><a name="l02660"></a><span class="lineno"> 2660</span>  <span class="keywordflow">if</span> (descriptor.m_BiasEnabled && !biases.has_value())</div> -<div class="line"><a name="l02661"></a><span class="lineno"> 2661</span>  {</div> -<div class="line"><a name="l02662"></a><span class="lineno"> 2662</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddTransposeConvolution2dLayer: Biases cannot be empty"</span>);</div> -<div class="line"><a name="l02663"></a><span class="lineno"> 2663</span>  }</div> -<div class="line"><a name="l02664"></a><span class="lineno"> 2664</span>  </div> -<div class="line"><a name="l02665"></a><span class="lineno"> 2665</span>  <span class="keyword">const</span> <span class="keyword">auto</span> layer = m_Graph->AddLayer<TransposeConvolution2dLayer>(descriptor, name);</div> -<div class="line"><a name="l02666"></a><span class="lineno"> 2666</span>  </div> -<div class="line"><a name="l02667"></a><span class="lineno"> 2667</span>  layer->m_Weight = std::make_shared<ScopedTensorHandle>(weights);</div> -<div class="line"><a name="l02668"></a><span class="lineno"> 2668</span>  </div> -<div class="line"><a name="l02669"></a><span class="lineno"> 2669</span>  <span class="keywordflow">if</span> (descriptor.m_BiasEnabled)</div> -<div class="line"><a name="l02670"></a><span class="lineno"> 2670</span>  {</div> -<div class="line"><a name="l02671"></a><span class="lineno"> 2671</span>  layer->m_Bias = std::make_shared<ScopedTensorHandle>(biases.value());</div> -<div class="line"><a name="l02672"></a><span class="lineno"> 2672</span>  }</div> -<div class="line"><a name="l02673"></a><span class="lineno"> 2673</span>  </div> -<div class="line"><a name="l02674"></a><span class="lineno"> 2674</span>  <span class="keywordflow">return</span> layer;</div> -<div class="line"><a name="l02675"></a><span class="lineno"> 2675</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02666">2666</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02670"></a><span class="lineno"> 2670</span> {</div> +<div class="line"><a name="l02671"></a><span class="lineno"> 2671</span>  <span class="keywordflow">if</span> (descriptor.m_BiasEnabled && !biases.has_value())</div> +<div class="line"><a name="l02672"></a><span class="lineno"> 2672</span>  {</div> +<div class="line"><a name="l02673"></a><span class="lineno"> 2673</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddTransposeConvolution2dLayer: Biases cannot be empty"</span>);</div> +<div class="line"><a name="l02674"></a><span class="lineno"> 2674</span>  }</div> +<div class="line"><a name="l02675"></a><span class="lineno"> 2675</span>  </div> +<div class="line"><a name="l02676"></a><span class="lineno"> 2676</span>  <span class="keyword">const</span> <span class="keyword">auto</span> layer = m_Graph->AddLayer<TransposeConvolution2dLayer>(descriptor, name);</div> +<div class="line"><a name="l02677"></a><span class="lineno"> 2677</span>  </div> +<div class="line"><a name="l02678"></a><span class="lineno"> 2678</span>  layer->m_Weight = std::make_shared<ScopedTensorHandle>(weights);</div> +<div class="line"><a name="l02679"></a><span class="lineno"> 2679</span>  </div> +<div class="line"><a name="l02680"></a><span class="lineno"> 2680</span>  <span class="keywordflow">if</span> (descriptor.m_BiasEnabled)</div> +<div class="line"><a name="l02681"></a><span class="lineno"> 2681</span>  {</div> +<div class="line"><a name="l02682"></a><span class="lineno"> 2682</span>  layer->m_Bias = std::make_shared<ScopedTensorHandle>(biases.value());</div> +<div class="line"><a name="l02683"></a><span class="lineno"> 2683</span>  }</div> +<div class="line"><a name="l02684"></a><span class="lineno"> 2684</span>  </div> +<div class="line"><a name="l02685"></a><span class="lineno"> 2685</span>  <span class="keywordflow">return</span> layer;</div> +<div class="line"><a name="l02686"></a><span class="lineno"> 2686</span> }</div> </div><!-- fragment --> <p class="reference">References <a class="el" href="_optional_8hpp_source.html#l00053">OptionalBase::has_value()</a>, <a class="el" href="_descriptors_8hpp_source.html#l01481">TransposeConvolution2dDescriptor::m_BiasEnabled</a>, <a class="el" href="_transpose_convolution2d_layer_8hpp_source.html#l00019">TransposeConvolution2dLayer::m_Weight</a>, and <a class="el" href="_optional_8hpp_source.html#l00146">OptionalReferenceSwitch< std::is_reference< T >::value, T >::value()</a>.</p> @@ -2860,10 +2860,10 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02677">2677</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02679"></a><span class="lineno"> 2679</span> {</div> -<div class="line"><a name="l02680"></a><span class="lineno"> 2680</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<TransposeLayer>(transposeDescriptor, name);</div> -<div class="line"><a name="l02681"></a><span class="lineno"> 2681</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02688">2688</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02690"></a><span class="lineno"> 2690</span> {</div> +<div class="line"><a name="l02691"></a><span class="lineno"> 2691</span>  <span class="keywordflow">return</span> m_Graph->AddLayer<TransposeLayer>(transposeDescriptor, name);</div> +<div class="line"><a name="l02692"></a><span class="lineno"> 2692</span> }</div> </div><!-- fragment --> </div> </div> @@ -2899,144 +2899,144 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02882">2882</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02886"></a><span class="lineno"> 2886</span> {</div> -<div class="line"><a name="l02887"></a><span class="lineno"> 2887</span>  <span class="keyword">const</span> <span class="keyword">auto</span> layer = m_Graph->AddLayer<UnidirectionalSequenceLstmLayer>(descriptor, name);</div> -<div class="line"><a name="l02888"></a><span class="lineno"> 2888</span>  </div> -<div class="line"><a name="l02889"></a><span class="lineno"> 2889</span>  <span class="comment">//Lstm Basic Parameters</span></div> -<div class="line"><a name="l02890"></a><span class="lineno"> 2890</span>  layer->m_BasicParameters.m_InputToForgetWeights =</div> -<div class="line"><a name="l02891"></a><span class="lineno"> 2891</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToForgetWeights));</div> -<div class="line"><a name="l02892"></a><span class="lineno"> 2892</span>  layer->m_BasicParameters.m_InputToCellWeights =</div> -<div class="line"><a name="l02893"></a><span class="lineno"> 2893</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToCellWeights));</div> -<div class="line"><a name="l02894"></a><span class="lineno"> 2894</span>  layer->m_BasicParameters.m_InputToOutputWeights =</div> -<div class="line"><a name="l02895"></a><span class="lineno"> 2895</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToOutputWeights));</div> -<div class="line"><a name="l02896"></a><span class="lineno"> 2896</span>  layer->m_BasicParameters.m_RecurrentToForgetWeights =</div> -<div class="line"><a name="l02897"></a><span class="lineno"> 2897</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToForgetWeights));</div> -<div class="line"><a name="l02898"></a><span class="lineno"> 2898</span>  layer->m_BasicParameters.m_RecurrentToCellWeights =</div> -<div class="line"><a name="l02899"></a><span class="lineno"> 2899</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToCellWeights));</div> -<div class="line"><a name="l02900"></a><span class="lineno"> 2900</span>  layer->m_BasicParameters.m_RecurrentToOutputWeights =</div> -<div class="line"><a name="l02901"></a><span class="lineno"> 2901</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToOutputWeights));</div> -<div class="line"><a name="l02902"></a><span class="lineno"> 2902</span>  layer->m_BasicParameters.m_ForgetGateBias =</div> -<div class="line"><a name="l02903"></a><span class="lineno"> 2903</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ForgetGateBias));</div> -<div class="line"><a name="l02904"></a><span class="lineno"> 2904</span>  layer->m_BasicParameters.m_CellBias =</div> -<div class="line"><a name="l02905"></a><span class="lineno"> 2905</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellBias));</div> -<div class="line"><a name="l02906"></a><span class="lineno"> 2906</span>  layer->m_BasicParameters.m_OutputGateBias =</div> -<div class="line"><a name="l02907"></a><span class="lineno"> 2907</span>  std::make_shared<ScopedTensorHandle>(*(params.m_OutputGateBias));</div> -<div class="line"><a name="l02908"></a><span class="lineno"> 2908</span>  </div> -<div class="line"><a name="l02909"></a><span class="lineno"> 2909</span>  <span class="comment">//Lstm Cifg parameters</span></div> -<div class="line"><a name="l02910"></a><span class="lineno"> 2910</span>  <span class="keywordflow">if</span>(!descriptor.m_CifgEnabled)</div> -<div class="line"><a name="l02911"></a><span class="lineno"> 2911</span>  {</div> -<div class="line"><a name="l02912"></a><span class="lineno"> 2912</span>  <span class="keywordflow">if</span>(params.m_InputToInputWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02913"></a><span class="lineno"> 2913</span>  {</div> -<div class="line"><a name="l02914"></a><span class="lineno"> 2914</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Input To Input Weights cannot be NULL "</span></div> -<div class="line"><a name="l02915"></a><span class="lineno"> 2915</span>  <span class="stringliteral">"when CIFG is disabled."</span>);</div> -<div class="line"><a name="l02916"></a><span class="lineno"> 2916</span>  }</div> -<div class="line"><a name="l02917"></a><span class="lineno"> 2917</span>  <span class="keywordflow">if</span>(params.m_RecurrentToInputWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02918"></a><span class="lineno"> 2918</span>  {</div> -<div class="line"><a name="l02919"></a><span class="lineno"> 2919</span>  <span class="keywordflow">throw</span> InvalidArgumentException(</div> -<div class="line"><a name="l02920"></a><span class="lineno"> 2920</span>  <span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Recurrent To Input Weights cannot be NULL "</span></div> -<div class="line"><a name="l02921"></a><span class="lineno"> 2921</span>  <span class="stringliteral">"when CIFG is disabled."</span>);</div> -<div class="line"><a name="l02922"></a><span class="lineno"> 2922</span>  }</div> -<div class="line"><a name="l02923"></a><span class="lineno"> 2923</span>  <span class="keywordflow">if</span>(params.m_InputGateBias == <span class="keyword">nullptr</span>)</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02893">2893</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02897"></a><span class="lineno"> 2897</span> {</div> +<div class="line"><a name="l02898"></a><span class="lineno"> 2898</span>  <span class="keyword">const</span> <span class="keyword">auto</span> layer = m_Graph->AddLayer<UnidirectionalSequenceLstmLayer>(descriptor, name);</div> +<div class="line"><a name="l02899"></a><span class="lineno"> 2899</span>  </div> +<div class="line"><a name="l02900"></a><span class="lineno"> 2900</span>  <span class="comment">//Lstm Basic Parameters</span></div> +<div class="line"><a name="l02901"></a><span class="lineno"> 2901</span>  layer->m_BasicParameters.m_InputToForgetWeights =</div> +<div class="line"><a name="l02902"></a><span class="lineno"> 2902</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToForgetWeights));</div> +<div class="line"><a name="l02903"></a><span class="lineno"> 2903</span>  layer->m_BasicParameters.m_InputToCellWeights =</div> +<div class="line"><a name="l02904"></a><span class="lineno"> 2904</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToCellWeights));</div> +<div class="line"><a name="l02905"></a><span class="lineno"> 2905</span>  layer->m_BasicParameters.m_InputToOutputWeights =</div> +<div class="line"><a name="l02906"></a><span class="lineno"> 2906</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToOutputWeights));</div> +<div class="line"><a name="l02907"></a><span class="lineno"> 2907</span>  layer->m_BasicParameters.m_RecurrentToForgetWeights =</div> +<div class="line"><a name="l02908"></a><span class="lineno"> 2908</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToForgetWeights));</div> +<div class="line"><a name="l02909"></a><span class="lineno"> 2909</span>  layer->m_BasicParameters.m_RecurrentToCellWeights =</div> +<div class="line"><a name="l02910"></a><span class="lineno"> 2910</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToCellWeights));</div> +<div class="line"><a name="l02911"></a><span class="lineno"> 2911</span>  layer->m_BasicParameters.m_RecurrentToOutputWeights =</div> +<div class="line"><a name="l02912"></a><span class="lineno"> 2912</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToOutputWeights));</div> +<div class="line"><a name="l02913"></a><span class="lineno"> 2913</span>  layer->m_BasicParameters.m_ForgetGateBias =</div> +<div class="line"><a name="l02914"></a><span class="lineno"> 2914</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ForgetGateBias));</div> +<div class="line"><a name="l02915"></a><span class="lineno"> 2915</span>  layer->m_BasicParameters.m_CellBias =</div> +<div class="line"><a name="l02916"></a><span class="lineno"> 2916</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellBias));</div> +<div class="line"><a name="l02917"></a><span class="lineno"> 2917</span>  layer->m_BasicParameters.m_OutputGateBias =</div> +<div class="line"><a name="l02918"></a><span class="lineno"> 2918</span>  std::make_shared<ScopedTensorHandle>(*(params.m_OutputGateBias));</div> +<div class="line"><a name="l02919"></a><span class="lineno"> 2919</span>  </div> +<div class="line"><a name="l02920"></a><span class="lineno"> 2920</span>  <span class="comment">//Lstm Cifg parameters</span></div> +<div class="line"><a name="l02921"></a><span class="lineno"> 2921</span>  <span class="keywordflow">if</span>(!descriptor.m_CifgEnabled)</div> +<div class="line"><a name="l02922"></a><span class="lineno"> 2922</span>  {</div> +<div class="line"><a name="l02923"></a><span class="lineno"> 2923</span>  <span class="keywordflow">if</span>(params.m_InputToInputWeights == <span class="keyword">nullptr</span>)</div> <div class="line"><a name="l02924"></a><span class="lineno"> 2924</span>  {</div> -<div class="line"><a name="l02925"></a><span class="lineno"> 2925</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Input Gate Bias cannot be NULL "</span></div> +<div class="line"><a name="l02925"></a><span class="lineno"> 2925</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Input To Input Weights cannot be NULL "</span></div> <div class="line"><a name="l02926"></a><span class="lineno"> 2926</span>  <span class="stringliteral">"when CIFG is disabled."</span>);</div> <div class="line"><a name="l02927"></a><span class="lineno"> 2927</span>  }</div> -<div class="line"><a name="l02928"></a><span class="lineno"> 2928</span>  layer->m_CifgParameters.m_InputToInputWeights =</div> -<div class="line"><a name="l02929"></a><span class="lineno"> 2929</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToInputWeights));</div> -<div class="line"><a name="l02930"></a><span class="lineno"> 2930</span>  layer->m_CifgParameters.m_RecurrentToInputWeights =</div> -<div class="line"><a name="l02931"></a><span class="lineno"> 2931</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToInputWeights));</div> -<div class="line"><a name="l02932"></a><span class="lineno"> 2932</span>  layer->m_CifgParameters.m_InputGateBias =</div> -<div class="line"><a name="l02933"></a><span class="lineno"> 2933</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputGateBias));</div> -<div class="line"><a name="l02934"></a><span class="lineno"> 2934</span>  }</div> -<div class="line"><a name="l02935"></a><span class="lineno"> 2935</span>  </div> -<div class="line"><a name="l02936"></a><span class="lineno"> 2936</span>  <span class="comment">//Lstm projection parameters</span></div> -<div class="line"><a name="l02937"></a><span class="lineno"> 2937</span>  <span class="keywordflow">if</span>(descriptor.m_ProjectionEnabled)</div> -<div class="line"><a name="l02938"></a><span class="lineno"> 2938</span>  {</div> -<div class="line"><a name="l02939"></a><span class="lineno"> 2939</span>  <span class="keywordflow">if</span>(params.m_ProjectionWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02940"></a><span class="lineno"> 2940</span>  {</div> -<div class="line"><a name="l02941"></a><span class="lineno"> 2941</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Projection Weights cannot be NULL "</span></div> -<div class="line"><a name="l02942"></a><span class="lineno"> 2942</span>  <span class="stringliteral">"when projection is enabled."</span>);</div> -<div class="line"><a name="l02943"></a><span class="lineno"> 2943</span>  }</div> -<div class="line"><a name="l02944"></a><span class="lineno"> 2944</span>  layer->m_ProjectionParameters.m_ProjectionWeights =</div> -<div class="line"><a name="l02945"></a><span class="lineno"> 2945</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ProjectionWeights));</div> -<div class="line"><a name="l02946"></a><span class="lineno"> 2946</span>  <span class="keywordflow">if</span>(params.m_ProjectionBias != <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02947"></a><span class="lineno"> 2947</span>  {</div> -<div class="line"><a name="l02948"></a><span class="lineno"> 2948</span>  layer->m_ProjectionParameters.m_ProjectionBias =</div> -<div class="line"><a name="l02949"></a><span class="lineno"> 2949</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ProjectionBias));</div> -<div class="line"><a name="l02950"></a><span class="lineno"> 2950</span>  }</div> -<div class="line"><a name="l02951"></a><span class="lineno"> 2951</span>  }</div> -<div class="line"><a name="l02952"></a><span class="lineno"> 2952</span>  </div> -<div class="line"><a name="l02953"></a><span class="lineno"> 2953</span>  <span class="comment">//Lstm Peephole params</span></div> -<div class="line"><a name="l02954"></a><span class="lineno"> 2954</span>  <span class="keywordflow">if</span>(descriptor.m_PeepholeEnabled)</div> -<div class="line"><a name="l02955"></a><span class="lineno"> 2955</span>  {</div> -<div class="line"><a name="l02956"></a><span class="lineno"> 2956</span>  <span class="keywordflow">if</span>(!descriptor.m_CifgEnabled)</div> -<div class="line"><a name="l02957"></a><span class="lineno"> 2957</span>  {</div> -<div class="line"><a name="l02958"></a><span class="lineno"> 2958</span>  <span class="keywordflow">if</span>(params.m_CellToInputWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02959"></a><span class="lineno"> 2959</span>  {</div> -<div class="line"><a name="l02960"></a><span class="lineno"> 2960</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Cell To Input Weights "</span></div> -<div class="line"><a name="l02961"></a><span class="lineno"> 2961</span>  <span class="stringliteral">"cannot be NULL when Peephole is enabled and CIFG disabled."</span>);</div> -<div class="line"><a name="l02962"></a><span class="lineno"> 2962</span>  }</div> +<div class="line"><a name="l02928"></a><span class="lineno"> 2928</span>  <span class="keywordflow">if</span>(params.m_RecurrentToInputWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02929"></a><span class="lineno"> 2929</span>  {</div> +<div class="line"><a name="l02930"></a><span class="lineno"> 2930</span>  <span class="keywordflow">throw</span> InvalidArgumentException(</div> +<div class="line"><a name="l02931"></a><span class="lineno"> 2931</span>  <span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Recurrent To Input Weights cannot be NULL "</span></div> +<div class="line"><a name="l02932"></a><span class="lineno"> 2932</span>  <span class="stringliteral">"when CIFG is disabled."</span>);</div> +<div class="line"><a name="l02933"></a><span class="lineno"> 2933</span>  }</div> +<div class="line"><a name="l02934"></a><span class="lineno"> 2934</span>  <span class="keywordflow">if</span>(params.m_InputGateBias == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02935"></a><span class="lineno"> 2935</span>  {</div> +<div class="line"><a name="l02936"></a><span class="lineno"> 2936</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Input Gate Bias cannot be NULL "</span></div> +<div class="line"><a name="l02937"></a><span class="lineno"> 2937</span>  <span class="stringliteral">"when CIFG is disabled."</span>);</div> +<div class="line"><a name="l02938"></a><span class="lineno"> 2938</span>  }</div> +<div class="line"><a name="l02939"></a><span class="lineno"> 2939</span>  layer->m_CifgParameters.m_InputToInputWeights =</div> +<div class="line"><a name="l02940"></a><span class="lineno"> 2940</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputToInputWeights));</div> +<div class="line"><a name="l02941"></a><span class="lineno"> 2941</span>  layer->m_CifgParameters.m_RecurrentToInputWeights =</div> +<div class="line"><a name="l02942"></a><span class="lineno"> 2942</span>  std::make_shared<ScopedTensorHandle>(*(params.m_RecurrentToInputWeights));</div> +<div class="line"><a name="l02943"></a><span class="lineno"> 2943</span>  layer->m_CifgParameters.m_InputGateBias =</div> +<div class="line"><a name="l02944"></a><span class="lineno"> 2944</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputGateBias));</div> +<div class="line"><a name="l02945"></a><span class="lineno"> 2945</span>  }</div> +<div class="line"><a name="l02946"></a><span class="lineno"> 2946</span>  </div> +<div class="line"><a name="l02947"></a><span class="lineno"> 2947</span>  <span class="comment">//Lstm projection parameters</span></div> +<div class="line"><a name="l02948"></a><span class="lineno"> 2948</span>  <span class="keywordflow">if</span>(descriptor.m_ProjectionEnabled)</div> +<div class="line"><a name="l02949"></a><span class="lineno"> 2949</span>  {</div> +<div class="line"><a name="l02950"></a><span class="lineno"> 2950</span>  <span class="keywordflow">if</span>(params.m_ProjectionWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02951"></a><span class="lineno"> 2951</span>  {</div> +<div class="line"><a name="l02952"></a><span class="lineno"> 2952</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Projection Weights cannot be NULL "</span></div> +<div class="line"><a name="l02953"></a><span class="lineno"> 2953</span>  <span class="stringliteral">"when projection is enabled."</span>);</div> +<div class="line"><a name="l02954"></a><span class="lineno"> 2954</span>  }</div> +<div class="line"><a name="l02955"></a><span class="lineno"> 2955</span>  layer->m_ProjectionParameters.m_ProjectionWeights =</div> +<div class="line"><a name="l02956"></a><span class="lineno"> 2956</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ProjectionWeights));</div> +<div class="line"><a name="l02957"></a><span class="lineno"> 2957</span>  <span class="keywordflow">if</span>(params.m_ProjectionBias != <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02958"></a><span class="lineno"> 2958</span>  {</div> +<div class="line"><a name="l02959"></a><span class="lineno"> 2959</span>  layer->m_ProjectionParameters.m_ProjectionBias =</div> +<div class="line"><a name="l02960"></a><span class="lineno"> 2960</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ProjectionBias));</div> +<div class="line"><a name="l02961"></a><span class="lineno"> 2961</span>  }</div> +<div class="line"><a name="l02962"></a><span class="lineno"> 2962</span>  }</div> <div class="line"><a name="l02963"></a><span class="lineno"> 2963</span>  </div> -<div class="line"><a name="l02964"></a><span class="lineno"> 2964</span>  layer->m_PeepholeParameters.m_CellToInputWeights =</div> -<div class="line"><a name="l02965"></a><span class="lineno"> 2965</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellToInputWeights));</div> -<div class="line"><a name="l02966"></a><span class="lineno"> 2966</span>  }</div> -<div class="line"><a name="l02967"></a><span class="lineno"> 2967</span>  </div> -<div class="line"><a name="l02968"></a><span class="lineno"> 2968</span>  <span class="keywordflow">if</span>(params.m_CellToForgetWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02969"></a><span class="lineno"> 2969</span>  {</div> -<div class="line"><a name="l02970"></a><span class="lineno"> 2970</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Cell To Forget Weights cannot be NULL "</span></div> -<div class="line"><a name="l02971"></a><span class="lineno"> 2971</span>  <span class="stringliteral">"when Peephole is enabled."</span>);</div> -<div class="line"><a name="l02972"></a><span class="lineno"> 2972</span>  }</div> -<div class="line"><a name="l02973"></a><span class="lineno"> 2973</span>  <span class="keywordflow">if</span>(params.m_CellToOutputWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02974"></a><span class="lineno"> 2974</span>  {</div> -<div class="line"><a name="l02975"></a><span class="lineno"> 2975</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Cell To Output Weights cannot be NULL "</span></div> -<div class="line"><a name="l02976"></a><span class="lineno"> 2976</span>  <span class="stringliteral">"when Peephole is enabled."</span>);</div> +<div class="line"><a name="l02964"></a><span class="lineno"> 2964</span>  <span class="comment">//Lstm Peephole params</span></div> +<div class="line"><a name="l02965"></a><span class="lineno"> 2965</span>  <span class="keywordflow">if</span>(descriptor.m_PeepholeEnabled)</div> +<div class="line"><a name="l02966"></a><span class="lineno"> 2966</span>  {</div> +<div class="line"><a name="l02967"></a><span class="lineno"> 2967</span>  <span class="keywordflow">if</span>(!descriptor.m_CifgEnabled)</div> +<div class="line"><a name="l02968"></a><span class="lineno"> 2968</span>  {</div> +<div class="line"><a name="l02969"></a><span class="lineno"> 2969</span>  <span class="keywordflow">if</span>(params.m_CellToInputWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02970"></a><span class="lineno"> 2970</span>  {</div> +<div class="line"><a name="l02971"></a><span class="lineno"> 2971</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Cell To Input Weights "</span></div> +<div class="line"><a name="l02972"></a><span class="lineno"> 2972</span>  <span class="stringliteral">"cannot be NULL when Peephole is enabled and CIFG disabled."</span>);</div> +<div class="line"><a name="l02973"></a><span class="lineno"> 2973</span>  }</div> +<div class="line"><a name="l02974"></a><span class="lineno"> 2974</span>  </div> +<div class="line"><a name="l02975"></a><span class="lineno"> 2975</span>  layer->m_PeepholeParameters.m_CellToInputWeights =</div> +<div class="line"><a name="l02976"></a><span class="lineno"> 2976</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellToInputWeights));</div> <div class="line"><a name="l02977"></a><span class="lineno"> 2977</span>  }</div> <div class="line"><a name="l02978"></a><span class="lineno"> 2978</span>  </div> -<div class="line"><a name="l02979"></a><span class="lineno"> 2979</span>  layer->m_PeepholeParameters.m_CellToForgetWeights =</div> -<div class="line"><a name="l02980"></a><span class="lineno"> 2980</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellToForgetWeights));</div> -<div class="line"><a name="l02981"></a><span class="lineno"> 2981</span>  layer->m_PeepholeParameters.m_CellToOutputWeights =</div> -<div class="line"><a name="l02982"></a><span class="lineno"> 2982</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellToOutputWeights));</div> -<div class="line"><a name="l02983"></a><span class="lineno"> 2983</span>  }</div> -<div class="line"><a name="l02984"></a><span class="lineno"> 2984</span>  </div> -<div class="line"><a name="l02985"></a><span class="lineno"> 2985</span>  <span class="comment">//Lstm Layer Normalization params</span></div> -<div class="line"><a name="l02986"></a><span class="lineno"> 2986</span>  <span class="keywordflow">if</span>(descriptor.m_LayerNormEnabled)</div> -<div class="line"><a name="l02987"></a><span class="lineno"> 2987</span>  {</div> -<div class="line"><a name="l02988"></a><span class="lineno"> 2988</span>  <span class="keywordflow">if</span>(!descriptor.m_CifgEnabled)</div> -<div class="line"><a name="l02989"></a><span class="lineno"> 2989</span>  {</div> -<div class="line"><a name="l02990"></a><span class="lineno"> 2990</span>  <span class="keywordflow">if</span>(params.m_InputLayerNormWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l02991"></a><span class="lineno"> 2991</span>  {</div> -<div class="line"><a name="l02992"></a><span class="lineno"> 2992</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Input layer normalization weights "</span></div> -<div class="line"><a name="l02993"></a><span class="lineno"> 2993</span>  <span class="stringliteral">"cannot be NULL when layer normalization is enabled and CIFG disabled."</span>);</div> -<div class="line"><a name="l02994"></a><span class="lineno"> 2994</span>  }</div> -<div class="line"><a name="l02995"></a><span class="lineno"> 2995</span>  layer->m_LayerNormParameters.m_InputLayerNormWeights =</div> -<div class="line"><a name="l02996"></a><span class="lineno"> 2996</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputLayerNormWeights));</div> -<div class="line"><a name="l02997"></a><span class="lineno"> 2997</span>  }</div> -<div class="line"><a name="l02998"></a><span class="lineno"> 2998</span>  </div> -<div class="line"><a name="l02999"></a><span class="lineno"> 2999</span>  <span class="keywordflow">if</span>(params.m_ForgetLayerNormWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02979"></a><span class="lineno"> 2979</span>  <span class="keywordflow">if</span>(params.m_CellToForgetWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02980"></a><span class="lineno"> 2980</span>  {</div> +<div class="line"><a name="l02981"></a><span class="lineno"> 2981</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Cell To Forget Weights cannot be NULL "</span></div> +<div class="line"><a name="l02982"></a><span class="lineno"> 2982</span>  <span class="stringliteral">"when Peephole is enabled."</span>);</div> +<div class="line"><a name="l02983"></a><span class="lineno"> 2983</span>  }</div> +<div class="line"><a name="l02984"></a><span class="lineno"> 2984</span>  <span class="keywordflow">if</span>(params.m_CellToOutputWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l02985"></a><span class="lineno"> 2985</span>  {</div> +<div class="line"><a name="l02986"></a><span class="lineno"> 2986</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Cell To Output Weights cannot be NULL "</span></div> +<div class="line"><a name="l02987"></a><span class="lineno"> 2987</span>  <span class="stringliteral">"when Peephole is enabled."</span>);</div> +<div class="line"><a name="l02988"></a><span class="lineno"> 2988</span>  }</div> +<div class="line"><a name="l02989"></a><span class="lineno"> 2989</span>  </div> +<div class="line"><a name="l02990"></a><span class="lineno"> 2990</span>  layer->m_PeepholeParameters.m_CellToForgetWeights =</div> +<div class="line"><a name="l02991"></a><span class="lineno"> 2991</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellToForgetWeights));</div> +<div class="line"><a name="l02992"></a><span class="lineno"> 2992</span>  layer->m_PeepholeParameters.m_CellToOutputWeights =</div> +<div class="line"><a name="l02993"></a><span class="lineno"> 2993</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellToOutputWeights));</div> +<div class="line"><a name="l02994"></a><span class="lineno"> 2994</span>  }</div> +<div class="line"><a name="l02995"></a><span class="lineno"> 2995</span>  </div> +<div class="line"><a name="l02996"></a><span class="lineno"> 2996</span>  <span class="comment">//Lstm Layer Normalization params</span></div> +<div class="line"><a name="l02997"></a><span class="lineno"> 2997</span>  <span class="keywordflow">if</span>(descriptor.m_LayerNormEnabled)</div> +<div class="line"><a name="l02998"></a><span class="lineno"> 2998</span>  {</div> +<div class="line"><a name="l02999"></a><span class="lineno"> 2999</span>  <span class="keywordflow">if</span>(!descriptor.m_CifgEnabled)</div> <div class="line"><a name="l03000"></a><span class="lineno"> 3000</span>  {</div> -<div class="line"><a name="l03001"></a><span class="lineno"> 3001</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Forget layer normalization weights "</span></div> -<div class="line"><a name="l03002"></a><span class="lineno"> 3002</span>  <span class="stringliteral">"cannot be NULL when layer normalization is enabled."</span>);</div> -<div class="line"><a name="l03003"></a><span class="lineno"> 3003</span>  }</div> -<div class="line"><a name="l03004"></a><span class="lineno"> 3004</span>  <span class="keywordflow">if</span>(params.m_CellLayerNormWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l03005"></a><span class="lineno"> 3005</span>  {</div> -<div class="line"><a name="l03006"></a><span class="lineno"> 3006</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Cell layer normalization weights "</span></div> -<div class="line"><a name="l03007"></a><span class="lineno"> 3007</span>  <span class="stringliteral">"cannot be NULL when layer normalization is enabled."</span>);</div> +<div class="line"><a name="l03001"></a><span class="lineno"> 3001</span>  <span class="keywordflow">if</span>(params.m_InputLayerNormWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l03002"></a><span class="lineno"> 3002</span>  {</div> +<div class="line"><a name="l03003"></a><span class="lineno"> 3003</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Input layer normalization weights "</span></div> +<div class="line"><a name="l03004"></a><span class="lineno"> 3004</span>  <span class="stringliteral">"cannot be NULL when layer normalization is enabled and CIFG disabled."</span>);</div> +<div class="line"><a name="l03005"></a><span class="lineno"> 3005</span>  }</div> +<div class="line"><a name="l03006"></a><span class="lineno"> 3006</span>  layer->m_LayerNormParameters.m_InputLayerNormWeights =</div> +<div class="line"><a name="l03007"></a><span class="lineno"> 3007</span>  std::make_shared<ScopedTensorHandle>(*(params.m_InputLayerNormWeights));</div> <div class="line"><a name="l03008"></a><span class="lineno"> 3008</span>  }</div> -<div class="line"><a name="l03009"></a><span class="lineno"> 3009</span>  <span class="keywordflow">if</span>(params.m_OutputLayerNormWeights == <span class="keyword">nullptr</span>)</div> -<div class="line"><a name="l03010"></a><span class="lineno"> 3010</span>  {</div> -<div class="line"><a name="l03011"></a><span class="lineno"> 3011</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Output layer normalization weights "</span></div> -<div class="line"><a name="l03012"></a><span class="lineno"> 3012</span>  <span class="stringliteral">"cannot be NULL when layer normalization is enabled."</span>);</div> -<div class="line"><a name="l03013"></a><span class="lineno"> 3013</span>  }</div> -<div class="line"><a name="l03014"></a><span class="lineno"> 3014</span>  layer->m_LayerNormParameters.m_ForgetLayerNormWeights =</div> -<div class="line"><a name="l03015"></a><span class="lineno"> 3015</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ForgetLayerNormWeights));</div> -<div class="line"><a name="l03016"></a><span class="lineno"> 3016</span>  layer->m_LayerNormParameters.m_CellLayerNormWeights =</div> -<div class="line"><a name="l03017"></a><span class="lineno"> 3017</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellLayerNormWeights));</div> -<div class="line"><a name="l03018"></a><span class="lineno"> 3018</span>  layer->m_LayerNormParameters.m_OutputLayerNormWeights =</div> -<div class="line"><a name="l03019"></a><span class="lineno"> 3019</span>  std::make_shared<ScopedTensorHandle>(*(params.m_OutputLayerNormWeights));</div> -<div class="line"><a name="l03020"></a><span class="lineno"> 3020</span>  }</div> -<div class="line"><a name="l03021"></a><span class="lineno"> 3021</span>  <span class="keywordflow">return</span> layer;</div> -<div class="line"><a name="l03022"></a><span class="lineno"> 3022</span> }</div> +<div class="line"><a name="l03009"></a><span class="lineno"> 3009</span>  </div> +<div class="line"><a name="l03010"></a><span class="lineno"> 3010</span>  <span class="keywordflow">if</span>(params.m_ForgetLayerNormWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l03011"></a><span class="lineno"> 3011</span>  {</div> +<div class="line"><a name="l03012"></a><span class="lineno"> 3012</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Forget layer normalization weights "</span></div> +<div class="line"><a name="l03013"></a><span class="lineno"> 3013</span>  <span class="stringliteral">"cannot be NULL when layer normalization is enabled."</span>);</div> +<div class="line"><a name="l03014"></a><span class="lineno"> 3014</span>  }</div> +<div class="line"><a name="l03015"></a><span class="lineno"> 3015</span>  <span class="keywordflow">if</span>(params.m_CellLayerNormWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l03016"></a><span class="lineno"> 3016</span>  {</div> +<div class="line"><a name="l03017"></a><span class="lineno"> 3017</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Cell layer normalization weights "</span></div> +<div class="line"><a name="l03018"></a><span class="lineno"> 3018</span>  <span class="stringliteral">"cannot be NULL when layer normalization is enabled."</span>);</div> +<div class="line"><a name="l03019"></a><span class="lineno"> 3019</span>  }</div> +<div class="line"><a name="l03020"></a><span class="lineno"> 3020</span>  <span class="keywordflow">if</span>(params.m_OutputLayerNormWeights == <span class="keyword">nullptr</span>)</div> +<div class="line"><a name="l03021"></a><span class="lineno"> 3021</span>  {</div> +<div class="line"><a name="l03022"></a><span class="lineno"> 3022</span>  <span class="keywordflow">throw</span> InvalidArgumentException(<span class="stringliteral">"AddUnidirectionalSequenceLstmLayer: Output layer normalization weights "</span></div> +<div class="line"><a name="l03023"></a><span class="lineno"> 3023</span>  <span class="stringliteral">"cannot be NULL when layer normalization is enabled."</span>);</div> +<div class="line"><a name="l03024"></a><span class="lineno"> 3024</span>  }</div> +<div class="line"><a name="l03025"></a><span class="lineno"> 3025</span>  layer->m_LayerNormParameters.m_ForgetLayerNormWeights =</div> +<div class="line"><a name="l03026"></a><span class="lineno"> 3026</span>  std::make_shared<ScopedTensorHandle>(*(params.m_ForgetLayerNormWeights));</div> +<div class="line"><a name="l03027"></a><span class="lineno"> 3027</span>  layer->m_LayerNormParameters.m_CellLayerNormWeights =</div> +<div class="line"><a name="l03028"></a><span class="lineno"> 3028</span>  std::make_shared<ScopedTensorHandle>(*(params.m_CellLayerNormWeights));</div> +<div class="line"><a name="l03029"></a><span class="lineno"> 3029</span>  layer->m_LayerNormParameters.m_OutputLayerNormWeights =</div> +<div class="line"><a name="l03030"></a><span class="lineno"> 3030</span>  std::make_shared<ScopedTensorHandle>(*(params.m_OutputLayerNormWeights));</div> +<div class="line"><a name="l03031"></a><span class="lineno"> 3031</span>  }</div> +<div class="line"><a name="l03032"></a><span class="lineno"> 3032</span>  <span class="keywordflow">return</span> layer;</div> +<div class="line"><a name="l03033"></a><span class="lineno"> 3033</span> }</div> </div><!-- fragment --> <p class="reference">References <a class="el" href="_unidirectional_sequence_lstm_layer_8hpp_source.html#l00020">UnidirectionalSequenceLstmLayer::m_BasicParameters</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00053">LstmInputParams::m_CellBias</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00059">LstmInputParams::m_CellLayerNormWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00049">LstmInputParams::m_CellToForgetWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00048">LstmInputParams::m_CellToInputWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00050">LstmInputParams::m_CellToOutputWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00052">LstmInputParams::m_ForgetGateBias</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00058">LstmInputParams::m_ForgetLayerNormWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00051">LstmInputParams::m_InputGateBias</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00057">LstmInputParams::m_InputLayerNormWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00042">LstmInputParams::m_InputToCellWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00041">LstmInputParams::m_InputToForgetWeights</a>, <a class="el" href="_lstm_parameters_8hpp_source.html#l00057">LstmBasicParameters::m_InputToForgetWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00040">LstmInputParams::m_InputToInputWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00043">LstmInputParams::m_InputToOutputWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00054">LstmInputParams::m_OutputGateBias</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00060">LstmInputParams::m_OutputLayerNormWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00056">LstmInputParams::m_ProjectionBias</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00055">LstmInputParams::m_ProjectionWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00046">LstmInputParams::m_RecurrentToCellWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00045">LstmInputParams::m_RecurrentToForgetWeights</a>, <a class="el" href="_lstm_params_8hpp_source.html#l00044">LstmInputParams::m_RecurrentToInputWeights</a>, and <a class="el" href="_lstm_params_8hpp_source.html#l00047">LstmInputParams::m_RecurrentToOutputWeights</a>.</p> @@ -3058,13 +3058,13 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l03077">3077</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l03078"></a><span class="lineno"> 3078</span> {</div> -<div class="line"><a name="l03079"></a><span class="lineno"> 3079</span>  <span class="keywordflow">for</span> (<span class="keyword">auto</span> layer : <a class="code" href="classarmnn_1_1_network_impl.html#afe0a4f719f9752a405e71878da7012ba">GetGraph</a>())</div> -<div class="line"><a name="l03080"></a><span class="lineno"> 3080</span>  {</div> -<div class="line"><a name="l03081"></a><span class="lineno"> 3081</span>  layer->ExecuteStrategy(strategy);</div> -<div class="line"><a name="l03082"></a><span class="lineno"> 3082</span>  };</div> -<div class="line"><a name="l03083"></a><span class="lineno"> 3083</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l03088">3088</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l03089"></a><span class="lineno"> 3089</span> {</div> +<div class="line"><a name="l03090"></a><span class="lineno"> 3090</span>  <span class="keywordflow">for</span> (<span class="keyword">auto</span> layer : <a class="code" href="classarmnn_1_1_network_impl.html#afe0a4f719f9752a405e71878da7012ba">GetGraph</a>())</div> +<div class="line"><a name="l03091"></a><span class="lineno"> 3091</span>  {</div> +<div class="line"><a name="l03092"></a><span class="lineno"> 3092</span>  layer->ExecuteStrategy(strategy);</div> +<div class="line"><a name="l03093"></a><span class="lineno"> 3093</span>  };</div> +<div class="line"><a name="l03094"></a><span class="lineno"> 3094</span> }</div> </div><!-- fragment --> <p class="reference">References <a class="el" href="_network_8hpp_source.html#l00038">NetworkImpl::GetGraph()</a>.</p> @@ -3096,7 +3096,7 @@ Public Member Functions</h2></td></tr> <p class="definition">Definition at line <a class="el" href="_network_8hpp_source.html#l00038">38</a> of file <a class="el" href="_network_8hpp_source.html">Network.hpp</a>.</p> <div class="fragment"><div class="line"><a name="l00039"></a><span class="lineno"> 39</span>  { <span class="keywordflow">return</span> *m_Graph; }</div> </div><!-- fragment --> -<p class="reference">Referenced by <a class="el" href="_network_8cpp_source.html#l03077">NetworkImpl::ExecuteStrategy()</a>.</p> +<p class="reference">Referenced by <a class="el" href="_network_8cpp_source.html#l03088">NetworkImpl::ExecuteStrategy()</a>.</p> </div> </div> @@ -3115,11 +3115,11 @@ Public Member Functions</h2></td></tr> </table> </div><div class="memdoc"> -<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02171">2171</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> -<div class="fragment"><div class="line"><a name="l02172"></a><span class="lineno"> 2172</span> {</div> -<div class="line"><a name="l02173"></a><span class="lineno"> 2173</span>  m_Graph->Print();</div> -<div class="line"><a name="l02174"></a><span class="lineno"> 2174</span>  <span class="keywordflow">return</span> <a class="code" href="namespacearmnn.html#a67a0db04d321a74b7e7fcfd3f1a3f70ba505a83f220c02df2f85c3810cd9ceb38">Status::Success</a>;</div> -<div class="line"><a name="l02175"></a><span class="lineno"> 2175</span> }</div> +<p class="definition">Definition at line <a class="el" href="_network_8cpp_source.html#l02182">2182</a> of file <a class="el" href="_network_8cpp_source.html">Network.cpp</a>.</p> +<div class="fragment"><div class="line"><a name="l02183"></a><span class="lineno"> 2183</span> {</div> +<div class="line"><a name="l02184"></a><span class="lineno"> 2184</span>  m_Graph->Print();</div> +<div class="line"><a name="l02185"></a><span class="lineno"> 2185</span>  <span class="keywordflow">return</span> <a class="code" href="namespacearmnn.html#a67a0db04d321a74b7e7fcfd3f1a3f70ba505a83f220c02df2f85c3810cd9ceb38">Status::Success</a>;</div> +<div class="line"><a name="l02186"></a><span class="lineno"> 2186</span> }</div> </div><!-- fragment --> <p class="reference">References <a class="el" href="namespacearmnn.html#a67a0db04d321a74b7e7fcfd3f1a3f70ba505a83f220c02df2f85c3810cd9ceb38">armnn::Success</a>.</p> @@ -3137,7 +3137,7 @@ Public Member Functions</h2></td></tr> <div id="nav-path" class="navpath"><!-- id is needed for treeview function! --> <ul> <li class="navelem"><a class="el" href="namespacearmnn.html">armnn</a></li><li class="navelem"><a class="el" href="classarmnn_1_1_network_impl.html">NetworkImpl</a></li> - <li class="footer">Generated on Wed Nov 22 2023 10:57:15 for Arm NN by + <li class="footer">Generated on Wed Feb 14 2024 16:36:24 for Arm NN by <a href="http://www.doxygen.org/index.html"> <img class="footer" src="doxygen.png" alt="doxygen"/></a> 1.8.17 </li> </ul> |