From 8efb48a6847c5cd166c561127ae6611150963ce3 Mon Sep 17 00:00:00 2001 From: Nikhil Raj Date: Fri, 19 May 2023 11:14:28 +0100 Subject: Update Doxygen docu for 23.05 Signed-off-by: Nikhil Raj Change-Id: I0a992286f14fa68fcc6e5eba31ac39fed003cbbe --- 23.05/_ref_log_softmax_workload_8cpp_source.xhtml | 189 ++++++++++++++++++++++ 1 file changed, 189 insertions(+) create mode 100644 23.05/_ref_log_softmax_workload_8cpp_source.xhtml (limited to '23.05/_ref_log_softmax_workload_8cpp_source.xhtml') diff --git a/23.05/_ref_log_softmax_workload_8cpp_source.xhtml b/23.05/_ref_log_softmax_workload_8cpp_source.xhtml new file mode 100644 index 0000000000..a573b76daf --- /dev/null +++ b/23.05/_ref_log_softmax_workload_8cpp_source.xhtml @@ -0,0 +1,189 @@ + + + + + + + + + + + + + +ArmNN: src/backends/reference/workloads/RefLogSoftmaxWorkload.cpp Source File + + + + + + + + + + + + + + + + +
+
+ + + + ArmNN + + + +
+
+  23.05 +
+
+
+ + + + + + + +
+
+ +
+
+
+ +
+ +
+
+ + +
+ +
+ +
+
+
RefLogSoftmaxWorkload.cpp
+
+
+Go to the documentation of this file.
1 //
+
2 // Copyright © 2022 Arm Ltd and Contributors. All rights reserved.
+
3 // SPDX-License-Identifier: MIT
+
4 //
+
5 
+ +
7 
+
8 #include "Decoders.hpp"
+
9 #include "Encoders.hpp"
+
10 #include "LogSoftmax.hpp"
+
11 #include "RefWorkloadUtils.hpp"
+
12 
+
13 #include <Profiling.hpp>
+
14 
+
15 #include <armnn/utility/Assert.hpp>
+
16 
+
17 namespace armnn
+
18 {
+
19 
+ +
21 {
+ +
23 }
+
24 
+ +
26 {
+
27  WorkingMemDescriptor* workingMemDescriptor = static_cast<WorkingMemDescriptor*>(executionData.m_Data);
+
28  Execute(workingMemDescriptor->m_Inputs, workingMemDescriptor->m_Outputs);
+
29 }
+
30 
+
31 void RefLogSoftmaxWorkload::Execute(std::vector<ITensorHandle*> inputs, std::vector<ITensorHandle*> outputs) const
+
32 {
+
33  ARMNN_SCOPED_PROFILING_EVENT(Compute::CpuRef, "RefLogSoftmaxWorkload_Execute");
+
34 
+
35  const TensorInfo& inputInfo = GetTensorInfo(inputs[0]);
+
36  const TensorInfo& outputInfo = GetTensorInfo(outputs[0]);
+
37 
+
38  std::unique_ptr<Decoder<float>> decoder = MakeDecoder<float>(inputInfo, inputs[0]->Map());
+
39  std::unique_ptr<Encoder<float>> encoder = MakeEncoder<float>(outputInfo, outputs[0]->Map());
+
40 
+
41  ARMNN_ASSERT(decoder != nullptr);
+
42  ARMNN_ASSERT(encoder != nullptr);
+
43 
+
44  LogSoftmax(*decoder, *encoder, inputInfo, m_Data.m_Parameters);
+
45 }
+
46 
+
47 } // namespace armnn
+
+
+
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers
+ + + + + + + +
Copyright (c) 2021 ARM Limited and Contributors.
+
void ExecuteAsync(ExecutionData &executionData) override
+ +
void LogSoftmax(Decoder< float > &input, Encoder< float > &output, const TensorInfo &inputInfo, const LogSoftmaxDescriptor &descriptor)
Definition: LogSoftmax.cpp:29
+
#define ARMNN_SCOPED_PROFILING_EVENT(backendId, name)
Definition: Profiling.hpp:220
+ + + + + +
#define ARMNN_ASSERT(COND)
Definition: Assert.hpp:14
+ + +
std::vector< ITensorHandle * > m_Outputs
+ +
@ CpuRef
CPU Execution: Reference C++ kernels.
+
std::vector< ITensorHandle * > m_Inputs
+ + + + + -- cgit v1.2.1