From cb0630959aeae05bc2ae9f6d80cf5f5983a8fb77 Mon Sep 17 00:00:00 2001 From: Nikhil Raj Date: Wed, 23 Nov 2022 11:05:29 +0000 Subject: IVGCVSW-7075 Update Doxygen for 22.11 Release Signed-off-by: Nikhil Raj Change-Id: Ib5669e8fd3739d1b10f5dd694d020d51799896dc --- 22.11/_neon_quantize_workload_8cpp_source.xhtml | 133 ++++++++++++++++++++++++ 1 file changed, 133 insertions(+) create mode 100644 22.11/_neon_quantize_workload_8cpp_source.xhtml (limited to '22.11/_neon_quantize_workload_8cpp_source.xhtml') diff --git a/22.11/_neon_quantize_workload_8cpp_source.xhtml b/22.11/_neon_quantize_workload_8cpp_source.xhtml new file mode 100644 index 0000000000..16cdfc65a5 --- /dev/null +++ b/22.11/_neon_quantize_workload_8cpp_source.xhtml @@ -0,0 +1,133 @@ + + + + + + + + + + + + + +ArmNN: src/backends/neon/workloads/NeonQuantizeWorkload.cpp Source File + + + + + + + + + + + + + + + + +
+
+ + + + ArmNN + + + +
+
+  22.11 +
+
+
+ + + + + + + +
+
+ +
+
+
+ +
+ +
+
+ + +
+ +
+ +
+
+
NeonQuantizeWorkload.cpp
+
+
+Go to the documentation of this file.
1 //
2 // Copyright © 2017 Arm Ltd and Contributors. All rights reserved.
3 // SPDX-License-Identifier: MIT
4 //
5 
7 #include "NeonWorkloadUtils.hpp"
8 
12 #include <arm_compute/core/Types.h>
13 
14 namespace armnn
15 {
16 using namespace armcomputetensorutils;
17 
19 {
20  const arm_compute::TensorInfo neonInputInfo = armcomputetensorutils::BuildArmComputeTensorInfo(input);
21  const arm_compute::TensorInfo neonOutputInfo = armcomputetensorutils::BuildArmComputeTensorInfo(output);
22 
23  return arm_compute::NEQuantizationLayer::validate(&neonInputInfo, &neonOutputInfo);
24 }
25 
27  const WorkloadInfo& workloadInfo)
28  : NeonBaseWorkload<QuantizeQueueDescriptor>(descriptor, workloadInfo)
29 {
30  m_Data.ValidateInputsOutputs("NeonQuantizeWorkload", 1, 1);
31 
32  arm_compute::ITensor& input = PolymorphicPointerDowncast<IAclTensorHandle>(
33  m_Data.m_Inputs[0])->GetTensor();
34  arm_compute::ITensor& output = PolymorphicPointerDowncast<IAclTensorHandle>(
35  m_Data.m_Outputs[0])->GetTensor();
36 
37  m_Layer.reset(new arm_compute::NEQuantizationLayer());
38  m_Layer->configure(&input, &output);
39  m_Layer->prepare();
40 }
41 
43 {
44  if (m_Layer)
45  {
46  ARMNN_SCOPED_PROFILING_EVENT_NEON_GUID("NeonQuantizeWorkload_Execute", this->GetGuid());
47  m_Layer->run();
48  }
49 }
50 
51 } // namespace armnn
+ + +
arm::pipe::ProfilingGuid GetGuid() const final
Definition: Workload.hpp:61
+ +
void ValidateInputsOutputs(const std::string &descName, unsigned int numExpectedIn, unsigned int numExpectedOut) const
+
Copyright (c) 2021 ARM Limited and Contributors.
+
arm_compute::Status NeonQuantizeWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
+ + + + +
Status
enumeration
Definition: Types.hpp:42
+ +
std::vector< ITensorHandle * > m_Outputs
+
NeonQuantizeWorkload(const QuantizeQueueDescriptor &descriptor, const WorkloadInfo &workloadInfo)
+
Contains information about TensorInfos of a layer.
+
std::vector< ITensorHandle * > m_Inputs
+
#define ARMNN_SCOPED_PROFILING_EVENT_NEON_GUID(name, guid)
+ +
+
+ + + + -- cgit v1.2.1