From fd627ffaec8fd8801d980b4c91ee7c0607ab6aaf Mon Sep 17 00:00:00 2001 From: Jan Eilers Date: Thu, 25 Feb 2021 17:44:00 +0000 Subject: IVGCVSW-5687 Update Doxygen Docu * Update Doxygen Documentation for 21.02 release Signed-off-by: Jan Eilers Change-Id: I9ed2f9caab038836ea99d7b378d7899fe431a4e5 --- ...onvert_fp32_to_fp16_test_impl_8cpp_source.xhtml | 132 +++++++++++++++++++++ 1 file changed, 132 insertions(+) create mode 100644 21.02/_convert_fp32_to_fp16_test_impl_8cpp_source.xhtml (limited to '21.02/_convert_fp32_to_fp16_test_impl_8cpp_source.xhtml') diff --git a/21.02/_convert_fp32_to_fp16_test_impl_8cpp_source.xhtml b/21.02/_convert_fp32_to_fp16_test_impl_8cpp_source.xhtml new file mode 100644 index 0000000000..d79525f535 --- /dev/null +++ b/21.02/_convert_fp32_to_fp16_test_impl_8cpp_source.xhtml @@ -0,0 +1,132 @@ + + + + + + + + + + + + + +ArmNN: src/backends/backendsCommon/test/layerTests/ConvertFp32ToFp16TestImpl.cpp Source File + + + + + + + + + + + + + + + + +
+
+ + + + ArmNN + + + +
+
+  21.02 +
+
+
+ + + + + + + +
+
+ +
+
+
+ +
+ +
+
+ + +
+ +
+ +
+
+
ConvertFp32ToFp16TestImpl.cpp
+
+
+Go to the documentation of this file.
1 //
2 // Copyright © 2017 Arm Ltd and Contributors. All rights reserved.
3 // SPDX-License-Identifier: MIT
4 //
5 
7 
8 
11 
12 #include <test/TensorHelpers.hpp>
13 
15  armnn::IWorkloadFactory& workloadFactory,
17  const armnn::ITensorHandleFactory& tensorHandleFactory)
18 {
19  IgnoreUnused(memoryManager);
20  using namespace half_float::literal;
21 
22  const armnn::TensorInfo inputTensorInfo({1, 3, 2, 3}, armnn::DataType::Float32);
23  const armnn::TensorInfo outputTensorInfo({1, 3, 2, 3}, armnn::DataType::Float16);
24 
25  auto input = MakeTensor<float, 4>(inputTensorInfo,
26  { -37.5f, -15.2f, -8.76f, -2.0f, -1.5f, -1.3f, -0.5f, -0.4f, 0.0f,
27  1.0f, 0.4f, 0.5f, 1.3f, 1.5f, 2.0f, 8.76f, 15.2f, 37.5f });
28 
29  LayerTestResult<armnn::Half, 4> ret(outputTensorInfo);
30  ret.outputExpected = MakeTensor<armnn::Half, 4>(outputTensorInfo,
31  { -37.5_h, -15.2_h, -8.76_h, -2.0_h, -1.5_h, -1.3_h, -0.5_h, -0.4_h, 0.0_h,
32  1.0_h, 0.4_h, 0.5_h, 1.3_h, 1.5_h, 2.0_h, 8.76_h, 15.2_h, 37.5_h });
33 
34  std::unique_ptr<armnn::ITensorHandle> inputHandle = tensorHandleFactory.CreateTensorHandle(inputTensorInfo);
35  std::unique_ptr<armnn::ITensorHandle> outputHandle = tensorHandleFactory.CreateTensorHandle(outputTensorInfo);
36 
39  AddInputToWorkload(data, info, inputTensorInfo, inputHandle.get());
40  AddOutputToWorkload(data, info, outputTensorInfo, outputHandle.get());
41 
42  std::unique_ptr<armnn::IWorkload> workload = workloadFactory.CreateConvertFp32ToFp16(data, info);
43 
44  inputHandle->Allocate();
45  outputHandle->Allocate();
46 
47  CopyDataToITensorHandle(inputHandle.get(), &input[0][0][0][0]);
48 
49  workload->Execute();
50 
51  CopyDataFromITensorHandle(&ret.output[0][0][0][0], outputHandle.get());
52 
53  return ret;
54 }
LayerTestResult< armnn::Half, 4 > SimpleConvertFp32ToFp16Test(armnn::IWorkloadFactory &workloadFactory, const armnn::IBackendInternal::IMemoryManagerSharedPtr &memoryManager, const armnn::ITensorHandleFactory &tensorHandleFactory)
+ + + + +
void IgnoreUnused(Ts &&...)
+ +
std::shared_ptr< IMemoryManager > IMemoryManagerSharedPtr
+ + +
void CopyDataFromITensorHandle(void *memory, const armnn::ITensorHandle *tensorHandle)
+ +
virtual std::unique_ptr< IWorkload > CreateConvertFp32ToFp16(const ConvertFp32ToFp16QueueDescriptor &descriptor, const WorkloadInfo &info) const
+ + +
Contains information about inputs and outputs to a layer.
+ +
virtual std::unique_ptr< ITensorHandle > CreateTensorHandle(const TensorInfo &tensorInfo) const =0
+
void CopyDataToITensorHandle(armnn::ITensorHandle *tensorHandle, const void *memory)
+
+
+ + + + -- cgit v1.2.1