1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
|
//
// Copyright © 2017 Arm Ltd. All rights reserved.
// See LICENSE file in the project root for full license information.
//
#pragma once
#include <armnn/ArmNN.hpp>
#include <armnn/Tensor.hpp>
#include <armnn/TypesUtils.hpp>
#include <backends/WorkloadInfo.hpp>
#include "test/TensorHelpers.hpp"
#include "QuantizeHelper.hpp"
#include "backends/CpuTensorHandle.hpp"
#include "backends/WorkloadFactory.hpp"
template<typename T>
LayerTestResult<T, 4> SimplePermuteTestImpl(
armnn::IWorkloadFactory& workloadFactory,
armnn::PermuteDescriptor descriptor,
armnn::TensorInfo inputTensorInfo,
armnn::TensorInfo outputTensorInfo,
const std::vector<T>& inputData,
const std::vector<T>& outputExpectedData)
{
auto input = MakeTensor<T, 4>(inputTensorInfo, inputData);
LayerTestResult<T, 4> ret(outputTensorInfo);
ret.outputExpected = MakeTensor<T, 4>(outputTensorInfo, outputExpectedData);
std::unique_ptr<armnn::ITensorHandle> inputHandle = workloadFactory.CreateTensorHandle(inputTensorInfo);
std::unique_ptr<armnn::ITensorHandle> outputHandle = workloadFactory.CreateTensorHandle(outputTensorInfo);
armnn::PermuteQueueDescriptor data;
data.m_Parameters = descriptor;
armnn::WorkloadInfo info;
AddInputToWorkload(data, info, inputTensorInfo, inputHandle.get());
AddOutputToWorkload(data, info, outputTensorInfo, outputHandle.get());
std::unique_ptr<armnn::IWorkload> workload = workloadFactory.CreatePermute(data, info);
inputHandle->Allocate();
outputHandle->Allocate();
CopyDataToITensorHandle(inputHandle.get(), &input[0][0][0][0]);
workload->Execute();
CopyDataFromITensorHandle(&ret.output[0][0][0][0], outputHandle.get());
return ret;
}
LayerTestResult<float, 4> SimplePermuteFloat32TestCommon(armnn::IWorkloadFactory& workloadFactory)
{
armnn::TensorInfo inputTensorInfo;
armnn::TensorInfo outputTensorInfo;
unsigned int inputShape[] = { 1, 2, 2, 2 };
unsigned int outputShape[] = { 1, 2, 2, 2 };
armnn::PermuteDescriptor descriptor;
descriptor.m_DimMappings = {0U, 3U, 1U, 2U};
inputTensorInfo = armnn::TensorInfo(4, inputShape, armnn::DataType::Float32);
outputTensorInfo = armnn::TensorInfo(4, outputShape, armnn::DataType::Float32);
std::vector<float> input = std::vector<float>(
{
1.0f, 2.0f,
3.0f, 4.0f,
5.0f, 6.0f,
7.0f, 8.0f
});
std::vector<float> outputExpected = std::vector<float>(
{
1.0f, 5.0f, 2.0f, 6.0f,
3.0f, 7.0f, 4.0f, 8.0f
});
return SimplePermuteTestImpl<float>(workloadFactory, descriptor, inputTensorInfo,
outputTensorInfo, input, outputExpected);
}
LayerTestResult<uint8_t, 4> SimplePermuteUint8TestCommon(armnn::IWorkloadFactory& workloadFactory)
{
armnn::TensorInfo inputTensorInfo;
armnn::TensorInfo outputTensorInfo;
unsigned int inputShape[] = { 1, 2, 2, 2 };
unsigned int outputShape[] = { 1, 2, 2, 2 };
armnn::PermuteDescriptor descriptor;
descriptor.m_DimMappings = {0U, 3U, 1U, 2U};
inputTensorInfo = armnn::TensorInfo(4, inputShape, armnn::DataType::QuantisedAsymm8);
inputTensorInfo.SetQuantizationScale(1.0f);
outputTensorInfo = armnn::TensorInfo(4, outputShape, armnn::DataType::QuantisedAsymm8);
outputTensorInfo.SetQuantizationScale(1.0f);
std::vector<uint8_t> input = std::vector<uint8_t>(
{
1, 2,
3, 4,
5, 6,
7, 8
});
std::vector<uint8_t> outputExpected = std::vector<uint8_t>(
{
1, 5, 2, 6,
3, 7, 4, 8
});
return SimplePermuteTestImpl<uint8_t>(workloadFactory, descriptor, inputTensorInfo,
outputTensorInfo, input, outputExpected);
}
|