28 std::vector<float>& inputData,
29 std::vector<float>& outputExpectedData,
31 const float qScale = 1.0f,
32 const int32_t qOffset = 0)
42 std::vector<float> inputTmp(inputData.size());
44 inputData.data(), inputTmp.data(),
sizeof(float));
47 std::vector<float> outputTmp(outputExpectedData.size());
49 outputExpectedData.data(), outputTmp.data(),
sizeof(float));
50 outputExpectedData = outputTmp;
53 if(armnn::IsQuantizedType<T>())
61 boost::multi_array<T, 4> input = MakeTensor<T, 4>(inputTensorInfo,
62 armnnUtils::QuantizedVector<T>(inputData, qScale, qOffset));
65 ret.outputExpected = MakeTensor<T, 4>(outputTensorInfo,
66 armnnUtils::QuantizedVector<T>(outputExpectedData, qScale, qOffset));
68 std::unique_ptr<armnn::ITensorHandle> inputHandle = workloadFactory.
CreateTensorHandle(inputTensorInfo);
69 std::unique_ptr<armnn::ITensorHandle> outputHandle = workloadFactory.
CreateTensorHandle(outputTensorInfo);
72 AddInputToWorkload(descriptor, info, inputTensorInfo, inputHandle.get());
73 AddOutputToWorkload(descriptor, info, outputTensorInfo, outputHandle.get());
75 std::unique_ptr<armnn::IWorkload> workload = workloadFactory.
CreateSpaceToDepth(descriptor, info);
77 inputHandle->Allocate();
78 outputHandle->Allocate();
89 template<armnn::DataType ArmnnType,
typename T = armnn::ResolveType<ArmnnType>>
95 unsigned int inputShape[] = {1, 2, 2, 1};
96 unsigned int outputShape[] = {1, 1, 1, 4};
98 std::vector<float> input = std::vector<float>(
100 1.0f, 2.0f, 3.0f, 4.0f
103 std::vector<float> outputExpected = std::vector<float>(
105 1.0f, 2.0f, 3.0f, 4.0f
118 return SpaceToDepthTestImpl<T>(
119 workloadFactory, memoryManager, inputTensorInfo, outputTensorInfo, input, outputExpected, desc);
122 template<armnn::DataType ArmnnType,
typename T = armnn::ResolveType<ArmnnType>>
128 unsigned int inputShape[] = {1, 2, 2, 2};
129 unsigned int outputShape[] = {1, 1, 1, 8};
131 std::vector<float> input = std::vector<float>(
133 1.4f, 2.3f, 3.2f, 4.1f, 5.4f, 6.3f, 7.2f, 8.1f
136 std::vector<float> outputExpected = std::vector<float>(
138 1.4f, 2.3f, 3.2f, 4.1f, 5.4f, 6.3f, 7.2f, 8.1f
151 return SpaceToDepthTestImpl<T>(
152 workloadFactory, memoryManager, inputTensorInfo, outputTensorInfo, input, outputExpected, desc);
161 return SpaceToDepthSimpleTest1<armnn::DataType::QAsymmU8>(
170 return SpaceToDepthSimpleTest1<armnn::DataType::QAsymmU8>(
180 return SpaceToDepthSimpleTest1<armnn::DataType::Float16>(
189 return SpaceToDepthSimpleTest1<armnn::DataType::Float16>(
199 return SpaceToDepthSimpleTest1<armnn::DataType::Float32>(
208 return SpaceToDepthSimpleTest1<armnn::DataType::Float32>(
218 return SpaceToDepthSimpleTest2<armnn::DataType::Float32>(
227 return SpaceToDepthSimpleTest2<armnn::DataType::Float32>(
237 return SpaceToDepthSimpleTest2<armnn::DataType::QSymmS16>(
246 return SpaceToDepthSimpleTest2<armnn::DataType::QSymmS16>(
const TensorShape & GetShape() const
LayerTestResult< float, 4 > SpaceToDepthNhwcFloat32Test2(armnn::IWorkloadFactory &workloadFactory, const armnn::IBackendInternal::IMemoryManagerSharedPtr &memoryManager)
LayerTestResult< float, 4 > SpaceToDepthNhwcFloat32Test1(armnn::IWorkloadFactory &workloadFactory, const armnn::IBackendInternal::IMemoryManagerSharedPtr &memoryManager)
LayerTestResult< float, 4 > SpaceToDepthNchwFloat32Test2(armnn::IWorkloadFactory &workloadFactory, const armnn::IBackendInternal::IMemoryManagerSharedPtr &memoryManager)
void IgnoreUnused(Ts &&...)
LayerDescriptor m_Parameters
LayerTestResult< int16_t, 4 > SpaceToDepthNhwcQSymm16Test(armnn::IWorkloadFactory &workloadFactory, const armnn::IBackendInternal::IMemoryManagerSharedPtr &memoryManager)
void Permute(const armnn::TensorShape &dstShape, const armnn::PermutationVector &mappings, const void *src, void *dst, size_t dataTypeSize)
LayerTestResult< armnn::Half, 4 > SpaceToDepthNchwFloat16Test(armnn::IWorkloadFactory &workloadFactory, const armnn::IBackendInternal::IMemoryManagerSharedPtr &memoryManager)
LayerTestResult< float, 4 > SpaceToDepthNchwFloat32Test1(armnn::IWorkloadFactory &workloadFactory, const armnn::IBackendInternal::IMemoryManagerSharedPtr &memoryManager)
std::shared_ptr< IMemoryManager > IMemoryManagerSharedPtr
void SetQuantizationScale(float scale)
void CopyDataFromITensorHandle(void *memory, const armnn::ITensorHandle *tensorHandle)
LayerTestResult< armnn::Half, 4 > SpaceToDepthNhwcFloat16Test(armnn::IWorkloadFactory &workloadFactory, const armnn::IBackendInternal::IMemoryManagerSharedPtr &memoryManager)
virtual std::unique_ptr< ITensorHandle > CreateTensorHandle(const TensorInfo &tensorInfo, const bool IsMemoryManaged=true) const =0
unsigned int m_BlockSize
Scalar specifying the input block size. It must be >= 1.
virtual std::unique_ptr< IWorkload > CreateSpaceToDepth(const SpaceToDepthQueueDescriptor &descriptor, const WorkloadInfo &info) const
DataLayout m_DataLayout
The data layout to be used (NCHW, NHWC).
LayerTestResult< int16_t, 4 > SpaceToDepthNchwQSymm16Test(armnn::IWorkloadFactory &workloadFactory, const armnn::IBackendInternal::IMemoryManagerSharedPtr &memoryManager)
LayerTestResult< uint8_t, 4 > SpaceToDepthNhwcAsymmQ8Test(armnn::IWorkloadFactory &workloadFactory, const armnn::IBackendInternal::IMemoryManagerSharedPtr &memoryManager)
Contains information about inputs and outputs to a layer.
void SetQuantizationOffset(int32_t offset)
armnn::TensorShape Permuted(const armnn::TensorShape &srcShape, const armnn::PermutationVector &mappings)
void CopyDataToITensorHandle(armnn::ITensorHandle *tensorHandle, const void *memory)
LayerTestResult< uint8_t, 4 > SpaceToDepthNchwAsymmQ8Test(armnn::IWorkloadFactory &workloadFactory, const armnn::IBackendInternal::IMemoryManagerSharedPtr &memoryManager)