aboutsummaryrefslogtreecommitdiff
path: root/src/backends/neon/workloads/NeonBatchToSpaceNdWorkload.cpp
diff options
context:
space:
mode:
Diffstat (limited to 'src/backends/neon/workloads/NeonBatchToSpaceNdWorkload.cpp')
-rw-r--r--src/backends/neon/workloads/NeonBatchToSpaceNdWorkload.cpp149
1 files changed, 123 insertions, 26 deletions
diff --git a/src/backends/neon/workloads/NeonBatchToSpaceNdWorkload.cpp b/src/backends/neon/workloads/NeonBatchToSpaceNdWorkload.cpp
index 05b5899bdd..f66849a88d 100644
--- a/src/backends/neon/workloads/NeonBatchToSpaceNdWorkload.cpp
+++ b/src/backends/neon/workloads/NeonBatchToSpaceNdWorkload.cpp
@@ -1,17 +1,12 @@
//
-// Copyright © 2020, 2023 Arm Ltd and Contributors. All rights reserved.
+// Copyright © 2020-2023 Arm Ltd and Contributors. All rights reserved.
// SPDX-License-Identifier: MIT
//
#include "NeonBatchToSpaceNdWorkload.hpp"
-#include "NeonWorkloadUtils.hpp"
-
-#include <armnn/utility/NumericCast.hpp>
#include <armnn/utility/PolymorphicDowncast.hpp>
-#include <ResolveType.hpp>
-
namespace armnn
{
@@ -21,21 +16,71 @@ arm_compute::Status NeonBatchToSpaceNdWorkloadValidate(const TensorInfo& input,
const TensorInfo& output,
const BatchToSpaceNdDescriptor& descriptor)
{
- const arm_compute::TensorInfo aclInputInfo = BuildArmComputeTensorInfo(input, descriptor.m_DataLayout);
- const arm_compute::TensorInfo aclOutputInfo = BuildArmComputeTensorInfo(output, descriptor.m_DataLayout);
+ arm_compute::TensorInfo aclInputInfo = BuildArmComputeTensorInfo(input, descriptor.m_DataLayout);
+ arm_compute::TensorInfo aclOutputInfo = BuildArmComputeTensorInfo(output, descriptor.m_DataLayout);
+
+ arm_compute::Status statusBatchToSpace = arm_compute::Status(arm_compute::ErrorCode::OK);
+ arm_compute::Status statusReshapeInput = arm_compute::Status(arm_compute::ErrorCode::OK);
+ arm_compute::Status statusReshapeOutput = arm_compute::Status(arm_compute::ErrorCode::OK);
+
+ arm_compute::TensorInfo aclReshapeInputInfo = aclInputInfo;
+ arm_compute::TensorInfo aclReshapeOutputInfo = aclOutputInfo;
+
+ // When a spacial dimension is missing (rank=3) set W to 1
+ const unsigned int rank = input.GetNumDimensions();
+ if (rank == 3)
+ {
+ const arm_compute::TensorShape inputShape = aclInputInfo.tensor_shape();
+ const arm_compute::TensorShape outputShape = aclOutputInfo.tensor_shape();
+
+ if (descriptor.m_DataLayout == armnn::DataLayout::NHWC)
+ {
+ // In ACL dimensions are right to left: C, W, H, N
+ aclReshapeInputInfo.set_tensor_shape({inputShape.x(), 1, inputShape.y(), inputShape.z()});
+ aclReshapeOutputInfo.set_tensor_shape({outputShape.x(), 1, outputShape.y(), outputShape.z()});
+ }
+ else if (descriptor.m_DataLayout == armnn::DataLayout::NCHW)
+ {
+ // In ACL dimensions are right to left: W, H, C, N
+ aclReshapeInputInfo.set_tensor_shape({1, inputShape.x(), inputShape.y(), inputShape.z()});
+ aclReshapeOutputInfo.set_tensor_shape({1, outputShape.x(), outputShape.y(), outputShape.z()});
+ }
+ else
+ {
+ throw InvalidArgumentException("Unsupported or unknown DataLayout", CHECK_LOCATION());
+ }
+
+ statusReshapeInput = arm_compute::NEReshapeLayer::validate(&aclInputInfo, &aclReshapeInputInfo);
+ statusReshapeOutput = arm_compute::NEReshapeLayer::validate(&aclReshapeOutputInfo, &aclOutputInfo);
+ }
- // ArmNN blockShape is [H, W] Cl asks for W, H
+ // ArmNN blockShape is [H, W] ACl asks for W, H
int32_t blockHeight = armnn::numeric_cast<int32_t>(descriptor.m_BlockShape[0]);
- int32_t blockWidth = armnn::numeric_cast<int32_t>(descriptor.m_BlockShape[1]);
+ int32_t blockWidth = (rank == 3) ? 1 : armnn::numeric_cast<int32_t>(descriptor.m_BlockShape[1]);
+
+ const arm_compute::CropInfo cropInfo = BuildArmComputeCropInfo(descriptor, rank);
- const arm_compute::CropInfo cropInfo = BuildArmComputeCropInfo(descriptor);
+ statusBatchToSpace = arm_compute::NEBatchToSpaceLayer::validate(rank == 3 ? &aclReshapeInputInfo : &aclInputInfo,
+ blockWidth,
+ blockHeight,
+ rank == 3 ? &aclReshapeOutputInfo : &aclOutputInfo,
+ cropInfo);
- const arm_compute::Status aclStatus = arm_compute::NEBatchToSpaceLayer::validate(&aclInputInfo,
- blockWidth,
- blockHeight,
- &aclOutputInfo,
- cropInfo);
- return aclStatus;
+ if (statusReshapeInput.error_code() == arm_compute::ErrorCode::OK &&
+ statusReshapeOutput.error_code() == arm_compute::ErrorCode::OK &&
+ statusBatchToSpace.error_code() == arm_compute::ErrorCode::OK)
+ {
+ return arm_compute::Status(arm_compute::ErrorCode::OK,
+ "All BatchToSpace layers validate status OK.");
+ }
+ else
+ {
+ return arm_compute::Status(arm_compute::ErrorCode::RUNTIME_ERROR,
+ "BatchToSpace layer validate status failed."
+ + statusBatchToSpace.error_description()
+ + statusReshapeInput.error_description()
+ + statusReshapeOutput.error_description());
+ }
}
NeonBatchToSpaceNdWorkload::NeonBatchToSpaceNdWorkload(const BatchToSpaceNdQueueDescriptor& descriptor,
@@ -50,33 +95,85 @@ NeonBatchToSpaceNdWorkload::NeonBatchToSpaceNdWorkload(const BatchToSpaceNdQueue
m_Data.ValidateInputsOutputs("NeonBatchToSpaceNdWorkload", 1, 1);
- arm_compute::ITensor& input =
- armnn::PolymorphicPointerDowncast<IAclTensorHandle>(m_Data.m_Inputs[0])->GetTensor();
- arm_compute::ITensor& output =
- armnn::PolymorphicPointerDowncast<IAclTensorHandle>(m_Data.m_Outputs[0])->GetTensor();
+ arm_compute::ITensor& input = PolymorphicPointerDowncast<IAclTensorHandle>(m_Data.m_Inputs[0])->GetTensor();
+ arm_compute::ITensor& output = PolymorphicPointerDowncast<IAclTensorHandle>(m_Data.m_Outputs[0])->GetTensor();
arm_compute::DataLayout aclDataLayout = ConvertDataLayout(m_Data.m_Parameters.m_DataLayout);
input.info()->set_data_layout(aclDataLayout);
output.info()->set_data_layout(aclDataLayout);
- // ArmNN blockShape is [H, W] Cl asks for W, H
+ arm_compute::TensorInfo aclReshapeInputInfo = BuildArmComputeTensorInfo(info.m_InputTensorInfos[0],
+ m_Data.m_Parameters.m_DataLayout);
+ arm_compute::TensorInfo aclReshapeOutputInfo = BuildArmComputeTensorInfo(info.m_OutputTensorInfos[0],
+ m_Data.m_Parameters.m_DataLayout);
+
+ const unsigned int rank = info.m_InputTensorInfos[0].GetNumDimensions();
+ if (rank == 3)
+ {
+ const arm_compute::TensorShape inputShape = input.info()->tensor_shape();
+ const arm_compute::TensorShape outputShape = output.info()->tensor_shape();
+
+ // When a spacial dimension is missing set W to 1
+ if (m_Data.m_Parameters.m_DataLayout == armnn::DataLayout::NHWC)
+ {
+ // In ACL dimensions are right to left: C, W, H, N
+ aclReshapeInputInfo.set_tensor_shape({inputShape.x(), 1, inputShape.y(), inputShape.z()});
+ aclReshapeOutputInfo.set_tensor_shape({outputShape.x(), 1, outputShape.y(), outputShape.z()});
+ }
+ else if (m_Data.m_Parameters.m_DataLayout == armnn::DataLayout::NCHW)
+ {
+ // In ACL dimensions are right to left: W, H, C, N
+ aclReshapeInputInfo.set_tensor_shape({1, inputShape.x(), inputShape.y(), inputShape.z()});
+ aclReshapeOutputInfo.set_tensor_shape({1, outputShape.x(), outputShape.y(), outputShape.z()});
+ }
+ else
+ {
+ throw InvalidArgumentException("Unsupported or unknown DataLayout", CHECK_LOCATION());
+ }
+
+ m_ReshapeInputTensor.allocator()->init(aclReshapeInputInfo);
+ m_ReshapeOutputTensor.allocator()->init(aclReshapeOutputInfo);
+
+ InitialiseArmComputeTensorEmpty(m_ReshapeInputTensor);
+ InitialiseArmComputeTensorEmpty(m_ReshapeOutputTensor);
+
+ m_LayerReshapeInput.reset(new arm_compute::NEReshapeLayer());
+ m_LayerReshapeOutput.reset(new arm_compute::NEReshapeLayer());
+
+ m_LayerReshapeInput->configure(&input, &m_ReshapeInputTensor);
+ m_LayerReshapeOutput->configure(&m_ReshapeOutputTensor, &output);
+ }
+
+ // ArmNN blockShape is [H, W] ACl asks for W, H
int32_t blockHeight = armnn::numeric_cast<int32_t>(descriptor.m_Parameters.m_BlockShape[0]);
- int32_t blockWidth = armnn::numeric_cast<int32_t>(descriptor.m_Parameters.m_BlockShape[1]);
+ int32_t blockWidth = (rank == 3) ? 1 : armnn::numeric_cast<int32_t>(descriptor.m_Parameters.m_BlockShape[1]);
- const arm_compute::CropInfo cropInfo = BuildArmComputeCropInfo(descriptor.m_Parameters);
+ const arm_compute::CropInfo cropInfo = BuildArmComputeCropInfo(descriptor.m_Parameters, rank);
m_Layer.reset(new arm_compute::NEBatchToSpaceLayer());
- m_Layer->configure(&input, blockWidth, blockHeight, &output, cropInfo);
+ m_Layer->configure(rank == 3 ? &m_ReshapeInputTensor : &input,
+ blockWidth,
+ blockHeight,
+ rank == 3 ? &m_ReshapeOutputTensor : &output,
+ cropInfo);
m_Layer->prepare();
}
void NeonBatchToSpaceNdWorkload::Execute() const
{
+ ARMNN_SCOPED_PROFILING_EVENT_NEON_GUID("NeonBatchToSpaceNdWorkload_Execute", this->GetGuid());
+ if (m_LayerReshapeInput)
+ {
+ m_LayerReshapeInput->run();
+ }
if (m_Layer)
{
- ARMNN_SCOPED_PROFILING_EVENT_NEON_GUID("NeonSpaceToBatchNdWorkload_Execute", this->GetGuid());
m_Layer->run();
}
+ if (m_LayerReshapeOutput)
+ {
+ m_LayerReshapeOutput->run();
+ }
}
} //namespace armnn