aboutsummaryrefslogtreecommitdiff
path: root/src/backends/cl/workloads
diff options
context:
space:
mode:
Diffstat (limited to 'src/backends/cl/workloads')
-rw-r--r--src/backends/cl/workloads/ClAdditionWorkload.cpp15
-rw-r--r--src/backends/cl/workloads/ClAdditionWorkload.hpp7
-rw-r--r--src/backends/cl/workloads/ClConvertFp16ToFp32Workload.cpp16
-rw-r--r--src/backends/cl/workloads/ClConvertFp16ToFp32Workload.hpp4
-rw-r--r--src/backends/cl/workloads/ClConvertFp32ToFp16Workload.cpp16
-rw-r--r--src/backends/cl/workloads/ClConvertFp32ToFp16Workload.hpp4
-rw-r--r--src/backends/cl/workloads/ClPadWorkload.cpp15
-rw-r--r--src/backends/cl/workloads/ClPadWorkload.hpp7
-rw-r--r--src/backends/cl/workloads/ClSubtractionWorkload.cpp15
-rw-r--r--src/backends/cl/workloads/ClSubtractionWorkload.hpp7
10 files changed, 29 insertions, 77 deletions
diff --git a/src/backends/cl/workloads/ClAdditionWorkload.cpp b/src/backends/cl/workloads/ClAdditionWorkload.cpp
index aa032e872c..ec0dd30275 100644
--- a/src/backends/cl/workloads/ClAdditionWorkload.cpp
+++ b/src/backends/cl/workloads/ClAdditionWorkload.cpp
@@ -37,10 +37,9 @@ void ClAdditionWorkload<T...>::Execute() const
m_Layer.run();
}
-bool ClAdditionValidate(const TensorInfo& input0,
- const TensorInfo& input1,
- const TensorInfo& output,
- std::string* reasonIfUnsupported)
+arm_compute::Status ClAdditionValidate(const TensorInfo& input0,
+ const TensorInfo& input1,
+ const TensorInfo& output)
{
const arm_compute::TensorInfo aclInput0Info = BuildArmComputeTensorInfo(input0);
const arm_compute::TensorInfo aclInput1Info = BuildArmComputeTensorInfo(input1);
@@ -51,13 +50,7 @@ bool ClAdditionValidate(const TensorInfo& input0,
&aclOutputInfo,
g_AclConvertPolicy);
- const bool supported = (aclStatus.error_code() == arm_compute::ErrorCode::OK);
- if (!supported && reasonIfUnsupported)
- {
- *reasonIfUnsupported = aclStatus.error_description();
- }
-
- return supported;
+ return aclStatus;
}
} //namespace armnn
diff --git a/src/backends/cl/workloads/ClAdditionWorkload.hpp b/src/backends/cl/workloads/ClAdditionWorkload.hpp
index 3e4ee26793..c5e6affac8 100644
--- a/src/backends/cl/workloads/ClAdditionWorkload.hpp
+++ b/src/backends/cl/workloads/ClAdditionWorkload.hpp
@@ -24,8 +24,7 @@ private:
mutable arm_compute::CLArithmeticAddition m_Layer;
};
-bool ClAdditionValidate(const TensorInfo& input0,
- const TensorInfo& input1,
- const TensorInfo& output,
- std::string* reasonIfUnsupported);
+arm_compute::Status ClAdditionValidate(const TensorInfo& input0,
+ const TensorInfo& input1,
+ const TensorInfo& output);
} //namespace armnn
diff --git a/src/backends/cl/workloads/ClConvertFp16ToFp32Workload.cpp b/src/backends/cl/workloads/ClConvertFp16ToFp32Workload.cpp
index e7663b4ca4..2c9a0e1fc2 100644
--- a/src/backends/cl/workloads/ClConvertFp16ToFp32Workload.cpp
+++ b/src/backends/cl/workloads/ClConvertFp16ToFp32Workload.cpp
@@ -32,19 +32,15 @@ void ClConvertFp16ToFp32Workload::Execute() const
m_Layer.run();
}
-arm_compute::Status ClConvertFp16ToFp32WorkloadValidate(const TensorInfo& input,
- const TensorInfo& output,
- std::string* reasonIfUnsupported)
+arm_compute::Status ClConvertFp16ToFp32WorkloadValidate(const TensorInfo& input, const TensorInfo& output)
{
if (input.GetDataType() != DataType::Float16)
{
- *reasonIfUnsupported = "Input should be Float16";
- return arm_compute::Status(arm_compute::ErrorCode::RUNTIME_ERROR, *reasonIfUnsupported);
+ return arm_compute::Status(arm_compute::ErrorCode::RUNTIME_ERROR, "Input should be Float16");
}
if (output.GetDataType() != DataType::Float32)
{
- *reasonIfUnsupported = "Output should be Float32";
- return arm_compute::Status(arm_compute::ErrorCode::RUNTIME_ERROR, *reasonIfUnsupported);
+ return arm_compute::Status(arm_compute::ErrorCode::RUNTIME_ERROR, "Output should be Float32");
}
const arm_compute::TensorInfo aclInputInfo = BuildArmComputeTensorInfo(input);
@@ -53,12 +49,6 @@ arm_compute::Status ClConvertFp16ToFp32WorkloadValidate(const TensorInfo& input,
const arm_compute::Status aclStatus = arm_compute::CLDepthConvertLayer::validate(
&aclInputInfo, &aclOutputInfo, g_AclConvertPolicy, 0);
- const bool supported = (aclStatus.error_code() == arm_compute::ErrorCode::OK);
- if (!supported && reasonIfUnsupported)
- {
- *reasonIfUnsupported = aclStatus.error_description();
- }
-
return aclStatus;
}
diff --git a/src/backends/cl/workloads/ClConvertFp16ToFp32Workload.hpp b/src/backends/cl/workloads/ClConvertFp16ToFp32Workload.hpp
index b6447488f7..f5f230d869 100644
--- a/src/backends/cl/workloads/ClConvertFp16ToFp32Workload.hpp
+++ b/src/backends/cl/workloads/ClConvertFp16ToFp32Workload.hpp
@@ -23,8 +23,6 @@ private:
mutable arm_compute::CLDepthConvertLayer m_Layer;
};
-arm_compute::Status ClConvertFp16ToFp32WorkloadValidate(const TensorInfo& input,
- const TensorInfo& output,
- std::string* reasonIfUnsupported);
+arm_compute::Status ClConvertFp16ToFp32WorkloadValidate(const TensorInfo& input, const TensorInfo& output);
} //namespace armnn
diff --git a/src/backends/cl/workloads/ClConvertFp32ToFp16Workload.cpp b/src/backends/cl/workloads/ClConvertFp32ToFp16Workload.cpp
index 2ae4adc424..6758180a6e 100644
--- a/src/backends/cl/workloads/ClConvertFp32ToFp16Workload.cpp
+++ b/src/backends/cl/workloads/ClConvertFp32ToFp16Workload.cpp
@@ -32,19 +32,15 @@ void ClConvertFp32ToFp16Workload::Execute() const
m_Layer.run();
}
-arm_compute::Status ClConvertFp32ToFp16WorkloadValidate(const TensorInfo& input,
- const TensorInfo& output,
- std::string* reasonIfUnsupported)
+arm_compute::Status ClConvertFp32ToFp16WorkloadValidate(const TensorInfo& input, const TensorInfo& output)
{
if (input.GetDataType() != DataType::Float32)
{
- *reasonIfUnsupported = "Input should be Float32";
- return arm_compute::Status(arm_compute::ErrorCode::RUNTIME_ERROR, *reasonIfUnsupported);
+ return arm_compute::Status(arm_compute::ErrorCode::RUNTIME_ERROR, "Input should be Float32");
}
if (output.GetDataType() != DataType::Float16)
{
- *reasonIfUnsupported = "Output should be Float16";
- return arm_compute::Status(arm_compute::ErrorCode::RUNTIME_ERROR, *reasonIfUnsupported);
+ return arm_compute::Status(arm_compute::ErrorCode::RUNTIME_ERROR, "Output should be Float16");
}
const arm_compute::TensorInfo aclInputInfo = BuildArmComputeTensorInfo(input);
@@ -53,12 +49,6 @@ arm_compute::Status ClConvertFp32ToFp16WorkloadValidate(const TensorInfo& input,
const arm_compute::Status aclStatus = arm_compute::CLDepthConvertLayer::validate(
&aclInputInfo, &aclOutputInfo, g_AclConvertPolicy, 0);
- const bool supported = (aclStatus.error_code() == arm_compute::ErrorCode::OK);
- if (!supported && reasonIfUnsupported)
- {
- *reasonIfUnsupported = aclStatus.error_description();
- }
-
return aclStatus;
}
diff --git a/src/backends/cl/workloads/ClConvertFp32ToFp16Workload.hpp b/src/backends/cl/workloads/ClConvertFp32ToFp16Workload.hpp
index 95d19905d7..28d0bfa9b0 100644
--- a/src/backends/cl/workloads/ClConvertFp32ToFp16Workload.hpp
+++ b/src/backends/cl/workloads/ClConvertFp32ToFp16Workload.hpp
@@ -23,8 +23,6 @@ private:
mutable arm_compute::CLDepthConvertLayer m_Layer;
};
-arm_compute::Status ClConvertFp32ToFp16WorkloadValidate(const TensorInfo& input,
- const TensorInfo& output,
- std::string* reasonIfUnsupported);
+arm_compute::Status ClConvertFp32ToFp16WorkloadValidate(const TensorInfo& input, const TensorInfo& output);
} //namespace armnn
diff --git a/src/backends/cl/workloads/ClPadWorkload.cpp b/src/backends/cl/workloads/ClPadWorkload.cpp
index 45dc5e8be7..89b0d8fde9 100644
--- a/src/backends/cl/workloads/ClPadWorkload.cpp
+++ b/src/backends/cl/workloads/ClPadWorkload.cpp
@@ -35,10 +35,9 @@ void ClPadWorkload<T...>::Execute() const
m_Layer.run();
}
-bool ClPadValidate(const TensorInfo& input,
- const TensorInfo& output,
- const PadDescriptor& descriptor,
- std::string* reasonIfUnsupported)
+arm_compute::Status ClPadValidate(const TensorInfo& input,
+ const TensorInfo& output,
+ const PadDescriptor& descriptor)
{
const arm_compute::TensorInfo aclInputInfo = BuildArmComputeTensorInfo(input);
const arm_compute::TensorInfo aclOutputInfo = BuildArmComputeTensorInfo(output);
@@ -48,13 +47,7 @@ bool ClPadValidate(const TensorInfo& input,
&aclOutputInfo,
padList);
- const bool supported = (aclStatus.error_code() == arm_compute::ErrorCode::OK);
- if (!supported && reasonIfUnsupported)
- {
- *reasonIfUnsupported = aclStatus.error_description();
- }
-
- return supported;
+ return aclStatus;
}
} // namespace armnn
diff --git a/src/backends/cl/workloads/ClPadWorkload.hpp b/src/backends/cl/workloads/ClPadWorkload.hpp
index a7ad6670a7..97f57fdeac 100644
--- a/src/backends/cl/workloads/ClPadWorkload.hpp
+++ b/src/backends/cl/workloads/ClPadWorkload.hpp
@@ -23,10 +23,9 @@ private:
mutable arm_compute::CLPadLayer m_Layer;
};
-bool ClPadValidate(const TensorInfo& input,
- const TensorInfo& output,
- const PadDescriptor& descriptor,
- std::string* reasonIfUnsupported);
+arm_compute::Status ClPadValidate(const TensorInfo& input,
+ const TensorInfo& output,
+ const PadDescriptor& descriptor);
} //namespace armnn
diff --git a/src/backends/cl/workloads/ClSubtractionWorkload.cpp b/src/backends/cl/workloads/ClSubtractionWorkload.cpp
index 8efed94293..1967fae354 100644
--- a/src/backends/cl/workloads/ClSubtractionWorkload.cpp
+++ b/src/backends/cl/workloads/ClSubtractionWorkload.cpp
@@ -35,10 +35,9 @@ void ClSubtractionWorkload::Execute() const
m_Layer.run();
}
-bool ClSubtractionValidate(const TensorInfo& input0,
- const TensorInfo& input1,
- const TensorInfo& output,
- std::string* reasonIfUnsupported)
+arm_compute::Status ClSubtractionValidate(const TensorInfo& input0,
+ const TensorInfo& input1,
+ const TensorInfo& output)
{
const arm_compute::TensorInfo aclInput0Info = BuildArmComputeTensorInfo(input0);
const arm_compute::TensorInfo aclInput1Info = BuildArmComputeTensorInfo(input1);
@@ -49,13 +48,7 @@ bool ClSubtractionValidate(const TensorInfo& input0,
&aclOutputInfo,
g_AclConvertPolicy);
- const bool supported = (aclStatus.error_code() == arm_compute::ErrorCode::OK);
- if (!supported && reasonIfUnsupported)
- {
- *reasonIfUnsupported = aclStatus.error_description();
- }
-
- return supported;
+ return aclStatus;
}
} //namespace armnn
diff --git a/src/backends/cl/workloads/ClSubtractionWorkload.hpp b/src/backends/cl/workloads/ClSubtractionWorkload.hpp
index 7dd608bf8a..3a4210da07 100644
--- a/src/backends/cl/workloads/ClSubtractionWorkload.hpp
+++ b/src/backends/cl/workloads/ClSubtractionWorkload.hpp
@@ -23,8 +23,7 @@ private:
mutable arm_compute::CLArithmeticSubtraction m_Layer;
};
-bool ClSubtractionValidate(const TensorInfo& input0,
- const TensorInfo& input1,
- const TensorInfo& output,
- std::string* reasonIfUnsupported);
+arm_compute::Status ClSubtractionValidate(const TensorInfo& input0,
+ const TensorInfo& input1,
+ const TensorInfo& output);
} //namespace armnn