aboutsummaryrefslogtreecommitdiff
path: root/src
diff options
context:
space:
mode:
Diffstat (limited to 'src')
-rw-r--r--src/armnn/DeviceSpec.hpp2
-rw-r--r--src/armnn/Network.cpp14
-rw-r--r--src/armnn/test/EndToEndTest.cpp20
-rw-r--r--src/armnn/test/JsonPrinterTests.cpp34
-rw-r--r--src/armnn/test/NetworkTests.cpp26
-rw-r--r--src/armnn/test/ProfilerTests.cpp2
-rw-r--r--src/armnn/test/RuntimeTests.cpp26
-rw-r--r--src/armnnCaffeParser/test/TestInputs.cpp2
-rw-r--r--src/armnnTfLiteParser/test/ParserFlatbuffersFixture.hpp2
-rw-r--r--src/armnnUtils/ParserPrototxtFixture.hpp2
-rw-r--r--src/backends/cl/test/Fp16SupportTest.cpp2
11 files changed, 66 insertions, 66 deletions
diff --git a/src/armnn/DeviceSpec.hpp b/src/armnn/DeviceSpec.hpp
index 34acbcbdec..af0d8f578b 100644
--- a/src/armnn/DeviceSpec.hpp
+++ b/src/armnn/DeviceSpec.hpp
@@ -22,7 +22,7 @@ public:
return std::vector<IBackendSharedPtr>();
}
- std::set<Compute> m_SupportedComputeDevices;
+ std::set<BackendId> m_SupportedComputeDevices;
};
}
diff --git a/src/armnn/Network.cpp b/src/armnn/Network.cpp
index c43f336145..8c70e5d793 100644
--- a/src/armnn/Network.cpp
+++ b/src/armnn/Network.cpp
@@ -94,7 +94,7 @@ bool CheckScaleSetOnQuantizedType(Layer* layer, Optional<std::vector<std::string
}
IOptimizedNetworkPtr Optimize(const INetwork& inNetwork,
- const std::vector<armnn::Compute>& backendPreferences,
+ const std::vector<BackendId>& backendPreferences,
const IDeviceSpec& deviceSpec,
const OptimizerOptions& options,
Optional<std::vector<std::string>&> errMessages)
@@ -133,8 +133,8 @@ IOptimizedNetworkPtr Optimize(const INetwork& inNetwork,
// determine which of the preferred backends we have available for use
// and whether we have specified CpuRef as one of those backends.
bool cpuRefUsed = false;
- std::vector<armnn::Compute> availablePreferredBackends;
- for (const armnn::Compute& backend : backendPreferences)
+ std::vector<BackendId> availablePreferredBackends;
+ for (const auto& backend : backendPreferences)
{
// Check if the backend is in the available backend devices.
if (std::find(spec.m_SupportedComputeDevices.begin(),
@@ -142,7 +142,7 @@ IOptimizedNetworkPtr Optimize(const INetwork& inNetwork,
spec.m_SupportedComputeDevices.end())
{
availablePreferredBackends.push_back(backend);
- if (armnn::Compute::CpuRef == backend) {
+ if (backend == armnn::Compute::CpuRef) {
cpuRefUsed = true;
}
}
@@ -183,7 +183,7 @@ IOptimizedNetworkPtr Optimize(const INetwork& inNetwork,
// which haven't had a scale set and report them all back.
bErrorFound = true;
}
- for (const armnn::Compute& backend : availablePreferredBackends)
+ for (const auto& backend : availablePreferredBackends)
{
// need to set the compute device on the layer
// before we can check if it is supported
@@ -205,7 +205,7 @@ IOptimizedNetworkPtr Optimize(const INetwork& inNetwork,
InsertConvertFp32ToFp16LayersAfter(optNetObjPtr->GetGraph(), *layer);
// Assign a supported backend to the newly introduced conversion layers
- auto AssignFirstSupportedBackend = [&](Layer* layer, Compute preferredBackend)
+ auto AssignFirstSupportedBackend = [&](Layer* layer, BackendId preferredBackend)
{
bool supportedBackendFound = false;
std::string reasonIfUnsupported;
@@ -218,7 +218,7 @@ IOptimizedNetworkPtr Optimize(const INetwork& inNetwork,
}
else
{
- for (const Compute& backend : availablePreferredBackends)
+ for (const auto& backend : availablePreferredBackends)
{
// Skip preferred backend (we already determined that it is not supported)
if (backend == preferredBackend)
diff --git a/src/armnn/test/EndToEndTest.cpp b/src/armnn/test/EndToEndTest.cpp
index 98b18411d4..d34bf69548 100644
--- a/src/armnn/test/EndToEndTest.cpp
+++ b/src/armnn/test/EndToEndTest.cpp
@@ -74,7 +74,7 @@ BOOST_AUTO_TEST_CASE(Unsigned8)
softmax->GetOutputSlot(0).SetTensorInfo(outputTensorInfo);
// optimize the network
- std::vector<armnn::Compute> backends = {armnn::Compute::CpuRef};
+ std::vector<armnn::BackendId> backends = {armnn::Compute::CpuRef};
IOptimizedNetworkPtr optNet = Optimize(*net, backends, runtime->GetDeviceSpec());
// Loads it into the runtime.
@@ -110,7 +110,7 @@ BOOST_AUTO_TEST_CASE(Unsigned8)
}
template <typename T>
-void ConstantUsageTest(const std::vector<armnn::Compute>& computeDevice,
+void ConstantUsageTest(const std::vector<armnn::BackendId>& computeDevice,
const armnn::TensorInfo& commonTensorInfo,
const std::vector<T>& inputData,
const std::vector<T>& constantData,
@@ -165,7 +165,7 @@ void ConstantUsageTest(const std::vector<armnn::Compute>& computeDevice,
BOOST_TEST(outputData == expectedOutputData);
}
-static void ConstantUsageFloat32Test(const std::vector<armnn::Compute>& computeDevice)
+static void ConstantUsageFloat32Test(const std::vector<armnn::BackendId>& computeDevice)
{
const armnn::TensorInfo commonTensorInfo({ 2, 3 }, armnn::DataType::Float32);
@@ -177,7 +177,7 @@ static void ConstantUsageFloat32Test(const std::vector<armnn::Compute>& computeD
);
}
-static void ConstantUsageUint8Test(const std::vector<armnn::Compute>& computeDevice)
+static void ConstantUsageUint8Test(const std::vector<armnn::BackendId>& computeDevice)
{
armnn::TensorInfo commonTensorInfo({ 2, 3 }, armnn::DataType::QuantisedAsymm8);
@@ -197,7 +197,7 @@ static void ConstantUsageUint8Test(const std::vector<armnn::Compute>& computeDev
BOOST_AUTO_TEST_CASE(ConstantUsage_Ref_Float32)
{
- std::vector<armnn::Compute> backends = {armnn::Compute::CpuRef};
+ std::vector<armnn::BackendId> backends = {armnn::Compute::CpuRef};
ConstantUsageFloat32Test(backends);
}
@@ -217,7 +217,7 @@ BOOST_AUTO_TEST_CASE(ConstantUsage_Cl_Float32)
BOOST_AUTO_TEST_CASE(ConstantUsage_Ref_Uint8)
{
- std::vector<armnn::Compute> backends = {armnn::Compute::CpuRef};
+ std::vector<armnn::BackendId> backends = {armnn::Compute::CpuRef};
ConstantUsageUint8Test(backends);
}
@@ -250,7 +250,7 @@ BOOST_AUTO_TEST_CASE(TrivialAdd)
add->GetOutputSlot(0).SetTensorInfo(tensorInfo);
// optimize the network
- std::vector<armnn::Compute> backends = {armnn::Compute::CpuRef};
+ std::vector<armnn::BackendId> backends = {armnn::Compute::CpuRef};
IOptimizedNetworkPtr optNet = Optimize(*net, backends, runtime->GetDeviceSpec());
// Loads it into the runtime.
@@ -349,7 +349,7 @@ BOOST_AUTO_TEST_CASE(MultipleOutputs)
activation3->GetOutputSlot(0).SetTensorInfo(tensorInfo);
// optimize the network
- std::vector<armnn::Compute> backends = {armnn::Compute::CpuRef};
+ std::vector<armnn::BackendId> backends = {armnn::Compute::CpuRef};
IOptimizedNetworkPtr optNet = Optimize(*net, backends, runtime->GetDeviceSpec());
// Loads it into the runtime.
@@ -410,7 +410,7 @@ BOOST_AUTO_TEST_CASE(FallbackToCpuRef)
pooling->GetOutputSlot(0).SetTensorInfo(TensorInfo({ 1, 1, 4, 4 }, DataType::Float32));
// optimize the network
- std::vector<Compute> backends = {Compute::CpuAcc, Compute::CpuRef};
+ std::vector<BackendId> backends = {Compute::CpuAcc, Compute::CpuRef};
IOptimizedNetworkPtr optNet = Optimize(*net, backends, runtime->GetDeviceSpec());
// Load it into the runtime. It should pass.
@@ -446,7 +446,7 @@ BOOST_AUTO_TEST_CASE(ErrorOnLoadNetwork)
pooling->GetOutputSlot(0).SetTensorInfo(TensorInfo({ 1, 1, 4, 4 }, DataType::Float32));
// optimize the network
- std::vector<Compute> backends = {Compute::CpuAcc};
+ std::vector<BackendId> backends = {Compute::CpuAcc};
IOptimizedNetworkPtr optNet = Optimize(*net, backends, runtime->GetDeviceSpec());
BOOST_CHECK(!optNet);
}
diff --git a/src/armnn/test/JsonPrinterTests.cpp b/src/armnn/test/JsonPrinterTests.cpp
index 44609ea6f9..93f32cc540 100644
--- a/src/armnn/test/JsonPrinterTests.cpp
+++ b/src/armnn/test/JsonPrinterTests.cpp
@@ -117,7 +117,7 @@ std::vector<std::string> ExtractSections(const std::string& exp)
return sections;
}
-std::string SoftmaxProfilerTestSetupHelper(const std::vector<armnn::Compute>& backends)
+std::string SoftmaxProfilerTestSetupHelper(const std::vector<armnn::BackendId>& backends)
{
using namespace armnn;
@@ -239,7 +239,7 @@ void SoftmaxProfilerTestValidationHelper(std::string& result, const std::string&
}
void SetupSoftmaxProfilerWithSpecifiedBackendsAndValidateJSONPrinterResult(
- const std::vector<armnn::Compute>& backends)
+ const std::vector<armnn::BackendId>& backends)
{
// setup the test fixture and obtain JSON Printer result
std::string result = SoftmaxProfilerTestSetupHelper(backends);
@@ -250,10 +250,10 @@ void SetupSoftmaxProfilerWithSpecifiedBackendsAndValidateJSONPrinterResult(
std::string changeLine40;
std::string changeLine45;
- switch(backends[0]) {
- case armnn::Compute::GpuAcc: backend = "Cl";
- changeLine31 = ",\n\"OpenClKernelTimer/: softmax_layer_max_shift_exp_sum_quantized_serial GWS[,,]\": {";
- changeLine39 = R"(us"
+ if (backends[0] == armnn::Compute::GpuAcc) {
+ backend = "Cl";
+ changeLine31 = ",\n\"OpenClKernelTimer/: softmax_layer_max_shift_exp_sum_quantized_serial GWS[,,]\": {";
+ changeLine39 = R"(us"
},
"OpenClKernelTimer/: softmax_layer_norm_quantized GWS[,,]": {
"raw": [
@@ -263,7 +263,7 @@ void SetupSoftmaxProfilerWithSpecifiedBackendsAndValidateJSONPrinterResult(
],
"unit": "us")";
- changeLine40 = R"(
+ changeLine40 = R"(
},
"CopyMemGeneric_Execute": {
"raw": [
@@ -272,11 +272,13 @@ void SetupSoftmaxProfilerWithSpecifiedBackendsAndValidateJSONPrinterResult(
],
"unit": "us")";
- changeLine45 = "}\n";
- break;
- case armnn::Compute::CpuAcc: backend = "Neon";
- changeLine31 = ",\n\"NeonKernelTimer/: NEFillBorderKernel\": {";
- changeLine39 = R"(us"
+ changeLine45 = "}\n";
+ }
+ else if (backends[0] == armnn::Compute::CpuAcc)
+ {
+ backend = "Neon";
+ changeLine31 = ",\n\"NeonKernelTimer/: NEFillBorderKernel\": {";
+ changeLine39 = R"(us"
},
"NeonKernelTimer/: NELogitsDMaxKernel": {
"raw": [
@@ -293,7 +295,7 @@ void SetupSoftmaxProfilerWithSpecifiedBackendsAndValidateJSONPrinterResult(
],
"unit": "us")";
- changeLine40 = R"(
+ changeLine40 = R"(
},
"CopyMemGeneric_Execute": {
"raw": [
@@ -302,11 +304,9 @@ void SetupSoftmaxProfilerWithSpecifiedBackendsAndValidateJSONPrinterResult(
],
"unit": "us")";
- changeLine45 = "}\n";
- break;
- default:
- break;
+ changeLine45 = "}\n";
}
+
std::string testData = R"({
"ArmNN": {
"inference_measurements": {
diff --git a/src/armnn/test/NetworkTests.cpp b/src/armnn/test/NetworkTests.cpp
index 3b426fa8ab..4f8dd7ea7b 100644
--- a/src/armnn/test/NetworkTests.cpp
+++ b/src/armnn/test/NetworkTests.cpp
@@ -65,7 +65,7 @@ BOOST_AUTO_TEST_CASE(SerializeToDot)
armnn::IRuntime::CreationOptions options;
armnn::IRuntimePtr runtime(armnn::IRuntime::Create(options));
- std::vector<armnn::Compute> backends = {armnn::Compute::CpuRef};
+ std::vector<armnn::BackendId> backends = {armnn::Compute::CpuRef};
armnn::IOptimizedNetworkPtr optimizedNet = armnn::Optimize(net, backends, runtime->GetDeviceSpec());
std::ostringstream ss;
@@ -472,7 +472,7 @@ BOOST_AUTO_TEST_CASE(OptimizeValidateCpuRefWorkloads)
armnn::IRuntime::CreationOptions options;
armnn::IRuntimePtr runtime(armnn::IRuntime::Create(options));
- std::vector<armnn::Compute> backends = { armnn::Compute::CpuRef };
+ std::vector<armnn::BackendId> backends = { armnn::Compute::CpuRef };
armnn::IOptimizedNetworkPtr optNet = armnn::Optimize(net, backends, runtime->GetDeviceSpec());
static_cast<armnn::OptimizedNetwork*>(optNet.get())->GetGraph().AllocateDynamicBuffers();
BOOST_CHECK(optNet);
@@ -503,7 +503,7 @@ BOOST_AUTO_TEST_CASE(OptimizeValidateCpuAccDeviceSupportLayerNoFallback)
armnn::IRuntime::CreationOptions options;
armnn::IRuntimePtr runtime(armnn::IRuntime::Create(options));
- std::vector<armnn::Compute> backends = { armnn::Compute::CpuAcc };
+ std::vector<armnn::BackendId> backends = { armnn::Compute::CpuAcc };
armnn::IOptimizedNetworkPtr optNet = armnn::Optimize(*net, backends, runtime->GetDeviceSpec());
BOOST_CHECK(optNet);
// validate workloads
@@ -534,7 +534,7 @@ BOOST_AUTO_TEST_CASE(OptimizeValidateGpuDeviceSupportLayerNoFallback)
armnn::IRuntime::CreationOptions options;
armnn::IRuntimePtr runtime(armnn::IRuntime::Create(options));
- std::vector<armnn::Compute> backends = { armnn::Compute::GpuAcc };
+ std::vector<armnn::BackendId> backends = { armnn::Compute::GpuAcc };
armnn::IOptimizedNetworkPtr optNet = armnn::Optimize(*net, backends, runtime->GetDeviceSpec());
BOOST_CHECK(optNet);
// validate workloads
@@ -570,7 +570,7 @@ BOOST_AUTO_TEST_CASE(OptimizeValidateDeviceNonSupportLayerNoFallback)
armnn::IRuntime::CreationOptions options;
armnn::IRuntimePtr runtime(armnn::IRuntime::Create(options));
- std::vector<armnn::Compute> backends = { armnn::Compute::CpuAcc };
+ std::vector<armnn::BackendId> backends = { armnn::Compute::CpuAcc };
armnn::IOptimizedNetworkPtr optNet = armnn::Optimize(*net, backends, runtime->GetDeviceSpec());
BOOST_CHECK(!optNet);
}
@@ -597,7 +597,7 @@ BOOST_AUTO_TEST_CASE(OptimizeValidateDeviceNonSupportLayerWithFallback)
armnn::IRuntime::CreationOptions options;
armnn::IRuntimePtr runtime(armnn::IRuntime::Create(options));
- std::vector<armnn::Compute> backends = { armnn::Compute::CpuAcc, armnn::Compute::CpuRef };
+ std::vector<armnn::BackendId> backends = { armnn::Compute::CpuAcc, armnn::Compute::CpuRef };
armnn::IOptimizedNetworkPtr optNet = armnn::Optimize(*net, backends, runtime->GetDeviceSpec());
BOOST_REQUIRE(optNet);
@@ -676,7 +676,7 @@ BOOST_AUTO_TEST_CASE(OptimizeValidateWorkloadsUndefinedComputeDevice)
armnn::IRuntime::CreationOptions options;
armnn::IRuntimePtr runtime(armnn::IRuntime::Create(options));
- std::vector<armnn::Compute> backends = { armnn::Compute::Undefined };
+ std::vector<armnn::BackendId> backends = { armnn::Compute::Undefined };
armnn::IOptimizedNetworkPtr optNet = armnn::Optimize(net, backends, runtime->GetDeviceSpec());
BOOST_CHECK(!optNet);
@@ -738,7 +738,7 @@ BOOST_AUTO_TEST_CASE(OptimizeValidateWorkloadsUndefinedComputeDeviceWithFallback
armnn::IRuntime::CreationOptions options;
armnn::IRuntimePtr runtime(armnn::IRuntime::Create(options));
- std::vector<armnn::Compute> backends = { armnn::Compute::Undefined, armnn::Compute::CpuRef };
+ std::vector<armnn::BackendId> backends = { armnn::Compute::Undefined, armnn::Compute::CpuRef };
armnn::IOptimizedNetworkPtr optNet = armnn::Optimize(net, backends, runtime->GetDeviceSpec());
BOOST_CHECK(optNet);
@@ -774,7 +774,7 @@ BOOST_AUTO_TEST_CASE(OptimizeValidateWorkloadsDuplicateComputeDeviceWithFallback
armnn::IRuntime::CreationOptions options;
armnn::IRuntimePtr runtime(armnn::IRuntime::Create(options));
- std::vector<armnn::Compute> backends = { armnn::Compute::CpuAcc,
+ std::vector<armnn::BackendId> backends = { armnn::Compute::CpuAcc,
armnn::Compute::GpuAcc,
armnn::Compute::CpuRef };
@@ -818,7 +818,7 @@ BOOST_AUTO_TEST_CASE(OptimizeValidateWorkloadsCpuRefPermuteLayer)
armnn::IRuntime::CreationOptions options;
armnn::IRuntimePtr runtime(armnn::IRuntime::Create(options));
- std::vector<armnn::Compute> backends = {armnn::Compute::CpuRef};
+ std::vector<armnn::BackendId> backends = {armnn::Compute::CpuRef};
// build up the structure of the network
armnn::INetworkPtr net(armnn::INetwork::Create());
@@ -851,7 +851,7 @@ BOOST_AUTO_TEST_CASE(OptimizeValidateWorkloadsCpuRefMeanLayer)
armnn::IRuntime::CreationOptions options;
armnn::IRuntimePtr runtime(armnn::IRuntime::Create(options));
- std::vector<armnn::Compute> backends = {armnn::Compute::CpuRef};
+ std::vector<armnn::BackendId> backends = {armnn::Compute::CpuRef};
// build up the structure of the network
armnn::INetworkPtr net(armnn::INetwork::Create());
@@ -909,7 +909,7 @@ BOOST_AUTO_TEST_CASE(FP16TurboModeTestOnCpuRef)
armnn::IRuntime::CreationOptions options;
armnn::IRuntimePtr runtime(armnn::IRuntime::Create(options));
- std::vector<armnn::Compute> backends = {armnn::Compute::CpuRef};
+ std::vector<armnn::BackendId> backends = {armnn::Compute::CpuRef};
armnn::OptimizerOptions optimizerOptions;
optimizerOptions.m_ReduceFp32ToFp16 = true;
@@ -976,7 +976,7 @@ BOOST_AUTO_TEST_CASE(FP16TurboModeTestOnGpuAcc)
armnn::IRuntime::CreationOptions options;
armnn::IRuntimePtr runtime(armnn::IRuntime::Create(options));
- std::vector<armnn::Compute> backends = {armnn::Compute::GpuAcc};
+ std::vector<armnn::BackendId> backends = {armnn::Compute::GpuAcc};
armnn::OptimizerOptions optimizerOptions;
optimizerOptions.m_ReduceFp32ToFp16 = true;
diff --git a/src/armnn/test/ProfilerTests.cpp b/src/armnn/test/ProfilerTests.cpp
index 8d7684459a..650c13a4f4 100644
--- a/src/armnn/test/ProfilerTests.cpp
+++ b/src/armnn/test/ProfilerTests.cpp
@@ -163,7 +163,7 @@ BOOST_AUTO_TEST_CASE(RuntimeLoadNetwork)
armnn::NetworkId networkIdentifier = 1;
armnn::INetworkPtr mockNetwork(armnn::INetwork::Create());
mockNetwork->AddInputLayer(0, "test layer");
- std::vector<armnn::Compute> backends = { armnn::Compute::CpuRef };
+ std::vector<armnn::BackendId> backends = { armnn::Compute::CpuRef };
runtime->LoadNetwork(networkIdentifier, armnn::Optimize(*mockNetwork, backends, runtime->GetDeviceSpec()));
// Check that now there's a profiler registered for this thread (created and registered by the loading the network).
diff --git a/src/armnn/test/RuntimeTests.cpp b/src/armnn/test/RuntimeTests.cpp
index 0237387b93..76f5774a49 100644
--- a/src/armnn/test/RuntimeTests.cpp
+++ b/src/armnn/test/RuntimeTests.cpp
@@ -39,7 +39,7 @@ BOOST_AUTO_TEST_CASE(RuntimeUnloadNetwork)
armnn::NetworkId networkIdentifier1 = 1;
armnn::INetworkPtr mockNetwork1(armnn::INetwork::Create());
mockNetwork1->AddInputLayer(0, "test layer");
- std::vector<armnn::Compute> backends = {armnn::Compute::CpuRef};
+ std::vector<armnn::BackendId> backends = {armnn::Compute::CpuRef};
runtime->LoadNetwork(networkIdentifier1, Optimize(*mockNetwork1, backends, runtime->GetDeviceSpec()));
// Mock network 2.
@@ -71,7 +71,7 @@ struct DisableGlobalLeakChecking
BOOST_GLOBAL_FIXTURE(DisableGlobalLeakChecking);
-void CreateAndDropDummyNetwork(const std::vector<armnn::Compute>& backends, armnn::Runtime& runtime)
+void CreateAndDropDummyNetwork(const std::vector<armnn::BackendId>& backends, armnn::Runtime& runtime)
{
armnn::NetworkId networkIdentifier;
{
@@ -129,7 +129,7 @@ BOOST_AUTO_TEST_CASE(RuntimeMemoryLeaksGpuAcc)
armnn::Runtime runtime(options);
armnn::RuntimeLoadedNetworksReserve(&runtime);
- std::vector<armnn::Compute> backends = {armnn::Compute::GpuAcc};
+ std::vector<armnn::BackendId> backends = {armnn::Compute::GpuAcc};
{
// Do a warmup of this so we make sure that all one-time
// initialization happens before we do the leak checking.
@@ -158,7 +158,7 @@ BOOST_AUTO_TEST_CASE(RuntimeMemoryLeaksCpuAcc)
armnn::Runtime runtime(options);
armnn::RuntimeLoadedNetworksReserve(&runtime);
- std::vector<armnn::Compute> backends = {armnn::Compute::CpuAcc};
+ std::vector<armnn::BackendId> backends = {armnn::Compute::CpuAcc};
{
// Do a warmup of this so we make sure that all one-time
// initialization happens before we do the leak checking.
@@ -187,7 +187,7 @@ BOOST_AUTO_TEST_CASE(RuntimeMemoryLeaksCpuRef)
armnn::Runtime runtime(options);
armnn::RuntimeLoadedNetworksReserve(&runtime);
- std::vector<armnn::Compute> backends = {armnn::Compute::CpuRef};
+ std::vector<armnn::BackendId> backends = {armnn::Compute::CpuRef};
{
// Do a warmup of this so we make sure that all one-time
// initialization happens before we do the leak checking.
@@ -243,7 +243,7 @@ BOOST_AUTO_TEST_CASE(RuntimeMemoryUsage)
VALGRIND_COUNT_LEAKS(leakedBefore, dubious, reachableBefore, suppressed);
// build a mock-network and load it into the runtime
- std::vector<armnn::Compute> backends = {armnn::Compute::GpuAcc};
+ std::vector<armnn::BackendId> backends = {armnn::Compute::GpuAcc};
{
armnn::TensorInfo inputTensorInfo(armnn::TensorShape({ 7, 7 }), armnn::DataType::Float32);
armnn::TensorInfo outputTensorInfo(armnn::TensorShape({ 7, 7 }), armnn::DataType::Float32);
@@ -331,7 +331,7 @@ BOOST_AUTO_TEST_CASE(RuntimeMemoryLeak)
mockNetwork1->AddInputLayer(0, "test layer");
- std::vector<armnn::Compute> backends = {armnn::Compute::CpuRef};
+ std::vector<armnn::BackendId> backends = {armnn::Compute::CpuRef};
runtime.LoadNetwork(networkIdentifier1, Optimize(*mockNetwork1, backends, runtime.GetDeviceSpec()));
}
@@ -379,7 +379,7 @@ BOOST_AUTO_TEST_CASE(RuntimeValidateCpuAccDeviceSupportLayerNoFallback)
armnn::IRuntime::CreationOptions options;
armnn::IRuntimePtr runtime(armnn::IRuntime::Create(options));
- std::vector<armnn::Compute> backends = { armnn::Compute::CpuAcc };
+ std::vector<armnn::BackendId> backends = { armnn::Compute::CpuAcc };
armnn::IOptimizedNetworkPtr optNet = armnn::Optimize(*net, backends, runtime->GetDeviceSpec());
BOOST_CHECK(optNet);
@@ -406,7 +406,7 @@ BOOST_AUTO_TEST_CASE(RuntimeValidateGpuDeviceSupportLayerNoFallback)
armnn::IRuntime::CreationOptions options;
armnn::IRuntimePtr runtime(armnn::IRuntime::Create(options));
- std::vector<armnn::Compute> backends = { armnn::Compute::GpuAcc };
+ std::vector<armnn::BackendId> backends = { armnn::Compute::GpuAcc };
armnn::IOptimizedNetworkPtr optNet = armnn::Optimize(*net, backends, runtime->GetDeviceSpec());
BOOST_CHECK(optNet);
@@ -442,7 +442,7 @@ BOOST_AUTO_TEST_CASE(RuntimeCpuRef)
normalize->GetOutputSlot(0).SetTensorInfo(TensorInfo({ 1, 1, 4, 4 }, DataType::Float32));
// optimize the network
- std::vector<armnn::Compute> backends = { armnn::Compute::CpuRef };
+ std::vector<armnn::BackendId> backends = { armnn::Compute::CpuRef };
IOptimizedNetworkPtr optNet = Optimize(*net, backends, runtime->GetDeviceSpec());
// Load it into the runtime. It should success.
@@ -476,7 +476,7 @@ BOOST_AUTO_TEST_CASE(RuntimeFallbackToCpuRef)
normalize->GetOutputSlot(0).SetTensorInfo(TensorInfo({ 1, 1, 4, 4 }, DataType::Float32));
// Allow fallback to CpuRef.
- std::vector<armnn::Compute> backends = { armnn::Compute::CpuAcc, armnn::Compute::CpuRef };
+ std::vector<armnn::BackendId> backends = { armnn::Compute::CpuAcc, armnn::Compute::CpuRef };
// optimize the network
IOptimizedNetworkPtr optNet = Optimize(*net, backends, runtime->GetDeviceSpec());
@@ -524,7 +524,7 @@ BOOST_AUTO_TEST_CASE(IVGCVSW_1929_QuantizedSoftmaxIssue)
armnn::DataType::QuantisedAsymm8
));
- std::vector<armnn::Compute> backends = {armnn::Compute::CpuRef};
+ std::vector<armnn::BackendId> backends = {armnn::Compute::CpuRef};
std::vector<std::string> errMessages;
armnn::IOptimizedNetworkPtr optNet = Optimize(
*net,
@@ -533,7 +533,7 @@ BOOST_AUTO_TEST_CASE(IVGCVSW_1929_QuantizedSoftmaxIssue)
OptimizerOptions(),
errMessages
);
-
+
BOOST_TEST(errMessages.size() == 1);
BOOST_TEST(errMessages[0] ==
"ERROR: output 0 of layer Softmax (softmax) is of type "
diff --git a/src/armnnCaffeParser/test/TestInputs.cpp b/src/armnnCaffeParser/test/TestInputs.cpp
index 616d75d95c..bd13a0984d 100644
--- a/src/armnnCaffeParser/test/TestInputs.cpp
+++ b/src/armnnCaffeParser/test/TestInputs.cpp
@@ -42,7 +42,7 @@ BOOST_AUTO_TEST_CASE(InputShapes)
armnn::NetworkId netId;
// Check everything works normally
- std::vector<armnn::Compute> backends = {armnn::Compute::CpuRef};
+ std::vector<armnn::BackendId> backends = {armnn::Compute::CpuRef};
{
network = parser->CreateNetworkFromString(explicitInput.c_str(), {}, { "data" });
BOOST_TEST(network.get());
diff --git a/src/armnnTfLiteParser/test/ParserFlatbuffersFixture.hpp b/src/armnnTfLiteParser/test/ParserFlatbuffersFixture.hpp
index 404cbf0522..e9cbb24604 100644
--- a/src/armnnTfLiteParser/test/ParserFlatbuffersFixture.hpp
+++ b/src/armnnTfLiteParser/test/ParserFlatbuffersFixture.hpp
@@ -44,7 +44,7 @@ struct ParserFlatbuffersFixture
std::vector<uint8_t> m_GraphBinary;
std::string m_JsonString;
std::unique_ptr<ITfLiteParser, void (*)(ITfLiteParser *parser)> m_Parser;
- std::vector<std::pair<armnn::IRuntimePtr, armnn::Compute>> m_Runtimes;
+ std::vector<std::pair<armnn::IRuntimePtr, armnn::BackendId>> m_Runtimes;
armnn::NetworkId m_NetworkIdentifier;
/// If the single-input-single-output overload of Setup() is called, these will store the input and output name
diff --git a/src/armnnUtils/ParserPrototxtFixture.hpp b/src/armnnUtils/ParserPrototxtFixture.hpp
index c502ad9619..1adada7a26 100644
--- a/src/armnnUtils/ParserPrototxtFixture.hpp
+++ b/src/armnnUtils/ParserPrototxtFixture.hpp
@@ -59,7 +59,7 @@ struct ParserPrototxtFixture
std::string m_Prototext;
std::unique_ptr<TParser, void(*)(TParser* parser)> m_Parser;
- std::vector<std::pair<armnn::IRuntimePtr, armnn::Compute>> m_Runtimes;
+ std::vector<std::pair<armnn::IRuntimePtr, armnn::BackendId>> m_Runtimes;
armnn::NetworkId m_NetworkIdentifier;
/// If the single-input-single-output overload of Setup() is called, these will store the input and output name
diff --git a/src/backends/cl/test/Fp16SupportTest.cpp b/src/backends/cl/test/Fp16SupportTest.cpp
index 90bef3647b..e8d459170e 100644
--- a/src/backends/cl/test/Fp16SupportTest.cpp
+++ b/src/backends/cl/test/Fp16SupportTest.cpp
@@ -73,7 +73,7 @@ BOOST_AUTO_TEST_CASE(Fp16AdditionTest)
additionLayer->GetOutputSlot(0).SetTensorInfo(fp16TensorInfo);
// optimize the network
- std::vector<Compute> backends = {Compute::GpuAcc};
+ std::vector<BackendId> backends = {Compute::GpuAcc};
IOptimizedNetworkPtr optNet = Optimize(*net, backends, runtime->GetDeviceSpec());
// Loads it into the runtime.