diff options
Diffstat (limited to 'src/backends')
4 files changed, 15 insertions, 7 deletions
diff --git a/src/backends/backendsCommon/test/StridedSliceAsyncEndToEndTest.hpp b/src/backends/backendsCommon/test/StridedSliceAsyncEndToEndTest.hpp index 84bf34dc60..00be81dd3e 100644 --- a/src/backends/backendsCommon/test/StridedSliceAsyncEndToEndTest.hpp +++ b/src/backends/backendsCommon/test/StridedSliceAsyncEndToEndTest.hpp @@ -90,11 +90,12 @@ void AsyncThreadedEndToEndTestImpl(INetworkPtr network, InputTensors& inputTensors = inputTensorsVec[i]; OutputTensors& outputTensors = outputTensorsVec[i]; IWorkingMemHandle& workingMemHandle = *workingMemHandles[i].get(); - threads.emplace_back([&]() { +ARMNN_NO_DEPRECATE_WARN_BEGIN // Run the async network runtime->Execute(workingMemHandle, inputTensors, outputTensors); +ARMNN_NO_DEPRECATE_WARN_END }); } @@ -184,9 +185,10 @@ void AsyncEndToEndTestImpl(INetworkPtr network, // Create WorkingMemHandle for this async network std::unique_ptr<IWorkingMemHandle> workingMemHandle = runtime->CreateWorkingMemHandle(networkId); IWorkingMemHandle& workingMemHandleRef = *workingMemHandle.get(); - +ARMNN_NO_DEPRECATE_WARN_BEGIN // Run the async network runtime->Execute(workingMemHandleRef, inputTensors, outputTensorsVec[0]); +ARMNN_NO_DEPRECATE_WARN_END } else { diff --git a/src/backends/gpuFsa/GpuFsaBackend.hpp b/src/backends/gpuFsa/GpuFsaBackend.hpp index f5a866b186..92a1c15d25 100644 --- a/src/backends/gpuFsa/GpuFsaBackend.hpp +++ b/src/backends/gpuFsa/GpuFsaBackend.hpp @@ -50,11 +50,13 @@ const BackendCapabilities gpuFsaCapabilities("GpuFsa", {"MultiAxisPacking", false}, {"SingleAxisPacking", false} }); - +ARMNN_NO_DEPRECATE_WARN_BEGIN class GpuFsaBackend : public IBackendInternal { public: + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("The GpuFsa backend will be removed from Arm NN in 24.08", "24.08") GpuFsaBackend() : m_CustomAllocator(nullptr) {}; + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("The GpuFsa backend will be removed from Arm NN in 24.08", "24.08") GpuFsaBackend(std::shared_ptr<ICustomAllocator> allocator) { UseCustomMemoryAllocator(allocator, armnn::EmptyOptional()); @@ -301,5 +303,6 @@ public: std::shared_ptr<GpuFsaBackendCustomAllocatorWrapper> m_CustomAllocator; bool m_UsingCustomAllocator = false; }; +ARMNN_NO_DEPRECATE_WARN_END } // namespace armnn diff --git a/src/backends/gpuFsa/GpuFsaRegistryInitializer.cpp b/src/backends/gpuFsa/GpuFsaRegistryInitializer.cpp index 9efb300576..ea0bbc299a 100644 --- a/src/backends/gpuFsa/GpuFsaRegistryInitializer.cpp +++ b/src/backends/gpuFsa/GpuFsaRegistryInitializer.cpp @@ -1,5 +1,5 @@ // -// Copyright © 2022-2023 Arm Ltd and Contributors. All rights reserved. +// Copyright © 2022-2024 Arm Ltd and Contributors. All rights reserved. // SPDX-License-Identifier: MIT // @@ -15,7 +15,9 @@ static BackendRegistry::StaticRegistryInitializer g_RegisterHelper GpuFsaBackend::GetIdStatic(), []() { - return IBackendInternalUniquePtr(new GpuFsaBackend); +ARMNN_NO_DEPRECATE_WARN_BEGIN + return IBackendInternalUniquePtr(new GpuFsaBackend); +ARMNN_NO_DEPRECATE_WARN_END } }; } // Anonymous namespace
\ No newline at end of file diff --git a/src/backends/gpuFsa/test/GpuFsaWorkloadFactoryHelper.hpp b/src/backends/gpuFsa/test/GpuFsaWorkloadFactoryHelper.hpp index c1d75d625b..e2cb4b925e 100644 --- a/src/backends/gpuFsa/test/GpuFsaWorkloadFactoryHelper.hpp +++ b/src/backends/gpuFsa/test/GpuFsaWorkloadFactoryHelper.hpp @@ -1,5 +1,5 @@ // -// Copyright © 2022-2023 Arm Ltd and Contributors. All rights reserved. +// Copyright © 2022-2024 Arm Ltd and Contributors. All rights reserved. // SPDX-License-Identifier: MIT // @@ -13,9 +13,9 @@ #include <gpuFsa/GpuFsaWorkloadFactory.hpp> #include "gpuFsa/GpuFsaTensorHandleFactory.hpp" +ARMNN_NO_DEPRECATE_WARN_BEGIN namespace { - template<> struct WorkloadFactoryHelper<armnn::GpuFsaWorkloadFactory> { @@ -43,3 +43,4 @@ struct WorkloadFactoryHelper<armnn::GpuFsaWorkloadFactory> using GpuFsaWorkloadFactoryHelper = WorkloadFactoryHelper<armnn::GpuFsaWorkloadFactory>; } // anonymous namespace +ARMNN_NO_DEPRECATE_WARN_END
\ No newline at end of file |