From ca3a3e0fd86a07de9e073ba31dc2b42d6ca84536 Mon Sep 17 00:00:00 2001 From: Finn Williams Date: Fri, 11 Jun 2021 15:04:02 +0100 Subject: IVGCVSW-6062 Rework the async threadpool !armnn:5801 Signed-off-by: Finn Williams Change-Id: I9964d0899ce752441f380edddbd974010257b2dd --- 1.3/ArmnnDriverImpl.cpp | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) (limited to '1.3') diff --git a/1.3/ArmnnDriverImpl.cpp b/1.3/ArmnnDriverImpl.cpp index 5c5e6071..6f1b5098 100644 --- a/1.3/ArmnnDriverImpl.cpp +++ b/1.3/ArmnnDriverImpl.cpp @@ -202,8 +202,7 @@ Return ArmnnDriverImpl::prepareArmnnModel_1_3( std::string msg; armnn::INetworkProperties networkProperties(options.isAsyncModelExecutionEnabled(), MemorySource::Undefined, - MemorySource::Undefined, - options.getNoOfArmnnThreads()); + MemorySource::Undefined); try { if (runtime->LoadNetwork(netId, move(optNet), msg, networkProperties) != armnn::Status::Success) @@ -234,7 +233,8 @@ Return ArmnnDriverImpl::prepareArmnnModel_1_3( options.GetRequestInputsAndOutputsDumpDir(), options.IsGpuProfilingEnabled(), priority, - options.isAsyncModelExecutionEnabled())); + options.isAsyncModelExecutionEnabled(), + options.getNoOfArmnnThreads())); // Run a single 'dummy' inference of the model. This means that CL kernels will get compiled (and tuned if // this is enabled) before the first 'real' inference which removes the overhead of the first inference. -- cgit v1.2.1