Age | Commit message (Collapse) | Author |
|
* Change to check for success instead of specific failure
* Fix which map index is used when assigning outputs
Signed-off-by: Ryan OShea <ryan.oshea3@arm.com>
Change-Id: I13d8e989e35789ad3e2465d595905c5a5603ae0f
|
|
* Enabled using same instance of SimpleLogger
* Removed some trailing new lines on some log messages
Signed-off-by: Sadik Armagan <sadik.armagan@arm.com>
Change-Id: I4b917c0ca5011afc9b39dad50715290ba15a1246
|
|
Signed-off-by: Finn Williams <finn.williams@arm.com>
Change-Id: Ic5ebf7b80468b7751c234c43a90ec4cbf4c59ffe
|
|
Enforce serialized access to TimelineDecoder::m_Model by removing public
access funtion and replacing with an 'Apply' method taking a lambda and
uses a std::lock.
Use the new lambda when invoking callbacks.
Change-Id: I6ea2fbca990736f3be63e80897f175421f19f0c1
Signed-off-by: Matthew Bentham <matthew.bentham@arm.com>
|
|
* Created individual IRuntime sharedptr in ExecuteNetwork main() each time
MainImpl() is called. Prevents additional runtime being created when the
delegate is used.
Signed-off-by: Cathal Corbett <cathal.corbett@arm.com>
Change-Id: Ia4b508fbf2bbd25467c6235fed2f05662a7aecc0
|
|
* Adds ExecuteNetwork when building the delegate only
* Adds timings to delegate subgraph creation
* Adds executions times
Signed-off-by: Jan Eilers <jan.eilers@arm.com>
Change-Id: Ieff2f67ea8dbb6c2a708f8810e84a20485b7a631
|
|
Deep_speech models
* Fixed output bindings in ExecuteNetwork when using delegate
* Added extra comments (for my own comprehension)
Change-Id: Ia1330903a2b724be7d637f6c5b6e4349e7d33e2e
Signed-off-by: Tamas Nyiri <tamas.nyiri@arm.com>
|
|
to comply with inclusive language guidelines.
Signed-off-by: Teresa Charlin <teresa.charlinreyes@arm.com>
Change-Id: Iff3c9fd1107ddd97b1547f676e6aec06d7e2d28a
|
|
* add fsrcnn and mobilebert memory profiles to the strategy benchmark
Signed-off-by: Finn Williams <finn.williams@arm.com>
Change-Id: Ibd8b26f2153c561e5c5bec477f6246d0e8ffa4af
|
|
This reverts commit 2d9956162dd002a41f7fb4fa6753195d33524c7f.
Reason for revert: After some discussion, this does technically implement Float16 support for ExecuteNetwork, but not in a way which matches most use cases and is likely to cause issues in the future. Reverting for now.
Change-Id: I4ce6de6879216e694631f5dc68e46fb793fae0a9
|
|
* Allows the user to specify float16 as a datatype
* Does not contain support for float16 on the TfLiteDelegate via
ExecuteNetwork
Signed-off-by: David Monahan <David.Monahan@arm.com>
Change-Id: Icba56feedab32662e2cf671cc46ada899cf40c6c
|
|
* In ExecuteNetwork MainImpl compare the data types of outputs on the
loaded model with those specified by the user through --output-type.
Issue a warning if there is a mismatch.
Signed-off-by: Colm Donelan <colm.donelan@arm.com>
Change-Id: Ic5add9734dc239eddca0972a9e560e54abdb1093
|
|
* Move TContainer to armnnUtils library
Signed-off-by: Francis Murtagh <francis.murtagh@arm.com>
Change-Id: I3c0f895d11b66f6ee224ac689a19d0477f990b98
|
|
* Pass through the value of m_EnableProfiling from Executenetwork to
DelegateOptions.
* If internal profiling is enabled print it out from inside the delegate.
* Remove an unnecessary ProfilerImpl instance from WorkingMemhandle.hpp
* Remove an unnecessary parameter from TfLiteDelegateMainImpl in
ExecuteNetwork.
Signed-off-by: Colm Donelan <colm.donelan@arm.com>
Change-Id: Ia1d4b1eb3a05ca5b4d80cc39e138c7fac182d948
|
|
* Refactor the strategy library to be more generic
* Shorten the names of the current strategies
* Change validatorStrat to throw exceptions
Change-Id: I0d9c9ef609b2d8675e5788610d1accac6767c660
Signed-off-by: Finn Williams <finwil01@e127804.cambridge.arm.com>
Signed-off-by: Jim Flynn <jim.flynn@arm.com>
|
|
* Added a single definition of TContainer to include/armnn/Utils.hpp
* Change all files which contained their own identical definitions of TContainer to use the new one
Signed-off-by: David Monahan <David.Monahan@arm.com>
Change-Id: I63e633693a430bbbd6a29001cafa19742ef8309a
|
|
* Introduce a mechanism to construct a DelegateOptions from an
ExecuteNetworkParams.
* Modify ExecuteNetwork to use this constructed DelegateOptions.
Signed-off-by: Colm Donelan <Colm.Donelan@arm.com>
Change-Id: Ied663a1e00ac3eece42244ed313ddafd6d2ce078
|
|
Signed-off-by: Narumol Prangnawarat <narumol.prangnawarat@arm.com>
Change-Id: Ia7dbf0735619d406d6b4e34a71f14f20d92586e6
|
|
Signed-off-by: Derek Lamberti <derek.lamberti@arm.com>
Change-Id: I04fb80c967bba4bb377de419bde618c1cbb80075
|
|
Signed-off-by: Teresa Charlin Reyes <tercha01@e127807.cambridge.arm.com>
Change-Id: I0e9f5d0386613f29601db554db29d15b6bd0fff4
|
|
Signed-off-by: Jan Eilers <jan.eilers@arm.com>
Change-Id: I01aa6ea4b21ad4504d6ae04850a2236588e5ddd3
|
|
Signed-off-by: Jan Eilers <jan.eilers@arm.com>
Change-Id: I10fb010ee8d3f813d2264cefb526f352e30d7046
|
|
* Fixed Segfault when optional input tensor shapes aren't provided
Signed-off-by: Mike Kelly <mike.kelly@arm.com>
Change-Id: Ib18be62d3654020037e6101976d58297b2a983c4
|
|
* Add functionality to only output network details in ExNet
Signed-off-by: Keith Davis <keith.davis@arm.com>
Change-Id: I0c45e67193f308ce7b86f1bb1a918a266fefba2e
|
|
* Add switch for network details during profiling
Signed-off-by: Keith Davis <keith.davis@arm.com>
Change-Id: I8bd49fd58f0e0255598106e9ab36806ee78391d6
|
|
* Pass m_shapeInferenceMethod to OptimizerOptions in ExecuteNetwork
Signed-off-by: Finn Williams <Finn.Williams@arm.com>
Change-Id: I90280fb7629092d3b66e8a3968ca9e35a0df854a
|
|
* Add new ProfilingDetails class to construct operator details string
* Add new macro which helps append layer details to ostream
* Add ProfilingEnabled to NetworkProperties so that profiling can be
realised when loading the network
* Add further optional info to WorkloadInfo specific to convolutions
* Generalise some JsonPrinter functions into JsonUtils for reusability
* Remove explicit enabling of profiling within InferenceModel as it is
done when loading network
* Add ProfilingDetails macros to ConvolutionWorkloads for validation
Signed-off-by: Keith Davis <keith.davis@arm.com>
Change-Id: Ie84bc7dc667e72e6bcb635544f9ead7af1765690
|
|
* Allows to supply different input data for every execution of a
model when using the 'iterations' option in ExecuteNetwork
* Removes the option 'simultaneous-iterations' because it's functionallity
is now covered by 'iterations'
* Adds a deprecation warning message to notify users
* Little refactor of warning messages
Signed-off-by: Jan Eilers <jan.eilers@arm.com>
Change-Id: Ib3ab0d6533f6952bfee20d098a890b653c34cc12
|
|
* Added support for qasymms8 (int8) to ImageTensorGenerator
* Added qasymmu8 as alias for qasymm8 in ImageTensorGenerator
* Added support for qasymms8 (int8) to ExecuteNetwork
* Added qasymmu8 as alias for qasymm8 in ExecuteNetwork
* Set tflite to be the default model format in ImageTensorGenerator as
it's the only supported model format.
Signed-off-by: Mike Kelly <mike.kelly@arm.com>
Change-Id: Ieda7b78e668ea390e3565cd65a41fe0a9c8a5b83
|
|
!android-nn-driver:5966
Change-Id: Ice0b4d2872bb0e09bfc0763034a206c3a8f24af4
Signed-off-by: Rob Hughes <robert.hughes@arm.com>
|
|
!android-nn-driver:5802
* Extract the threadpool from LoadedNetwork/Runtime
* Refactor the threadpool to be handle multiple networks
* Trim IAsyncExecutionCallback and add an InferenceId to AsyncExecutionCallback
* Add AsyncCallbackManager class
Signed-off-by: Finn Williams <Finn.Williams@arm.com>
Change-Id: I36aa2ad29c16bc10ee0706adfeb6b27f60012afb
|
|
profiling/common
Signed-off-by: Nikhil Raj <nikhil.raj@arm.com>
Change-Id: I83de0fe9c83ec7e698817095ce5cfe3319c36ea6
|
|
* Used doctest in ArmNN unit tests
Signed-off-by: Sadik Armagan <sadik.armagan@arm.com>
Change-Id: Ia9cf5fc72775878885c5f864abf2c56b3a935f1a
|
|
profiling common
!android-nn-driver:5691
Signed-off-by: Nikhil Raj <nikhil.raj@arm.com>
Change-Id: Ib71af0831e324ac6bd27b1a36f4a6ec1a703b14a
|
|
* Allow the user to use create a tread pool with a single thread
* This is in keeping with how the android-nn-driver was implemented
* Add it to ExecuteNetwork thread pool creation
Signed-off-by: Kevin May <kevin.may@arm.com>
Change-Id: I05b8048a9e0e45ae11d2b585080af28d9d008d81
|
|
* Remove concurrent flag from ExecuteNetwork as it is possible
to deduce if SimultaneousIterations > 1
* Add void RunAsync()
* Refactor some unit tests
Change-Id: I7021d4821b0e460470908294cbd9462850e8b361
Signed-off-by: Keith Davis <keith.davis@arm.com>
Signed-off-by: Kevin May <kevin.may@arm.com>
|
|
Signed-off-by: Colm Donelan <Colm.Donelan@arm.com>
Change-Id: Ic0a2e62e3808caf89ee429ec184d947a04340248
|
|
TensorHandleFactoryRegistry::GetFactory
* Modify Layer::CreateTensorHandles to include MemorySource
* Modify INetworkProperties to add MemorySource
* Disable Neon/Cl fallback tests until full import implementation complete
Change-Id: Ia4fff6ea3d4bf6afca33aae358125ccaec7f9a38
Signed-off-by: Francis Murtagh <francis.murtagh@arm.com>
|
|
* Remove from .md files and Doxygen
* Remove from armnn/docker build
* Remove Tensorflow model format from ExecuteNetworkParams
* Remove Tensorflow model format from ImageTensorGenerator
Signed-off-by: Kevin May <kevin.may@arm.com>
Change-Id: Id6ed4a7d90366c396e8e0395d0ce43a3bcddcee6
|
|
* Enabled async mode with '-n, concurrent' and 'simultaneous-iterations'
in ExecuteNetwork
* Number of input files provided should be equal to number of input files
provided multiply by number of simultaneous iterations divided by comma
!armnn:5443
Signed-off-by: Sadik Armagan <sadik.armagan@arm.com>
Change-Id: Ibeb318010430bf4ae61a02b18b1bf88f3657774c
|
|
Signed-off-by: Nikhil Raj <nikhil.raj@arm.com>
Change-Id: Ida37d3ee3a1af0c75aa905199bd861726c646846
|
|
Signed-off-by: Kevin May <kevin.may@arm.com>
Change-Id: I8a29df03afdd6f1cc8413b219e179272cd2d51cf
|
|
Signed-off-by: Nikhil Raj <nikhil.raj@arm.com>
Change-Id: Ib00be204f549efa9aa5971ecf65c2dec4a10b10f
|
|
* Change layer storage of ConstTensors to std::shared_ptr<ConstCpuTensorHandle>
* Change clone to share ConstTensor rather than copy
* Remove uses of non-const GetTensor() call
* Reduce scope of non-optimized network in ExeNet, so memory can be released after use
Signed-off-by: Finn Williams <Finn.Williams@arm.com>
Change-Id: Ibb2c7309d12411d21405bd6024c76bcdf5404545
|
|
* Enable GPU profiling on ExecuteNetwork
Signed-off-by: Sadik Armagan <sadik.armagan@arm.com>
Change-Id: I57bb4eeb45674e5218fce7e67b9bddf16ba0894d
|
|
abs is int type, use fabsf instead for float type
Signed-off-by: Keith Mok <ek9852@gmail.com>
Change-Id: I157d471da25d66cbe39fa9809ff9ee536d4e95a1
|
|
Instead of using the predefined backend ids,
we should just pass the backend string directly to the optimator,
otherwise we cannot use vendor specific backends or GpuAcc
Signed-off-by: Keith Mok <ek9852@gmail.com>
Change-Id: Ic52c81d48364a19f0fcc59bc3c975cf3bd740e21
|
|
* fix delegate perchannel quantization
* change delegate to check reshape options before inputs
* Add int8 "qsymms8" option to ExecuteNetwork
* Add option to run ExecuteNetwork on tflite w/o delegate
!referencetests:301301
Signed-off-by: Finn Williams <Finn.Williams@arm.com>
Change-Id: If3e12599b17aff1199d7ab0a55e1c901e480083d
|
|
The input tensor data pointer was passed incorrectly
Signed-off-by: Keith Mok <ek9852@gmail.com>
Change-Id: I03f872c57ec588fde0f7d444c80b38823ea4f9b4
|
|
inference time.
Usage: armnn_tfl_benchmark -m <model .tflite>
-m --model_file <.tflite file path>: .tflite model to be executed
-b --backend <device>: preferred backend device to run
layers on by default. Possible
choices: CpuAcc, CpuRef
-l --loops <int>: provide the number of time the
inference will be executed
(by default nb_loops=1)
Signed-off-by: Vincent ABRIOU <vincent.abriou@st.com>
Signed-off-by: Jim Flynn <jim.flynn@arm.com>
Change-Id: Ia26fafd4f382f0ad03856436dcae6e71b5abbd26
|