ArmNN
 21.05
Runtime.hpp
Go to the documentation of this file.
1 //
2 // Copyright © 2017 Arm Ltd and Contributors. All rights reserved.
3 // SPDX-License-Identifier: MIT
4 //
5 #pragma once
6 
7 #include "LoadedNetwork.hpp"
8 #include "DeviceSpec.hpp"
9 
10 #include <armnn/INetwork.hpp>
11 #include <armnn/IRuntime.hpp>
12 #include <armnn/Tensor.hpp>
13 #include <armnn/BackendId.hpp>
14 
16 
17 #include <ProfilingService.hpp>
18 
19 #include <IProfilingService.hpp>
20 #include <IReportStructure.hpp>
21 
22 #include <mutex>
23 #include <unordered_map>
24 
25 namespace armnn
26 {
27 using LoadedNetworks = std::unordered_map<NetworkId, std::unique_ptr<LoadedNetwork>>;
29 
30 struct RuntimeImpl final : public IReportStructure
31 {
32 public:
33  /// Loads a complete network into the Runtime.
34  /// @param [out] networkIdOut - Unique identifier for the network is returned in this reference.
35  /// @param [in] network - Complete network to load into the Runtime.
36  /// The runtime takes ownership of the network once passed in.
37  /// @return armnn::Status
38  Status LoadNetwork(NetworkId& networkIdOut, IOptimizedNetworkPtr network);
39 
40  /// Load a complete network into the IRuntime.
41  /// @param [out] networkIdOut Unique identifier for the network is returned in this reference.
42  /// @param [in] network Complete network to load into the IRuntime.
43  /// @param [out] errorMessage Error message if there were any errors.
44  /// The runtime takes ownership of the network once passed in.
45  /// @return armnn::Status
46  Status LoadNetwork(NetworkId& networkIdOut,
47  IOptimizedNetworkPtr network,
48  std::string& errorMessage);
49 
50  Status LoadNetwork(NetworkId& networkIdOut,
51  IOptimizedNetworkPtr network,
52  std::string& errorMessage,
53  const INetworkProperties& networkProperties);
54 
55  TensorInfo GetInputTensorInfo(NetworkId networkId, LayerBindingId layerId) const;
56  TensorInfo GetOutputTensorInfo(NetworkId networkId, LayerBindingId layerId) const;
57 
58  // Evaluates network using input in inputTensors, outputs filled into outputTensors.
60  const InputTensors& inputTensors,
61  const OutputTensors& outputTensors);
62 
63  /// This is an experimental function.
64  /// Schedule a thread safe execution by taking the input tensors and an execution priority for Quality of Service.
65  /// The output tensors will then be filled and the callback object will notify that the execution has either
66  /// succeeded or failed.
67  void Schedule(NetworkId networkId,
68  const InputTensors& inputTensors,
69  const OutputTensors& outputTensors,
70  const QosExecPriority priority,
71  std::shared_ptr<IAsyncExecutionCallback> callback);
72 
73  /// This is an experimental function.
74  /// Evaluates a network using input in inputTensors and outputs filled into outputTensors.
75  /// This function performs a thread safe execution of the network. Returns once execution is complete.
76  /// Will block until this and any other thread using the same workingMem object completes.
77  Status Execute(IWorkingMemHandle& workingMemHandle,
78  const InputTensors& inputTensors,
79  const OutputTensors& outputTensors);
80 
81  /// Unloads a network from the Runtime.
82  /// At the moment this only removes the network from the m_Impl->m_Network.
83  /// This might need more work in the future to be AndroidNN compliant.
84  /// @param [in] networkId Unique identifier for the network to be unloaded. Generated in LoadNetwork().
85  /// @return armnn::Status
86  Status UnloadNetwork(NetworkId networkId);
87 
88  const IDeviceSpec& GetDeviceSpec() const { return m_DeviceSpec; }
89 
90  /// Gets the profiler corresponding to the given network id.
91  /// @param networkId The id of the network for which to get the profile.
92  /// @return A pointer to the requested profiler, or nullptr if not found.
93  const std::shared_ptr<IProfiler> GetProfiler(NetworkId networkId) const;
94 
95  /// Create a new unique WorkingMemHandle object. Create multiple handles if you wish to have
96  /// overlapped Execution by calling this function from different threads.
97  std::unique_ptr<IWorkingMemHandle> CreateWorkingMemHandle(NetworkId networkId);
98 
99  /// Registers a callback function to debug layers performing custom computations on intermediate tensors.
100  /// @param networkId The id of the network to register the callback.
101  /// @param func callback function to pass to the debug layer.
102  void RegisterDebugCallback(NetworkId networkId, const DebugCallbackFunction& func);
103 
104  /// Creates a runtime for workload execution.
105  RuntimeImpl(const IRuntime::CreationOptions& options);
106 
107  ~RuntimeImpl();
108 
109  //NOTE: we won't need the profiling service reference but it is good to pass the service
110  // in this way to facilitate other implementations down the road
111  void ReportStructure();
112 
113 private:
114  friend void RuntimeLoadedNetworksReserve(RuntimeImpl* runtime); // See RuntimeTests.cpp
115 
116  friend profiling::ProfilingService& GetProfilingService(RuntimeImpl* runtime); // See RuntimeTests.cpp
117 
118  int GenerateNetworkId();
119 
120  LoadedNetwork* GetLoadedNetworkPtr(NetworkId networkId) const;
121 
122  template<typename Func>
123  void LoadedNetworkFuncSafe(NetworkId networkId, Func f)
124  {
125  std::lock_guard<std::mutex> lockGuard(m_Mutex);
126  auto iter = m_LoadedNetworks.find(networkId);
127  if (iter != m_LoadedNetworks.end())
128  {
129  f(iter->second.get());
130  }
131  }
132 
133  /// Loads any available/compatible dynamic backend in the runtime.
134  void LoadDynamicBackends(const std::string& overrideBackendPath);
135 
136  mutable std::mutex m_Mutex;
137 
138  /// Map of Loaded Networks with associated GUID as key
139  LoadedNetworks m_LoadedNetworks;
140 
141  std::unordered_map<BackendId, IBackendInternal::IBackendContextPtr> m_BackendContexts;
142 
143  int m_NetworkIdCounter;
144 
145  DeviceSpec m_DeviceSpec;
146 
147  /// List of dynamic backends loaded in the runtime
148  std::vector<DynamicBackendPtr> m_DynamicBackends;
149 
150  /// Profiling Service Instance
151  profiling::ProfilingService m_ProfilingService;
152 };
153 
154 } // namespace armnn
TensorInfo GetInputTensorInfo(NetworkId networkId, LayerBindingId layerId) const
Definition: Runtime.cpp:394
Status UnloadNetwork(NetworkId networkId)
Unloads a network from the Runtime.
Definition: Runtime.cpp:190
Status LoadNetwork(NetworkId &networkIdOut, IOptimizedNetworkPtr network)
Loads a complete network into the Runtime.
Definition: Runtime.cpp:130
std::vector< std::pair< LayerBindingId, class ConstTensor > > InputTensors
Definition: Tensor.hpp:340
Status Execute(IWorkingMemHandle &workingMemHandle, const InputTensors &inputTensors, const OutputTensors &outputTensors)
This is an experimental function.
Definition: Runtime.cpp:438
TensorInfo GetOutputTensorInfo(NetworkId networkId, LayerBindingId layerId) const
Definition: Runtime.cpp:399
Copyright (c) 2021 ARM Limited and Contributors.
Status EnqueueWorkload(NetworkId networkId, const InputTensors &inputTensors, const OutputTensors &outputTensors)
Definition: Runtime.cpp:405
profiling::IReportStructure IReportStructure
Definition: Runtime.hpp:28
void RegisterDebugCallback(NetworkId networkId, const DebugCallbackFunction &func)
Registers a callback function to debug layers performing custom computations on intermediate tensors...
Definition: Runtime.cpp:523
const std::shared_ptr< IProfiler > GetProfiler(NetworkId networkId) const
Gets the profiler corresponding to the given network id.
Definition: Runtime.cpp:242
std::function< void(LayerGuid guid, unsigned int slotIndex, ITensorHandle *tensorHandle)> DebugCallbackFunction
Define the type of callback for the Debug layer to call.
Definition: Types.hpp:316
int LayerBindingId
Type of identifiers for bindable layers (inputs, outputs).
Definition: Types.hpp:243
friend profiling::ProfilingService & GetProfilingService(RuntimeImpl *runtime)
Definition: TestUtils.cpp:35
void Schedule(NetworkId networkId, const InputTensors &inputTensors, const OutputTensors &outputTensors, const QosExecPriority priority, std::shared_ptr< IAsyncExecutionCallback > callback)
This is an experimental function.
Definition: Runtime.cpp:462
void ReportStructure()
Definition: Runtime.cpp:254
const IDeviceSpec & GetDeviceSpec() const
Definition: Runtime.hpp:88
int NetworkId
Definition: IRuntime.hpp:22
std::vector< std::pair< LayerBindingId, class Tensor > > OutputTensors
Definition: Tensor.hpp:341
Status
enumeration
Definition: Types.hpp:30
std::unique_ptr< IOptimizedNetwork, void(*)(IOptimizedNetwork *network)> IOptimizedNetworkPtr
Definition: INetwork.hpp:174
Device specific knowledge to be passed to the optimizer.
Definition: Types.hpp:233
friend void RuntimeLoadedNetworksReserve(RuntimeImpl *runtime)
RuntimeImpl(const IRuntime::CreationOptions &options)
Creates a runtime for workload execution.
Definition: Runtime.cpp:269
QosExecPriority
Definition: Types.hpp:60
std::unordered_map< NetworkId, std::unique_ptr< LoadedNetwork > > LoadedNetworks
Definition: Runtime.hpp:27
std::unique_ptr< IWorkingMemHandle > CreateWorkingMemHandle(NetworkId networkId)
Create a new unique WorkingMemHandle object.
Definition: Runtime.cpp:492