22 template <
typename QueueDescriptor>
23 void AddInputToWorkload(QueueDescriptor& descriptor,
28 descriptor.m_Inputs.push_back(tensorHandle);
32 template <
typename QueueDescriptor>
33 void AddOutputToWorkload(QueueDescriptor& descriptor,
38 descriptor.m_Outputs.push_back(tensorHandle);
42 template <
typename QueueDescriptor>
43 void SetWorkloadInput(QueueDescriptor& descriptor,
49 descriptor.m_Inputs[index] = tensorHandle;
53 template <
typename QueueDescriptor>
54 void SetWorkloadOutput(QueueDescriptor& descriptor,
60 descriptor.m_Outputs[index] = tensorHandle;
66 bool memoryManagementRequested =
true)
68 const bool manageMemory = memoryManager && memoryManagementRequested;
73 memoryManager->Acquire();
85 memoryManager->Release();
96 switch(weightsType.
value())
109 BOOST_ASSERT_MSG(
false,
"GetBiasTypeFromWeightsType(): Unsupported data type.");
virtual void PostAllocationConfigure()=0
Copyright (c) 2020 ARM Limited.
armnn::Optional< armnn::DataType > GetBiasTypeFromWeightsType(armnn::Optional< armnn::DataType > weightsType)
std::vector< TensorInfo > m_InputTensorInfos
std::shared_ptr< IMemoryManager > IMemoryManagerSharedPtr
std::vector< TensorInfo > m_OutputTensorInfos
Workload interface to enqueue a layer computation.
EmptyOptional is used to initialize the Optional class in case we want to have default value for an O...
Contains information about inputs and outputs to a layer.
virtual void Execute() const =0