21.05
NeonConvertBf16ToFp32Workload.hpp
Go to the documentation of this file.
1
//
2
// Copyright © 2020 Arm Ltd. All rights reserved.
3
// SPDX-License-Identifier: MIT
4
//
5
6
#pragma once
7
8
#include <
backendsCommon/Workload.hpp
>
9
#include <
backendsCommon/WorkloadData.hpp
>
10
#include <
neon/workloads/NeonWorkloadUtils.hpp
>
11
12
namespace
armnn
13
{
14
15
class
NeonConvertBf16ToFp32Workload
:
public
BFloat16ToFloat32Workload
<ConvertBf16ToFp32QueueDescriptor>
16
{
17
public
:
18
NeonConvertBf16ToFp32Workload
(
const
ConvertBf16ToFp32QueueDescriptor
& descriptor,
const
WorkloadInfo
&
info
);
19
virtual
void
Execute
()
const override
;
20
21
private
:
22
using
TensorHandlePair = std::pair<const ITensorHandle*, ITensorHandle*>;
23
std::vector<TensorHandlePair> m_TensorHandlePairs;
24
};
25
26
}
//namespace armnn
WorkloadData.hpp
armnn::NeonConvertBf16ToFp32Workload::Execute
virtual void Execute() const override
Definition:
NeonConvertBf16ToFp32Workload.cpp:25
armnn::ConvertBf16ToFp32QueueDescriptor
Definition:
WorkloadData.hpp:444
armnn::NeonConvertBf16ToFp32Workload::NeonConvertBf16ToFp32Workload
NeonConvertBf16ToFp32Workload(const ConvertBf16ToFp32QueueDescriptor &descriptor, const WorkloadInfo &info)
Definition:
NeonConvertBf16ToFp32Workload.cpp:17
armnn
Copyright (c) 2021 ARM Limited and Contributors.
Definition:
01_00_software_tools.dox:6
armnn::NeonConvertBf16ToFp32Workload
Definition:
NeonConvertBf16ToFp32Workload.hpp:15
armnn::MultiTypedWorkload
Definition:
Workload.hpp:120
NeonWorkloadUtils.hpp
Workload.hpp
armnn::BoostLogSeverityMapping::info
armnn::WorkloadInfo
Contains information about inputs and outputs to a layer.
Definition:
WorkloadInfo.hpp:16
src
backends
neon
workloads
NeonConvertBf16ToFp32Workload.hpp
Generated on Mon May 10 2021 16:39:48 for ArmNN by
1.8.13