17 #define ARMNN_SCOPED_PROFILING_EVENT_NEON(name) \ 18 ARMNN_SCOPED_PROFILING_EVENT_WITH_INSTRUMENTS(armnn::Compute::CpuAcc, \ 21 armnn::WallClockTimer()) 31 InitialiseArmComputeTensorEmpty(dstTensor);
32 CopyArmComputeITensorData(srcData, dstTensor);
62 BOOST_ASSERT_MSG(
false,
"Unexpected tensor type.");
67 const std::vector<int>& m_end,
68 const std::vector<int>& m_stride)
74 unsigned int num_dims =
static_cast<unsigned int>(m_begin.size());
76 for (
unsigned int i = 0; i < num_dims; i++)
78 unsigned int revertedIndex = num_dims - i - 1;
80 starts.set(i, static_cast<int>(m_begin[revertedIndex]));
81 ends.set(i, static_cast<int>(m_end[revertedIndex]));
82 strides.set(i, static_cast<int>(m_stride[revertedIndex]));
85 return std::make_tuple(starts, ends, strides);
89 const std::vector<unsigned int>& m_size)
96 unsigned int num_dims =
static_cast<unsigned int>(m_begin.size());
101 for (
unsigned int i = 0; i < num_dims; i++)
103 unsigned int revertedIndex = num_dims - i - 1;
105 starts.set(i, static_cast<int>(m_begin[revertedIndex]));
106 ends.set(i, static_cast<int>(m_begin[revertedIndex] + m_size[revertedIndex]));
109 return std::make_tuple(starts, ends);
#define ARMNN_NO_DEPRECATE_WARN_BEGIN
std::array< unsigned int, MaxNumOfTensorDimensions > Coordinates
const T * GetConstTensor() const
Copyright (c) 2020 ARM Limited.
auto SetNeonStridedSliceData(const std::vector< int > &m_begin, const std::vector< int > &m_end, const std::vector< int > &m_stride)
#define ARMNN_NO_DEPRECATE_WARN_END
DataType GetDataType() const
auto SetNeonSliceData(const std::vector< unsigned int > &m_begin, const std::vector< unsigned int > &m_size)
#define ARMNN_FALLTHROUGH
void InitializeArmComputeTensorData(arm_compute::Tensor &tensor, const ConstCpuTensorHandle *handle)
void CopyArmComputeTensorData(arm_compute::Tensor &dstTensor, const T *srcData)
const TensorInfo & GetTensorInfo() const