diff options
Diffstat (limited to 'src/core/NEON/kernels/assembly')
-rw-r--r-- | src/core/NEON/kernels/assembly/INEGEMMWrapperKernel.cpp | 80 | ||||
-rw-r--r-- | src/core/NEON/kernels/assembly/NEGEMMNativeWrapperKernel.cpp | 119 |
2 files changed, 199 insertions, 0 deletions
diff --git a/src/core/NEON/kernels/assembly/INEGEMMWrapperKernel.cpp b/src/core/NEON/kernels/assembly/INEGEMMWrapperKernel.cpp new file mode 100644 index 0000000000..0fc3610014 --- /dev/null +++ b/src/core/NEON/kernels/assembly/INEGEMMWrapperKernel.cpp @@ -0,0 +1,80 @@ +/* + * Copyright (c) 2018 ARM Limited. + * + * SPDX-License-Identifier: MIT + * + * Permission is hereby granted, free of charge, to any person obtaining a copy + * of this software and associated documentation files (the "Software"), to + * deal in the Software without restriction, including without limitation the + * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or + * sell copies of the Software, and to permit persons to whom the Software is + * furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice shall be included in all + * copies or substantial portions of the Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE + * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE + * SOFTWARE. + */ + +#include "arm_compute/core/NEON/kernels/assembly/INEGEMMWrapperKernel.h" + +#include "arm_compute/core/Error.h" +#include "arm_compute/core/Helpers.h" +#include "arm_compute/core/ITensor.h" +#include "arm_compute/core/Validate.h" +#include "arm_compute/core/WindowIterator.h" + +using namespace arm_compute; + +INEGEMMWrapperKernel::INEGEMMWrapperKernel() + : _a(nullptr), _b(nullptr), _c(nullptr), _params(), _window3d(), _window_shape() +{ +} + +INEGEMMWrapperKernel::Params INEGEMMWrapperKernel::extract_parameters(const ITensor *a, const ITensor *b, const ITensor *c) +{ + Params p; + + ARM_COMPUTE_ERROR_ON_NULLPTR(a); + ARM_COMPUTE_ERROR_ON_NULLPTR(b); + ARM_COMPUTE_ERROR_ON_NULLPTR(c); + + p.M = c->info()->tensor_shape().y(); + p.N = c->info()->tensor_shape().x(); + p.K = a->info()->tensor_shape().x(); + p.multis = b->info()->tensor_shape().z(); + p.batches = c->info()->tensor_shape().total_size_upper(2) / p.multis; //COMPMID-1423: Agree on and document the layout of gemm inputs/outputs + + return p; +} + +void INEGEMMWrapperKernel::configure(const ITensor *a, const ITensor *b, ITensor *c, float alpha, float beta) +{ + _params = extract_parameters(a, b, c); + _a = a; + _b = b; + _c = c; + + _window3d = configure_internal(alpha, beta); + _window_shape = _window3d.shape(); + + // Convert the 3D window into a 1D window in order to allow the scheduler to arbitrary split it. + Window collapsed; + collapsed.set(0, Window::Dimension(0, _window3d.num_iterations_total())); + + INEKernel::configure(collapsed); +} + +void INEGEMMWrapperKernel::run(const Window &window, const ThreadInfo &info) +{ + const Coordinates start_offset = index2coords(_window_shape, window.x().start()); + const Coordinates end_offset = index2coords(_window_shape, window.x().end() - 1); + + run_internal(_window3d, start_offset, end_offset, info); +} diff --git a/src/core/NEON/kernels/assembly/NEGEMMNativeWrapperKernel.cpp b/src/core/NEON/kernels/assembly/NEGEMMNativeWrapperKernel.cpp new file mode 100644 index 0000000000..38eb40a06e --- /dev/null +++ b/src/core/NEON/kernels/assembly/NEGEMMNativeWrapperKernel.cpp @@ -0,0 +1,119 @@ +/* + * Copyright (c) 2018 ARM Limited. + * + * SPDX-License-Identifier: MIT + * + * Permission is hereby granted, free of charge, to any person obtaining a copy + * of this software and associated documentation files (the "Software"), to + * deal in the Software without restriction, including without limitation the + * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or + * sell copies of the Software, and to permit persons to whom the Software is + * furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice shall be included in all + * copies or substantial portions of the Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE + * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE + * SOFTWARE. + */ + +#include "arm_compute/core/NEON/kernels/assembly/NEGEMMNativeWrapperKernel.h" + +#include "arm_compute/core/ITensor.h" +#include "arm_compute/core/Utils.h" +#include "arm_compute/core/WindowIterator.h" + +#include "../arm_gemm/utils.hpp" +#include "arm_gemm.hpp" + +#include "../arm_gemm/mergeresults.hpp" +#include "../arm_gemm/transform.hpp" + +#include "../arm_gemm/kernels/a32_sgemm_8x6.hpp" +#include "../arm_gemm/kernels/a64_sgemm_12x8.hpp" +#include "../arm_gemm/kernels/a64_sgemm_native_16x4.hpp" +#include "../arm_gemm/kernels/a64_sgemv_pretransposed.hpp" +#include "../arm_gemm/kernels/a64_sgemv_trans.hpp" + +namespace arm_compute +{ +template <typename To, typename Tr> +struct Kernel +{ +}; + +#ifdef __aarch64__ +template <> +struct Kernel<float, float> +{ + using strategy = arm_gemm::sgemm_native_16x4; +}; +#endif /* __aarch64__ */ + +template <typename To, typename Tr> +Window NEGEMMNativeWrapperKernel<To, Tr>::configure_internal(float alpha, float beta) +{ + using strategy = typename Kernel<To, Tr>::strategy; + + _beta = beta; + + //Note: The window is shifted down by 1 dimension compare to the tensors + Window window; + window.set(Window::DimX, Window::Dimension(0, ceil_to_multiple(_params.M, strategy::out_height()), strategy::out_height())); + window.set(Window::DimY, Window::Dimension(0, _params.batches)); + window.set(Window::DimZ, Window::Dimension(0, _params.multis)); + + return window; +} + +template <typename To, typename Tr> +void NEGEMMNativeWrapperKernel<To, Tr>::run_internal(const Window &window, const Coordinates &start_offset, const Coordinates &end_offset, const ThreadInfo &info) +{ + using strategy = typename Kernel<To, Tr>::strategy; + + TensorAccessor<To> a(*_a); + TensorAccessor<To> b(*_b); + TensorAccessor<Tr> c(*_c); + + if(_a->info()->data_layout() == DataLayout::NHWC) + { + // In the case of NHWC we want to interpret the output shape as 3D. Thus, the batch stride for A is + // the relevant multiple of the row stride. + const size_t nhwc_batch_stride = _a->info()->strides_in_bytes().y() * _c->info()->dimension(1); + a.set_stride(2, nhwc_batch_stride); + } + + unsigned int m_end = 0; + + strategy strat(info.cpu_info); + auto window_iterator = arm_compute::create_window_iterator(window, start_offset, end_offset, [&](const Coordinates & id) + { + const unsigned int y0 = id.x(); + const unsigned int batch = id.y(); + const unsigned int multi = id.z(); + const unsigned int ymax = std::min(y0 + strategy::out_height(), m_end); + + strat.kernel(a(0, y0, batch, multi), a.stride(Window::DimY), + b(0, 0, 0, multi), b.stride(Window::DimY), + c(0, y0, batch, multi), c.stride(Window::DimY), + _beta, (ymax - y0), _params.N, _params.K); + }); + + auto on_new_row_size = [&](unsigned int start, unsigned int end) + { + m_end = std::min(end, _params.M); + }; + + window_iterator.iterate_3D(on_new_row_size); +} + +#ifdef __aarch64__ +template class NEGEMMNativeWrapperKernel<float, float>; +#endif /* __aarch64__ */ + +} // namespace arm_compute |