diff options
author | Georgios Pinitas <georgios.pinitas@arm.com> | 2021-07-23 20:38:47 +0100 |
---|---|---|
committer | Georgios Pinitas <georgios.pinitas@arm.com> | 2021-07-26 08:41:55 +0000 |
commit | e92c23eb7a91ddd12feeb40cd8bc5d766c6fe5c3 (patch) | |
tree | 9d9ed4ef73afb88d5951126675ad1bdd405b3469 /src/runtime/gpu/cl/operators | |
parent | 7075fe2c5ee6f7cfe7cfd9454d905235e70b9ac4 (diff) | |
download | ComputeLibrary-e92c23eb7a91ddd12feeb40cd8bc5d766c6fe5c3.tar.gz |
Fix allocation of prepare tensor on ClWinogradConv2d
Preparation tensor was allocated during execution. Avoid if GEMM used by
the Winograd function is reshaped.
Resolves: COMPMID-4640
Signed-off-by: Georgios Pinitas <georgios.pinitas@arm.com>
Change-Id: Id5b7e6bbfe9f69661baa9d42698cdc2cc1b422b2
Reviewed-on: https://review.mlplatform.org/c/ml/ComputeLibrary/+/5991
Tested-by: Arm Jenkins <bsgcomp@arm.com>
Reviewed-by: Michele Di Giorgio <michele.digiorgio@arm.com>
Comments-Addressed: Arm Jenkins <bsgcomp@arm.com>
Diffstat (limited to 'src/runtime/gpu/cl/operators')
-rw-r--r-- | src/runtime/gpu/cl/operators/ClWinogradConv2d.cpp | 19 |
1 files changed, 7 insertions, 12 deletions
diff --git a/src/runtime/gpu/cl/operators/ClWinogradConv2d.cpp b/src/runtime/gpu/cl/operators/ClWinogradConv2d.cpp index 2ca1ff59df..07f90ddaef 100644 --- a/src/runtime/gpu/cl/operators/ClWinogradConv2d.cpp +++ b/src/runtime/gpu/cl/operators/ClWinogradConv2d.cpp @@ -233,37 +233,32 @@ Status ClWinogradConv2d::validate(const ITensorInfo *src, const ITensorInfo *wei void ClWinogradConv2d::run(ITensorPack &tensors) { - prepare(tensors); + const bool is_gemm_reshaped = _aux_mem[3].lifetime == MemoryLifetime::Prepare; auto src = utils::cast::polymorphic_downcast<const ICLTensor *>(tensors.get_const_tensor(TensorType::ACL_SRC_0)); auto biases = utils::cast::polymorphic_downcast<const ICLTensor *>(tensors.get_const_tensor(TensorType::ACL_SRC_2)); auto dst = utils::cast::polymorphic_downcast<ICLTensor *>(tensors.get_tensor(TensorType::ACL_DST)); CLAuxTensorHandler input0(offset_int_vec(2), _input0, tensors, true); - CLAuxTensorHandler input1(offset_int_vec(3), _input1, tensors, true); + CLAuxTensorHandler input1(offset_int_vec(3), _input1, tensors, true, is_gemm_reshaped); CLAuxTensorHandler batched_mm_output(offset_int_vec(4), _batched_mm_output, tensors, true); + prepare(tensors); + // Run input transform ITensorPack pack_it { { TensorType::ACL_SRC, src }, { TensorType::ACL_DST, input0.get() }, }; - CLScheduler::get().enqueue_op(_border_handler, pack_it); - CLScheduler::get().enqueue_op(*_input_transform, pack_it); + CLScheduler::get().enqueue_op(_border_handler, pack_it, false); + CLScheduler::get().enqueue_op(*_input_transform, pack_it, false); // Run batched matrix multiplication ITensorPack pack_mm = tensors; pack_mm.add_const_tensor(TensorType::ACL_SRC_0, input0.get()); pack_mm.add_tensor(TensorType::ACL_DST, batched_mm_output.get()); - if(_aux_mem[3].lifetime == MemoryLifetime::Prepare) - { - pack_mm.remove_tensor(TensorType::ACL_SRC_1); - } - else - { - pack_mm.add_const_tensor(TensorType::ACL_SRC_1, input1.get()); - } + is_gemm_reshaped ? pack_mm.remove_tensor(TensorType::ACL_SRC_1) : pack_mm.add_const_tensor(TensorType::ACL_SRC_1, input1.get()); _batched_mm.run(pack_mm); // Run output transform |