Age | Commit message (Collapse) | Author |
|
* Enable fp_mixed_precision flag for the failing test case
Signed-off-by: Sadik Armagan <sadik.armagan@arm.com>
Change-Id: If13552165eb6598a84d213b82847b56a8c5f2783
|
|
* Pad value for QASYMM8 is no longer stored in quantized form.
Signed-off-by: Mike Kelly <mike.kelly@arm.com>
Change-Id: I048e1d233353c0560ae03a7cc1ed5199295352bc
|
|
Signed-off-by: Aron Virginas-Tar <Aron.Virginas-Tar@arm.com>
Change-Id: I3b021c0828d30298d99ddb211c9aae17fe3636f0
|
|
Change-Id: I0ab3339e8803a3e4e700d8fec9883eccc524b31e
|
|
* Unified ArmNN's weight format to [ M, I, H, W ] for the depthwise convolution
* Added conversion utilities to permute/reshape the weights as appropriate
when using CL and Neon backends
* Updated the reference implementation of the convolution
* Updated the relevant unit tests accordingly
!android-nn-driver:459
Change-Id: I07d0818efa9d1ca1e5dad82983aac1fe78eadb18
|
|
Factor out new BuildArmComputeReductionCoordinates function
from CL backend into ArmComputeTensorUtils.
Update NEON LayerSupport and WorkloadFactory objects
Change-Id: Icc975ec699199bffafbdb207323df509d35e1e04
|
|
!android-nn-driver:428
Change-Id: I42e59ad96d2c80f46b085182855d34b710a74dfe
|
|
This is one of the reasons why the tests in https://review.mlplatform.org/#/c/ml/armnn/+/237/
are failing (but not the only reason).
Change-Id: If485bade2a6dd013cba826cec71d748fc7747249
|
|
Change-Id: I663a0a0fccb43ee960ec070121a59df9db0bb04e
|
|
Change-Id: I400be8e7c0cc5a31eb9d2a7396da145d50d51b6e
|