aboutsummaryrefslogtreecommitdiff
path: root/src/cpu
AgeCommit message (Expand)Author
2022-11-01Fix fixed-point quantized additionViet-Hoa Do
2022-11-01Updateable weights in depthwise convolutionMilos Puzovic
2022-11-01Add threshold for floating-point SOFT_RELU activationMilos Puzovic
2022-11-01Add check for Batch Matmul in GemmAssemblyDispatchMohammed Suhail Munshi
2022-10-27Fix fixed-point quantized additionViet-Hoa Do
2022-10-20Update reinterpret tensor as 1D for CPU addViet-Hoa Do
2022-10-20Add test in GEMMLowp for batch matmulMohammed Suhail Munshi
2022-10-19Fix FFTConvolutionLayer testViet-Hoa Do
2022-10-12Optimize Neon™ Logistic ActivationMohammed Suhail Munshi
2022-10-12Adding documentation section explaining how BF16 is usedRamy Elgammal
2022-10-10Fix LUT-based activation layerViet-Hoa Do
2022-10-07Optimize Neon™ SUB operator by squashing execution windowJakub Sujak
2022-10-03Fix Batch Matmul nightly failureAdnan AlSinan
2022-10-03Optimize CPU add layer on quantized dataViet-Hoa Do
2022-09-26Add FP32 Neon™ swish activationJonathan Deakin
2022-09-22Fix unresolved symbol for target armv7a + AndroidPablo Marquez Tello
2022-09-16Fix bug in QASYMM8_SIGNED to F32 cast layerViet-Hoa Do
2022-09-16Optimize Quantized/Integer Bilinear Scale for Neon™Gunes Bayir
2022-09-14Interpreting tensor as 1D for CPU multiplicationViet-Hoa Do
2022-09-14Adding GELU activationMurray Kornelsen
2022-09-14INT8 Quantized MeanStdDevNorm (LayerNorm)Murray Kornelsen
2022-09-12Add test for NEGEMM to test a batched matrix multiplication with variable inp...Adnan AlSinan
2022-09-09Optimize FP32/16 Bilinear Scale Kernel for Neon™Gunes Bayir
2022-09-08Disable Winograd on fp16 if fast-math = falseRamy Elgammal
2022-09-02F16 Specialization for MeanStdDevNormMurray Kornelsen
2022-08-24Fix add for tensors with non-matching stridesJonathan Deakin
2022-08-18Use Neon™ kernels for FP Bilinear Resize for SVEGunes Bayir
2022-08-17Add LUT for quantized sigmoid functionViet-Hoa Do
2022-08-08Fix for AI benchmark ResNet regressionViet-Hoa Do
2022-08-04[ONCPUML-970] Fast math mode for fixed format kernelsPablo Marquez Tello
2022-08-03[ONCPUML-968] Fixed format kernel support in additional APIsMilos Puzovic
2022-08-01Optimize add layer by considering the input tensors as 1D arrayGunes Bayir
2022-07-27Fix compilation error rasied in Nightly_NEWRamy Elgammal
2022-07-26Fix for inclusion of "arm_gemm" from src into "Types.h" from coreRamy Elgammal
2022-07-25Enable march=armv8.6-a in non multi-isa buildsPablo Marquez Tello
2022-07-19[ONCPUML-951] Variable weight support for Convolution.Francesco Petrogalli
2022-07-14Integrate new winograd APIs from MLTechramelg01
2022-07-04Fix build errors on armv8.6 SVE2 with NDK 23 and 24Michalis Spyrou
2022-06-30Wrong arguments for running activation function in CpuGemmDirectConv2dMichalis Spyrou
2022-06-29Add LUT-based leaky relu for QASYMM8 on CPUViet-Hoa Do
2022-06-24Improve LUT Neon Hard-SwishPablo Marquez Tello
2022-06-23Select neon LUT Hard-Swish kernel on all devicesPablo Marquez Tello
2022-06-16Fix SVE2 implementation of quantized SoftMax 1DViet-Hoa Do
2022-06-09Fix crash in CpuActivationKernelPablo Marquez Tello
2022-06-07Compute Hard-Swish with a Lookup table for qasymm8.Pablo Marquez Tello
2022-05-24[arm_gemm] Import fixed-format kernels from gemm_linux.Francesco.Petrogalli@arm.com
2022-05-06Use svcreate instead of list initializations.Michalis Spyrou
2022-05-04Add a missing validation check to CPU Pool3dAdnan AlSinan
2022-04-25Update Neon™ pooling kernelramelg01
2022-04-25Add LU_BOUNDED_RELU support for QSYMM16Pablo Marquez Tello