aboutsummaryrefslogtreecommitdiff
path: root/tests/MobileNetSsdDatabase.hpp
AgeCommit message (Collapse)Author
2020-09-11IVGCVSW-5299 Remove some boost::numeric_cast from armnn/testsMatthew Sloyan
* Replaced with armnn/utility/NumericCast.hpp * Removed combinations without float implementation in NumericCast.hpp Signed-off-by: Matthew Sloyan <matthew.sloyan@arm.com> Change-Id: Ia4ec605f063cdb0071fff302ef48c610f9f9505e
2019-12-05Replace boost logging with simple loggerDerek Lamberti
!referencetests:214319 * Reduces arm nn binary size ~15% * Also fixed test logging black hole issues Change-Id: Iba27db304d9a8088fa46aeb0b52225d93bb56bc8 Signed-off-by: Derek Lamberti <derek.lamberti@arm.com>
2019-10-25IVGCVSW-4018 Move QuantizeHelper.hpp to armnnUtilsAron Virginas-Tar
* Moved QuntizeHelper.hpp to armnnUtils * Reordered parameters for QuantizedVector and added default values for qScale and qOffset to make life easier when using the function for non-quantized types such as Float16 Signed-off-by: Aron Virginas-Tar <Aron.Virginas-Tar@arm.com> Change-Id: I28c263dfa425f1316feccb4116839a84f5d568e5
2019-06-20IVGCVSW-3193 Update bounding box value of MobileNet SSD after refactorNarumol Prangnawarat
Convolution2d and DepthwiseConvolution2d to use Encoders and Decoders Signed-off-by: Narumol Prangnawarat <narumol.prangnawarat@arm.com> Change-Id: Ic87b8ce6f68343983f3c825902425516cbb23b06
2019-02-26IVGCVSW-2560 Verify Inference test for TensorFlow Lite MobileNet SSDNarumol Prangnawarat
* Assign output shape of MobileNet SSD to ArmNN network * Add m_OverridenOutputShapes to TfLiteParser to set shape in GetNetworkOutputBindingInfo * Use input quantization instead of output quantization params * Correct data and datatype in Inference test Change-Id: I01ac2e07ed08e8928ba0df33a4847399e1dd8394 Signed-off-by: Narumol Prangnawarat <narumol.prangnawarat@arm.com> Signed-off-by: Aron Virginas-Tar <Aron.Virginas-Tar@arm.com>
2019-02-20IVGCVSW-2436 Modify MobileNet SSD inference testNarumol Prangnawarat
* change MobileNet SSD input to uint8 * get quantization scale and offset from the model * change data layout to NHWC as TensorFlow lite layout * update expected output as result from TfLite with quantized data Change-Id: I07104d56286893935779169356234de53f1c9492 Signed-off-by: Narumol Prangnawarat <narumol.prangnawarat@arm.com>
2019-01-30IVGCVSW-2437 Inference test for TensorFlow Lite MobileNet SSDAron Virginas-Tar
Change-Id: If7ee1efa3ee79d9eca41c5a6219b3fc42e740efe