summaryrefslogtreecommitdiff
path: root/source/use_case/img_class
diff options
context:
space:
mode:
authoralexander <alexander.efremov@arm.com>2021-04-29 20:36:09 +0100
committerAlexander Efremov <alexander.efremov@arm.com>2021-05-04 19:57:44 +0000
commitc350cdced0a8a2ca17376f58813e6d48d796ac7c (patch)
treef732cde664837a7cb9429b17e1366bb31a635b15 /source/use_case/img_class
parent6448932cc1c612d78e62c778ebb228b3cbe96a58 (diff)
downloadml-embedded-evaluation-kit-c350cdced0a8a2ca17376f58813e6d48d796ac7c.tar.gz
MLECO-1868: Code static analyzer warnings fixes
Signed-off-by: alexander <alexander.efremov@arm.com> Change-Id: Ie423e9cad3fabec6ab077ded7236813fe4933dea
Diffstat (limited to 'source/use_case/img_class')
-rw-r--r--source/use_case/img_class/include/MobileNetModel.hpp4
-rw-r--r--source/use_case/img_class/src/UseCaseHandler.cc28
2 files changed, 16 insertions, 16 deletions
diff --git a/source/use_case/img_class/include/MobileNetModel.hpp b/source/use_case/img_class/include/MobileNetModel.hpp
index f0521ce..2540564 100644
--- a/source/use_case/img_class/include/MobileNetModel.hpp
+++ b/source/use_case/img_class/include/MobileNetModel.hpp
@@ -43,10 +43,10 @@ namespace app {
private:
/* Maximum number of individual operations that can be enlisted. */
- static constexpr int _ms_maxOpCnt = 7;
+ static constexpr int ms_maxOpCnt = 7;
/* A mutable op resolver instance. */
- tflite::MicroMutableOpResolver<_ms_maxOpCnt> _m_opResolver;
+ tflite::MicroMutableOpResolver<ms_maxOpCnt> _m_opResolver;
};
} /* namespace app */
diff --git a/source/use_case/img_class/src/UseCaseHandler.cc b/source/use_case/img_class/src/UseCaseHandler.cc
index f7e83f5..ffeb860 100644
--- a/source/use_case/img_class/src/UseCaseHandler.cc
+++ b/source/use_case/img_class/src/UseCaseHandler.cc
@@ -35,13 +35,13 @@ namespace app {
* @param[out] inputTensor Pointer to the input tensor to be populated.
* @return true if tensor is loaded, false otherwise.
**/
- static bool _LoadImageIntoTensor(uint32_t imIdx, TfLiteTensor* inputTensor);
+ static bool LoadImageIntoTensor(uint32_t imIdx, TfLiteTensor* inputTensor);
/**
* @brief Helper function to increment current image index.
* @param[in,out] ctx Pointer to the application context object.
**/
- static void _IncrementAppCtxImageIdx(ApplicationContext& ctx);
+ static void IncrementAppCtxImageIdx(ApplicationContext& ctx);
/**
* @brief Helper function to set the image index.
@@ -49,7 +49,7 @@ namespace app {
* @param[in] idx Value to be set.
* @return true if index is set, false otherwise.
**/
- static bool _SetAppCtxImageIdx(ApplicationContext& ctx, uint32_t idx);
+ static bool SetAppCtxImageIdx(ApplicationContext& ctx, uint32_t idx);
/**
* @brief Presents inference results using the data presentation
@@ -60,8 +60,8 @@ namespace app {
* otherwise, this can be passed in as 0.
* @return true if successful, false otherwise.
**/
- static bool _PresentInferenceResult(hal_platform& platform,
- const std::vector<ClassificationResult>& results);
+ static bool PresentInferenceResult(hal_platform& platform,
+ const std::vector<ClassificationResult>& results);
/**
* @brief Helper function to convert a UINT8 image to INT8 format.
@@ -89,7 +89,7 @@ namespace app {
/* If the request has a valid size, set the image index. */
if (imgIndex < NUMBER_OF_FILES) {
- if (!_SetAppCtxImageIdx(ctx, imgIndex)) {
+ if (!SetAppCtxImageIdx(ctx, imgIndex)) {
return false;
}
}
@@ -124,7 +124,7 @@ namespace app {
std::string str_inf{"Running inference... "};
/* Copy over the data. */
- _LoadImageIntoTensor(ctx.Get<uint32_t>("imgIndex"), inputTensor);
+ LoadImageIntoTensor(ctx.Get<uint32_t>("imgIndex"), inputTensor);
/* Display this image on the LCD. */
platform.data_psn->present_data_image(
@@ -164,20 +164,20 @@ namespace app {
arm::app::DumpTensor(outputTensor);
#endif /* VERIFY_TEST_OUTPUT */
- if (!_PresentInferenceResult(platform, results)) {
+ if (!PresentInferenceResult(platform, results)) {
return false;
}
profiler.PrintProfilingResult();
- _IncrementAppCtxImageIdx(ctx);
+ IncrementAppCtxImageIdx(ctx);
} while (runAll && ctx.Get<uint32_t>("imgIndex") != curImIdx);
return true;
}
- static bool _LoadImageIntoTensor(const uint32_t imIdx, TfLiteTensor* inputTensor)
+ static bool LoadImageIntoTensor(uint32_t imIdx, TfLiteTensor* inputTensor)
{
const size_t copySz = inputTensor->bytes < IMAGE_DATA_SIZE ?
inputTensor->bytes : IMAGE_DATA_SIZE;
@@ -193,7 +193,7 @@ namespace app {
return true;
}
- static void _IncrementAppCtxImageIdx(ApplicationContext& ctx)
+ static void IncrementAppCtxImageIdx(ApplicationContext& ctx)
{
auto curImIdx = ctx.Get<uint32_t>("imgIndex");
@@ -205,7 +205,7 @@ namespace app {
ctx.Set<uint32_t>("imgIndex", curImIdx);
}
- static bool _SetAppCtxImageIdx(ApplicationContext& ctx, const uint32_t idx)
+ static bool SetAppCtxImageIdx(ApplicationContext& ctx, uint32_t idx)
{
if (idx >= NUMBER_OF_FILES) {
printf_err("Invalid idx %u (expected less than %u)\n",
@@ -216,8 +216,8 @@ namespace app {
return true;
}
- static bool _PresentInferenceResult(hal_platform& platform,
- const std::vector<ClassificationResult>& results)
+ static bool PresentInferenceResult(hal_platform& platform,
+ const std::vector<ClassificationResult>& results)
{
constexpr uint32_t dataPsnTxtStartX1 = 150;
constexpr uint32_t dataPsnTxtStartY1 = 30;