/* * SPDX-FileCopyrightText: Copyright 2021-2022, 2024 Arm Limited and/or its affiliates * SPDX-License-Identifier: Apache-2.0 * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #ifndef USECASE_COMMON_UTILS_HPP #define USECASE_COMMON_UTILS_HPP #include "hal.h" #include "Model.hpp" #include "AppContext.hpp" #include "Profiler.hpp" #include "UseCaseHandler.hpp" /* Handlers for different user options. */ #include "Classifier.hpp" /* Classifier. */ #include "InputFiles.hpp" void DisplayCommonMenu(); /** * @brief Presents inference results using the data presentation * object. * @param[in] results Vector of classification results to be displayed. * @return true if successful, false otherwise. **/ bool PresentInferenceResult(const std::vector& results); /** * @brief Helper function to increment current input feature vector index. * @param[in,out] ctx Pointer to the application context object. * @param[in] useCase Use case name **/ void IncrementAppCtxIfmIdx(arm::app::ApplicationContext& ctx, const std::string& useCase); /** * @brief Helper function to set the input feature map index. * @param[in,out] ctx Pointer to the application context object. * @param[in] idx Value to be set. * @param[in] ctxIfmName Input Feature Map name * @return true if index is set, false otherwise. **/ bool SetAppCtxIfmIdx(arm::app::ApplicationContext& ctx, uint32_t idx, const std::string& ctxIfmName); namespace common { enum OPCODES { MENU_OPT_RUN_INF_NEXT = 1, /* Run on next vector. */ MENU_OPT_RUN_INF_CHOSEN, /* Run on a user provided vector index. */ MENU_OPT_RUN_INF_ALL, /* Run inference on all. */ MENU_OPT_SHOW_MODEL_INFO, /* Show model info. */ MENU_OPT_LIST_IFM, /* List the current IFM. */ #if VSI_ENABLED MENU_OPT_RUN_INF_VSI /* Run on input from VSI. */ #endif }; } namespace arm { namespace app { /** * @brief Run inference using given model * object. If profiling is enabled, it will log the * statistics too. * @param[in] model Reference to the initialised model. * @param[in] profiler Reference to the initialised profiler. * @return true if inference succeeds, false otherwise. **/ bool RunInference(arm::app::Model& model, Profiler& profiler); /** * @brief Read input and return as an integer. * @return Integer value corresponding to the user input. **/ int ReadUserInputAsInt(); #if VERIFY_TEST_OUTPUT /** * @brief Helper function to dump a tensor to stdout * @param[in] tensor tensor to be dumped * @param[in] lineBreakForNumElements number of elements * after which line break will be added. **/ void DumpTensor(const TfLiteTensor* tensor, size_t lineBreakForNumElements = 16); void DumpTensorData(const uint8_t* tensorData, size_t size, size_t lineBreakForNumElements = 16); #endif /* VERIFY_TEST_OUTPUT */ /** * @brief List the files baked in the application. * @param[in] ctx Reference to the application context. * @return true or false based on event being handled. **/ bool ListFilesHandler(ApplicationContext& ctx); } /* namespace app */ } /* namespace arm */ #endif /* USECASE_COMMON_UTILS_HPP */