diff options
Diffstat (limited to 'delegate/test')
-rw-r--r-- | delegate/test/DelegateOptionsTest.cpp | 2 | ||||
-rw-r--r-- | delegate/test/DelegateOptionsTestHelper.hpp | 18 | ||||
-rw-r--r-- | delegate/test/ElementwiseUnaryTest.cpp | 20 |
3 files changed, 30 insertions, 10 deletions
diff --git a/delegate/test/DelegateOptionsTest.cpp b/delegate/test/DelegateOptionsTest.cpp index fd1ef88645..349e5d0692 100644 --- a/delegate/test/DelegateOptionsTest.cpp +++ b/delegate/test/DelegateOptionsTest.cpp @@ -177,7 +177,7 @@ TEST_CASE ("ArmnnDelegateStringParsingOptionEnableTfLiteRuntimeFallback") std::vector<armnn::BackendId> backends = { armnn::Compute::CpuRef }; std::vector<int32_t> tensorShape { 1, 2, 2, 1 }; std::vector<float> inputData = { 0.1f, -2.1f, 3.0f, -4.6f }; - std::vector<float> expectedResult = { 1.0f, -2.0f, 3.0f, -4.0f }; + std::vector<float> expectedResult = { 0.995004177f, -0.504846036f, -0.989992499f, -0.112152621f }; // Create options_keys and options_values char array size_t num_options = keys.size(); diff --git a/delegate/test/DelegateOptionsTestHelper.hpp b/delegate/test/DelegateOptionsTestHelper.hpp index b6974c9fb6..148a6d2c22 100644 --- a/delegate/test/DelegateOptionsTestHelper.hpp +++ b/delegate/test/DelegateOptionsTestHelper.hpp @@ -152,10 +152,10 @@ std::vector<char> CreateAddDivTfLiteModel(tflite::TensorType tensorType, flatBufferBuilder.GetBufferPointer() + flatBufferBuilder.GetSize()); } -std::vector<char> CreateCeilTfLiteModel(tflite::TensorType tensorType, - const std::vector <int32_t>& tensorShape, - float quantScale = 1.0f, - int quantOffset = 0) +std::vector<char> CreateCosTfLiteModel(tflite::TensorType tensorType, + const std::vector <int32_t>& tensorShape, + float quantScale = 1.0f, + int quantOffset = 0) { using namespace tflite; flatbuffers::FlatBufferBuilder flatBufferBuilder; @@ -199,7 +199,7 @@ std::vector<char> CreateCeilTfLiteModel(tflite::TensorType tensorType, flatbuffers::Offset<flatbuffers::String> modelDescription = flatBufferBuilder.CreateString("ArmnnDelegate: CEIL Operator Model"); flatbuffers::Offset<OperatorCode> operatorCode = - CreateOperatorCode(flatBufferBuilder, tflite::BuiltinOperator_CEIL); + CreateOperatorCode(flatBufferBuilder, tflite::BuiltinOperator_COS); const std::vector<int32_t> subgraphInputs({0}); const std::vector<int32_t> subgraphOutputs({1}); @@ -277,10 +277,10 @@ void DelegateOptionNoFallbackTest(tflite::TensorType tensorType, int quantOffset = 0) { using namespace delegateTestInterpreter; - std::vector<char> modelBuffer = CreateCeilTfLiteModel(tensorType, - tensorShape, - quantScale, - quantOffset); + std::vector<char> modelBuffer = CreateCosTfLiteModel(tensorType, + tensorShape, + quantScale, + quantOffset); // Setup interpreter with just TFLite Runtime. auto tfLiteInterpreter = DelegateTestInterpreter(modelBuffer); diff --git a/delegate/test/ElementwiseUnaryTest.cpp b/delegate/test/ElementwiseUnaryTest.cpp index 6331436308..b4bdd295e1 100644 --- a/delegate/test/ElementwiseUnaryTest.cpp +++ b/delegate/test/ElementwiseUnaryTest.cpp @@ -295,6 +295,26 @@ TEST_CASE ("Abs_Float32_CpuRef_Test") ElementwiseUnaryFP32Test(tflite::BuiltinOperator_ABS, backends, inputValues, expectedOutputValues); } +TEST_CASE ("Ceil_Float32_CpuRef_Test") +{ + // Create the ArmNN Delegate + std::vector<armnn::BackendId> backends = { armnn::Compute::CpuRef }; + // Set input data + std::vector<float> inputValues + { + 0.0f, 1.1f, -16.1f, + 0.5f, -0.5f, -1.3f + }; + // Set output data + std::vector<float> expectedOutputValues + { + 0.0f, 2.0f, -16.0f, + 1.0f, 0.0f, -1.0f + }; + + ElementwiseUnaryFP32Test(tflite::BuiltinOperator_CEIL, backends, inputValues, expectedOutputValues); +} + TEST_CASE ("Exp_Float32_CpuRef_Test") { // Create the ArmNN Delegate |