From c577f2c6a3b4ddb6ba87a882723c53a248afbeba Mon Sep 17 00:00:00 2001 From: telsoa01 Date: Fri, 31 Aug 2018 09:22:23 +0100 Subject: Release 18.08 --- src/armnn/backends/test/ActivationTestImpl.hpp | 27 +++++++++++++------------- 1 file changed, 14 insertions(+), 13 deletions(-) (limited to 'src/armnn/backends/test/ActivationTestImpl.hpp') diff --git a/src/armnn/backends/test/ActivationTestImpl.hpp b/src/armnn/backends/test/ActivationTestImpl.hpp index 255a00ef0b..e699b2289b 100644 --- a/src/armnn/backends/test/ActivationTestImpl.hpp +++ b/src/armnn/backends/test/ActivationTestImpl.hpp @@ -53,7 +53,7 @@ LayerTestResult BoundedReLuTestCommon(armnn::IWorkloadFactory& workloadFac std::unique_ptr inputHandle = workloadFactory.CreateTensorHandle(inputTensorInfo); std::unique_ptr outputHandle = workloadFactory.CreateTensorHandle(outputTensorInfo); - // Setup bounded ReLu + // Setup bounded ReLu. armnn::ActivationQueueDescriptor descriptor; armnn::WorkloadInfo workloadInfo; AddInputToWorkload(descriptor, workloadInfo, inputTensorInfo, inputHandle.get()); @@ -94,7 +94,7 @@ LayerTestResult BoundedReLuUpperAndLowerBoundTest(armnn::IWorkloadFact 0.999f, 1.2f, 0.89f, 6.1f, }; - // Calculated manually + // Calculated manually. std::vector output = std::vector{ -1.0f, 0.1f, 0.5f, 1.0f, 0.786f, 0.9875f, -1.0f, 0.384f, @@ -122,7 +122,7 @@ LayerTestResult BoundedReLuUpperBoundOnlyTest(armnn::IWorkloadFactory& 0.999f, 1.2f, 0.89f, 6.1f, }; - // Calculated manually + // Calculated manually. std::vector output = std::vector{ 0.0f, 0.1f, 0.5f, 6.0f, 0.786f, 5.9875f, 0.0f, 0.384f, @@ -147,7 +147,7 @@ LayerTestResult BoundedReLuUint8UpperBoundOnlyTest(armnn::IWorkloadF 251, 8, 92 }; - // Calculated manually + // Calculated manually. std::vector output = std::vector{ 0, 122, 0, 255, 0, 58 @@ -176,7 +176,7 @@ LayerTestResult BoundedReLuUint8UpperAndLowerBoundTest(armnn::IWorkl 251, 8, 92 }; - // Calculated manually + // Calculated manually. std::vector output = std::vector{ 51, 192, 32, 192, 32, 92 @@ -186,7 +186,7 @@ LayerTestResult BoundedReLuUint8UpperAndLowerBoundTest(armnn::IWorkl float inputScale = 0.0125f; return BoundedReLuTestCommon(workloadFactory, 1.0f, -1.0f, - inputScale, inputOffset, inputScale, inputOffset, // input/output scale & offset same + inputScale, inputOffset, inputScale, inputOffset, // Input/output scale & offset same. input, output, inputWidth, inputHeight, inputChannels, inputBatchSize); } @@ -229,13 +229,14 @@ boost::multi_array BoundedReLuRandomInputTest(armnn::IWorkloadFactory& boost::multi_array output(GetTensorShapeAsArray<4>(outputTensorInfo)); - // min/max random values passed to MakeRandomTensor are purposely outside of the ReLu range [lowerBound, upperBound] + // Min/max random values passed to MakeRandomTensor are purposely outside of the ReLu + // range [lowerBound, upperBound]. auto input = MakeRandomTensor(inputTensorInfo, 4605828, lowerBound - 5.0f, upperBound * 2.0f); std::unique_ptr inputHandle = workloadFactory.CreateTensorHandle(inputTensorInfo); std::unique_ptr outputHandle = workloadFactory.CreateTensorHandle(outputTensorInfo); - // Setup bounded ReLu + // Set up bounded ReLu. armnn::ActivationQueueDescriptor descriptor; armnn::WorkloadInfo workloadInfo; AddInputToWorkload(descriptor, workloadInfo, inputTensorInfo, inputHandle.get()); @@ -308,7 +309,7 @@ LayerTestResult ConstantLinearActivationTestCommon(armnn::IWorkloadFactory& std::unique_ptr inputHandle = workloadFactory.CreateTensorHandle(inputTensorInfo); std::unique_ptr outputHandle = workloadFactory.CreateTensorHandle(outputTensorInfo); - // Do linear activation that should leave tensor unchanged + // Do linear activation that should leave the tensor unchanged. armnn::ActivationQueueDescriptor data; armnn::WorkloadInfo info; AddInputToWorkload(data, info, inputTensorInfo, inputHandle.get()); @@ -329,7 +330,7 @@ LayerTestResult ConstantLinearActivationTestCommon(armnn::IWorkloadFactory& CopyDataFromITensorHandle(&ret.output[0][0][0][0], outputHandle.get()); - // Ensure output equals input + // Ensure output equals input. ret.outputExpected = input; return ret; @@ -386,7 +387,7 @@ LayerTestResult SimpleActivationTest(armnn::IWorkloadFactory& workloadFact std::unique_ptr inputHandle = workloadFactory.CreateTensorHandle(inputTensorInfo); std::unique_ptr outputHandle = workloadFactory.CreateTensorHandle(outputTensorInfo); - // Setup bounded ReLu + // Setup bounded ReLu. armnn::ActivationQueueDescriptor descriptor; armnn::WorkloadInfo workloadInfo; AddInputToWorkload(descriptor, workloadInfo, inputTensorInfo, inputHandle.get()); @@ -407,7 +408,7 @@ LayerTestResult SimpleActivationTest(armnn::IWorkloadFactory& workloadFact CopyDataFromITensorHandle(&result.output[0][0][0][0], outputHandle.get()); - // Calculated manually + // Calculated manually. result.outputExpected = MakeTensor(outputTensorInfo, QuantizedVector(qScale, qOffset, outputExpectedData)); return result; @@ -423,7 +424,7 @@ LayerTestResult SimpleSigmoidTestCommon(armnn::IWorkloadFactory& workloadF 1.0f, 2.0f, 3.0f, 4.0f }; - // Calculate output values for input + // Calculate output values for input. auto f = [](float value) { return 1.0f / (1.0f + std::exp(-value)); -- cgit v1.2.1