// // Copyright © 2017 Arm Ltd. All rights reserved. // SPDX-License-Identifier: MIT // #include "ConstTensorLayerVisitor.hpp" #include "Network.hpp" #include namespace armnn { void TestConvolution2dLayerVisitor::CheckDescriptor(const Convolution2dDescriptor &convolution2dDescriptor) { BOOST_CHECK(m_Descriptor.m_PadLeft == convolution2dDescriptor.m_PadLeft); BOOST_CHECK(m_Descriptor.m_PadRight == convolution2dDescriptor.m_PadRight); BOOST_CHECK(m_Descriptor.m_PadTop == convolution2dDescriptor.m_PadTop); BOOST_CHECK(m_Descriptor.m_PadBottom == convolution2dDescriptor.m_PadBottom); BOOST_CHECK(m_Descriptor.m_StrideX == convolution2dDescriptor.m_StrideX); BOOST_CHECK(m_Descriptor.m_StrideY == convolution2dDescriptor.m_StrideY); BOOST_CHECK(m_Descriptor.m_BiasEnabled == convolution2dDescriptor.m_BiasEnabled); BOOST_CHECK(m_Descriptor.m_DataLayout == convolution2dDescriptor.m_DataLayout); } void TestDepthwiseConvolution2dLayerVisitor::CheckDescriptor( const DepthwiseConvolution2dDescriptor& convolution2dDescriptor) { BOOST_CHECK(m_Descriptor.m_PadLeft == convolution2dDescriptor.m_PadLeft); BOOST_CHECK(m_Descriptor.m_PadRight == convolution2dDescriptor.m_PadRight); BOOST_CHECK(m_Descriptor.m_PadTop == convolution2dDescriptor.m_PadTop); BOOST_CHECK(m_Descriptor.m_PadBottom == convolution2dDescriptor.m_PadBottom); BOOST_CHECK(m_Descriptor.m_StrideX == convolution2dDescriptor.m_StrideX); BOOST_CHECK(m_Descriptor.m_StrideY == convolution2dDescriptor.m_StrideY); BOOST_CHECK(m_Descriptor.m_BiasEnabled == convolution2dDescriptor.m_BiasEnabled); BOOST_CHECK(m_Descriptor.m_DataLayout == convolution2dDescriptor.m_DataLayout); } void TestFullyConnectedLayerVistor::CheckDescriptor(const FullyConnectedDescriptor& descriptor) { BOOST_CHECK(m_Descriptor.m_BiasEnabled == descriptor.m_BiasEnabled); BOOST_CHECK(m_Descriptor.m_TransposeWeightMatrix == descriptor.m_TransposeWeightMatrix); } void TestBatchNormalizationLayerVisitor::CheckDescriptor(const BatchNormalizationDescriptor& descriptor) { BOOST_CHECK(m_Descriptor.m_Eps == descriptor.m_Eps); BOOST_CHECK(m_Descriptor.m_DataLayout == descriptor.m_DataLayout); } void TestLstmLayerVisitor::CheckDescriptor(const LstmDescriptor& descriptor) { BOOST_CHECK(m_Descriptor.m_ActivationFunc == descriptor.m_ActivationFunc); BOOST_CHECK(m_Descriptor.m_ClippingThresCell == descriptor.m_ClippingThresCell); BOOST_CHECK(m_Descriptor.m_ClippingThresProj == descriptor.m_ClippingThresProj); BOOST_CHECK(m_Descriptor.m_CifgEnabled == descriptor.m_CifgEnabled); BOOST_CHECK(m_Descriptor.m_PeepholeEnabled == descriptor.m_PeepholeEnabled); BOOST_CHECK(m_Descriptor.m_ProjectionEnabled == descriptor.m_ProjectionEnabled); } void TestLstmLayerVisitor::CheckConstTensorPtrs(const std::string& name, const ConstTensor* expected, const ConstTensor* actual) { if (expected == nullptr) { BOOST_CHECK_MESSAGE(actual == nullptr, name + " actual should have been a nullptr"); } else { BOOST_CHECK_MESSAGE(actual != nullptr, name + " actual should have been set"); if (actual != nullptr) { CheckConstTensors(*expected, *actual); } } } void TestLstmLayerVisitor::CheckInputParameters(const LstmInputParams& inputParams) { CheckConstTensorPtrs("ProjectionBias", m_InputParams.m_ProjectionBias, inputParams.m_ProjectionBias); CheckConstTensorPtrs("ProjectionWeights", m_InputParams.m_ProjectionWeights, inputParams.m_ProjectionWeights); CheckConstTensorPtrs("OutputGateBias", m_InputParams.m_OutputGateBias, inputParams.m_OutputGateBias); CheckConstTensorPtrs("InputToInputWeights", m_InputParams.m_InputToInputWeights, inputParams.m_InputToInputWeights); CheckConstTensorPtrs("InputToForgetWeights", m_InputParams.m_InputToForgetWeights, inputParams.m_InputToForgetWeights); CheckConstTensorPtrs("InputToCellWeights", m_InputParams.m_InputToCellWeights, inputParams.m_InputToCellWeights); CheckConstTensorPtrs( "InputToOutputWeights", m_InputParams.m_InputToOutputWeights, inputParams.m_InputToOutputWeights); CheckConstTensorPtrs( "RecurrentToInputWeights", m_InputParams.m_RecurrentToInputWeights, inputParams.m_RecurrentToInputWeights); CheckConstTensorPtrs( "RecurrentToForgetWeights", m_InputParams.m_RecurrentToForgetWeights, inputParams.m_RecurrentToForgetWeights); CheckConstTensorPtrs( "RecurrentToCellWeights", m_InputParams.m_RecurrentToCellWeights, inputParams.m_RecurrentToCellWeights); CheckConstTensorPtrs( "RecurrentToOutputWeights", m_InputParams.m_RecurrentToOutputWeights, inputParams.m_RecurrentToOutputWeights); CheckConstTensorPtrs( "CellToInputWeights", m_InputParams.m_CellToInputWeights, inputParams.m_CellToInputWeights); CheckConstTensorPtrs( "CellToForgetWeights", m_InputParams.m_CellToForgetWeights, inputParams.m_CellToForgetWeights); CheckConstTensorPtrs( "CellToOutputWeights", m_InputParams.m_CellToOutputWeights, inputParams.m_CellToOutputWeights); CheckConstTensorPtrs("InputGateBias", m_InputParams.m_InputGateBias, inputParams.m_InputGateBias); CheckConstTensorPtrs("ForgetGateBias", m_InputParams.m_ForgetGateBias, inputParams.m_ForgetGateBias); CheckConstTensorPtrs("CellBias", m_InputParams.m_CellBias, inputParams.m_CellBias); } void TestQuantizedLstmLayerVisitor::CheckConstTensorPtrs(const std::string& name, const ConstTensor* expected, const ConstTensor* actual) { if (expected == nullptr) { BOOST_CHECK_MESSAGE(actual == nullptr, name + " actual should have been a nullptr"); } else { BOOST_CHECK_MESSAGE(actual != nullptr, name + " actual should have been set"); if (actual != nullptr) { CheckConstTensors(*expected, *actual); } } } void TestQuantizedLstmLayerVisitor::CheckInputParameters(const QuantizedLstmInputParams& inputParams) { CheckConstTensorPtrs("InputToInputWeights", m_InputParams.m_InputToInputWeights, inputParams.m_InputToInputWeights); CheckConstTensorPtrs("InputToForgetWeights", m_InputParams.m_InputToForgetWeights, inputParams.m_InputToForgetWeights); CheckConstTensorPtrs("InputToCellWeights", m_InputParams.m_InputToCellWeights, inputParams.m_InputToCellWeights); CheckConstTensorPtrs("InputToOutputWeights", m_InputParams.m_InputToOutputWeights, inputParams.m_InputToOutputWeights); CheckConstTensorPtrs("RecurrentToInputWeights", m_InputParams.m_RecurrentToInputWeights, inputParams.m_RecurrentToInputWeights); CheckConstTensorPtrs("RecurrentToForgetWeights", m_InputParams.m_RecurrentToForgetWeights, inputParams.m_RecurrentToForgetWeights); CheckConstTensorPtrs("RecurrentToCellWeights", m_InputParams.m_RecurrentToCellWeights, inputParams.m_RecurrentToCellWeights); CheckConstTensorPtrs("RecurrentToOutputWeights", m_InputParams.m_RecurrentToOutputWeights, inputParams.m_RecurrentToOutputWeights); CheckConstTensorPtrs("InputGateBias", m_InputParams.m_InputGateBias, inputParams.m_InputGateBias); CheckConstTensorPtrs("ForgetGateBias", m_InputParams.m_ForgetGateBias, inputParams.m_ForgetGateBias); CheckConstTensorPtrs("CellBias", m_InputParams.m_CellBias, inputParams.m_CellBias); CheckConstTensorPtrs("OutputGateBias", m_InputParams.m_OutputGateBias, inputParams.m_OutputGateBias); } BOOST_AUTO_TEST_SUITE(TestConstTensorLayerVisitor) BOOST_AUTO_TEST_CASE(CheckConvolution2dLayer) { Convolution2dDescriptor descriptor; descriptor.m_PadLeft = 2; descriptor.m_PadRight = 3; descriptor.m_PadBottom = 1; descriptor.m_PadTop = 5; descriptor.m_StrideX = 2; descriptor.m_StrideY = 3; descriptor.m_DataLayout = DataLayout::NHWC; std::vector data = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector dimensions = {1, 1, 3, 3}; ConstTensor weights(TensorInfo(4, dimensions.data(), DataType::Float32), data); TestConvolution2dLayerVisitor visitor(descriptor, weights, EmptyOptional()); Network net; IConnectableLayer* const layer = net.AddConvolution2dLayer(descriptor, weights, EmptyOptional()); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckNamedConvolution2dLayer) { const char* layerName = "Convolution2dLayer"; Convolution2dDescriptor descriptor; descriptor.m_PadLeft = 2; descriptor.m_PadRight = 3; descriptor.m_PadBottom = 1; descriptor.m_PadTop = 5; descriptor.m_StrideX = 2; descriptor.m_StrideY = 3; descriptor.m_DataLayout = DataLayout::NHWC; std::vector data = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector dimensions = {1, 1, 3, 3}; ConstTensor weights(TensorInfo(4, dimensions.data(), DataType::Float32), data); TestConvolution2dLayerVisitor visitor(descriptor, weights, EmptyOptional(), layerName); Network net; IConnectableLayer* const layer = net.AddConvolution2dLayer(descriptor, weights, EmptyOptional(), layerName); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckConvolution2dLayerWithBiases) { Convolution2dDescriptor descriptor; descriptor.m_PadLeft = 2; descriptor.m_PadRight = 3; descriptor.m_PadBottom = 1; descriptor.m_PadTop = 5; descriptor.m_StrideX = 2; descriptor.m_StrideY = 3; descriptor.m_DataLayout = DataLayout::NHWC; descriptor.m_BiasEnabled = true; std::vector data = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector dimensions = {1, 1, 3, 3}; ConstTensor weights(TensorInfo(4, dimensions.data(), DataType::Float32), data); std::vector biasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector biasDimensions = {1, 1, 3, 3}; ConstTensor biases(TensorInfo(4, biasDimensions.data(), DataType::Float32), biasData); Optional optionalBiases(biases); TestConvolution2dLayerVisitor visitor(descriptor, weights, optionalBiases); Network net; IConnectableLayer* const layer = net.AddConvolution2dLayer(descriptor, weights, optionalBiases); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckNamedConvolution2dLayerWithBiases) { const char* layerName = "Convolution2dLayer"; Convolution2dDescriptor descriptor; descriptor.m_PadLeft = 2; descriptor.m_PadRight = 3; descriptor.m_PadBottom = 1; descriptor.m_PadTop = 5; descriptor.m_StrideX = 2; descriptor.m_StrideY = 3; descriptor.m_DataLayout = DataLayout::NHWC; descriptor.m_BiasEnabled = true; std::vector data = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector dimensions = {1, 1, 3, 3}; ConstTensor weights(TensorInfo(4, dimensions.data(), DataType::Float32), data); std::vector biasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector biasDimensions = {1, 1, 3, 3}; ConstTensor biases(TensorInfo(4, biasDimensions.data(), DataType::Float32), biasData); Optional optionalBiases(biases); TestConvolution2dLayerVisitor visitor(descriptor, weights, optionalBiases, layerName); Network net; IConnectableLayer* const layer = net.AddConvolution2dLayer(descriptor, weights, optionalBiases, layerName); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckDepthwiseConvolution2dLayer) { DepthwiseConvolution2dDescriptor descriptor; descriptor.m_PadLeft = 2; descriptor.m_PadRight = 3; descriptor.m_PadBottom = 1; descriptor.m_PadTop = 5; descriptor.m_StrideX = 2; descriptor.m_StrideY = 3; descriptor.m_DataLayout = DataLayout::NHWC; std::vector data = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector dimensions = {1, 1, 3, 3}; ConstTensor weights(TensorInfo(4, dimensions.data(), DataType::Float32), data); TestDepthwiseConvolution2dLayerVisitor visitor(descriptor, weights, EmptyOptional()); Network net; IConnectableLayer* const layer = net.AddDepthwiseConvolution2dLayer(descriptor, weights, EmptyOptional()); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckNamedDepthwiseConvolution2dLayer) { const char* layerName = "DepthwiseConvolution2dLayer"; DepthwiseConvolution2dDescriptor descriptor; descriptor.m_PadLeft = 2; descriptor.m_PadRight = 3; descriptor.m_PadBottom = 1; descriptor.m_PadTop = 5; descriptor.m_StrideX = 2; descriptor.m_StrideY = 3; descriptor.m_DataLayout = DataLayout::NHWC; std::vector data = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector dimensions = {1, 1, 3, 3}; ConstTensor weights(TensorInfo(4, dimensions.data(), DataType::Float32), data); TestDepthwiseConvolution2dLayerVisitor visitor(descriptor, weights, EmptyOptional(), layerName); Network net; IConnectableLayer* const layer = net.AddDepthwiseConvolution2dLayer(descriptor, weights, EmptyOptional(), layerName); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckDepthwiseConvolution2dLayerWithBiases) { DepthwiseConvolution2dDescriptor descriptor; descriptor.m_PadLeft = 2; descriptor.m_PadRight = 3; descriptor.m_PadBottom = 1; descriptor.m_PadTop = 5; descriptor.m_StrideX = 2; descriptor.m_StrideY = 3; descriptor.m_DataLayout = DataLayout::NHWC; descriptor.m_BiasEnabled = true; std::vector data = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector dimensions = {1, 1, 3, 3}; ConstTensor weights(TensorInfo(4, dimensions.data(), DataType::Float32), data); std::vector biasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector biasDimensions = {1, 1, 3, 3}; ConstTensor biases(TensorInfo(4, biasDimensions.data(), DataType::Float32), biasData); Optional optionalBiases(biases); TestDepthwiseConvolution2dLayerVisitor visitor(descriptor, weights, optionalBiases); Network net; IConnectableLayer* const layer = net.AddDepthwiseConvolution2dLayer(descriptor, weights, optionalBiases); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckNamedDepthwiseConvolution2dLayerWithBiases) { const char* layerName = "DepthwiseConvolution2dLayer"; DepthwiseConvolution2dDescriptor descriptor; descriptor.m_PadLeft = 2; descriptor.m_PadRight = 3; descriptor.m_PadBottom = 1; descriptor.m_PadTop = 5; descriptor.m_StrideX = 2; descriptor.m_StrideY = 3; descriptor.m_DataLayout = DataLayout::NHWC; descriptor.m_BiasEnabled = true; std::vector data = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector dimensions = {1, 1, 3, 3}; ConstTensor weights(TensorInfo(4, dimensions.data(), DataType::Float32), data); std::vector biasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector biasDimensions = {1, 1, 3, 3}; ConstTensor biases(TensorInfo(4, biasDimensions.data(), DataType::Float32), biasData); Optional optionalBiases(biases); TestDepthwiseConvolution2dLayerVisitor visitor(descriptor, weights, optionalBiases, layerName); Network net; IConnectableLayer* const layer = net.AddDepthwiseConvolution2dLayer(descriptor, weights, optionalBiases, layerName); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckFullyConnectedLayer) { FullyConnectedDescriptor descriptor; descriptor.m_TransposeWeightMatrix = true; std::vector data = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector dimensions = {1, 1, 3, 3}; ConstTensor weights(TensorInfo(4, dimensions.data(), DataType::Float32), data); TestFullyConnectedLayerVistor visitor(descriptor, weights, EmptyOptional()); Network net; IConnectableLayer* const layer = net.AddFullyConnectedLayer(descriptor, weights, EmptyOptional()); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckNamedFullyConnectedLayer) { const char* layerName = "FullyConnectedLayer"; FullyConnectedDescriptor descriptor; descriptor.m_TransposeWeightMatrix = true; std::vector data = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector dimensions = {1, 1, 3, 3}; ConstTensor weights(TensorInfo(4, dimensions.data(), DataType::Float32), data); TestFullyConnectedLayerVistor visitor(descriptor, weights, EmptyOptional(), layerName); Network net; IConnectableLayer* const layer = net.AddFullyConnectedLayer(descriptor, weights, EmptyOptional(), layerName); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckFullyConnectedLayerWithBiases) { FullyConnectedDescriptor descriptor; descriptor.m_TransposeWeightMatrix = true; descriptor.m_BiasEnabled = true; std::vector data = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector dimensions = {1, 1, 3, 3}; ConstTensor weights(TensorInfo(4, dimensions.data(), DataType::Float32), data); std::vector biasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector biasDimensions = {1, 1, 3, 3}; ConstTensor biases(TensorInfo(4, biasDimensions.data(), DataType::Float32), biasData); Optional optionalBiases(biases); TestFullyConnectedLayerVistor visitor(descriptor, weights, optionalBiases); Network net; IConnectableLayer* const layer = net.AddFullyConnectedLayer(descriptor, weights, optionalBiases); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckNamedFullyConnectedLayerWithBiases) { const char* layerName = "FullyConnectedLayer"; FullyConnectedDescriptor descriptor; descriptor.m_TransposeWeightMatrix = true; descriptor.m_BiasEnabled = true; std::vector data = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector dimensions = {1, 1, 3, 3}; ConstTensor weights(TensorInfo(4, dimensions.data(), DataType::Float32), data); std::vector biasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector biasDimensions = {1, 1, 3, 3}; ConstTensor biases(TensorInfo(4, biasDimensions.data(), DataType::Float32), biasData); Optional optionalBiases(biases); TestFullyConnectedLayerVistor visitor(descriptor, weights, optionalBiases, layerName); Network net; IConnectableLayer* const layer = net.AddFullyConnectedLayer(descriptor, weights, optionalBiases, layerName); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckBatchNormalizationLayer) { BatchNormalizationDescriptor descriptor; descriptor.m_Eps = 0.0002f; descriptor.m_DataLayout = DataLayout::NHWC; std::vector data = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector dimensions = {1, 1, 3, 3}; ConstTensor mean(TensorInfo(4, dimensions.data(), DataType::Float32), data); std::vector varianceData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector varianceDimensions = {1, 1, 3, 3}; ConstTensor variance(TensorInfo(4, varianceDimensions.data(), DataType::Float32), varianceData); std::vector betaData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector betaDimensions = {1, 1, 3, 3}; ConstTensor beta(TensorInfo(4, betaDimensions.data(), DataType::Float32), betaData); std::vector gammaData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector gammaDimensions = {1, 1, 3, 3}; ConstTensor gamma(TensorInfo(4, gammaDimensions.data(), DataType::Float32), gammaData); TestBatchNormalizationLayerVisitor visitor(descriptor, mean, variance, beta, gamma); Network net; IConnectableLayer* const layer = net.AddBatchNormalizationLayer(descriptor, mean, variance, beta, gamma); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckNamedBatchNormalizationLayer) { const char* layerName = "BatchNormalizationLayer"; BatchNormalizationDescriptor descriptor; descriptor.m_Eps = 0.0002f; descriptor.m_DataLayout = DataLayout::NHWC; std::vector data = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector dimensions = {1, 1, 3, 3}; ConstTensor mean(TensorInfo(4, dimensions.data(), DataType::Float32), data); std::vector varianceData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector varianceDimensions = {1, 1, 3, 3}; ConstTensor variance(TensorInfo(4, varianceDimensions.data(), DataType::Float32), varianceData); std::vector betaData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector betaDimensions = {1, 1, 3, 3}; ConstTensor beta(TensorInfo(4, betaDimensions.data(), DataType::Float32), betaData); std::vector gammaData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector gammaDimensions = {1, 1, 3, 3}; ConstTensor gamma(TensorInfo(4, gammaDimensions.data(), DataType::Float32), gammaData); TestBatchNormalizationLayerVisitor visitor(descriptor, mean, variance, beta, gamma, layerName); Network net; IConnectableLayer* const layer = net.AddBatchNormalizationLayer( descriptor, mean, variance, beta, gamma, layerName); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckConstLayer) { std::vector data = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector dimensions = {1, 1, 3, 3}; ConstTensor input(TensorInfo(4, dimensions.data(), DataType::Float32), data); TestConstantLayerVisitor visitor(input); Network net; IConnectableLayer* const layer = net.AddConstantLayer(input); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckNamedConstLayer) { const char* layerName = "ConstantLayer"; std::vector data = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector dimensions = {1, 1, 3, 3}; ConstTensor input(TensorInfo(4, dimensions.data(), DataType::Float32), data); TestConstantLayerVisitor visitor(input, layerName); Network net; IConnectableLayer* const layer = net.AddConstantLayer(input, layerName); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckLstmLayerBasic) { LstmDescriptor descriptor; descriptor.m_ActivationFunc = 3; descriptor.m_ClippingThresProj = 0.5f; descriptor.m_ClippingThresCell = 0.3f; descriptor.m_CifgEnabled = true; // if this is true then we DON'T need to set the OptCifgParams std::vector inputToForgetWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToForgetWeights( TensorInfo(4, inputToForgetWeightsDimensions.data(), DataType::Float32), inputToForgetWeightsData); std::vector inputToCellWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToCellWeights( TensorInfo(4, inputToCellWeightsDimensions.data(), DataType::Float32), inputToCellWeightsData); std::vector inputToOutputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToOutputWeights( TensorInfo(4, inputToOutputWeightsDimensions.data(), DataType::Float32), inputToOutputWeightsData); std::vector recurrentToForgetWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToForgetWeights(TensorInfo( 4, recurrentToForgetWeightsDimensions.data(), DataType::Float32), recurrentToForgetWeightsData); std::vector recurrentToCellWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToCellWeights(TensorInfo( 4, recurrentToCellWeightsDimensions.data(), DataType::Float32), recurrentToCellWeightsData); std::vector recurrentToOutputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToOutputWeights(TensorInfo( 4, recurrentToOutputWeightsDimensions.data(), DataType::Float32), recurrentToOutputWeightsData); std::vector forgetGateBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector forgetGateBiasDimensions = {1, 1, 3, 3}; ConstTensor forgetGateBias(TensorInfo( 4, forgetGateBiasDimensions.data(), DataType::Float32), forgetGateBiasData); std::vector cellBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector cellBiasDimensions = {1, 1, 3, 3}; ConstTensor cellBias(TensorInfo( 4, cellBiasDimensions.data(), DataType::Float32), cellBiasData); std::vector outputGateBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector outputGateBiasDimensions = {1, 1, 3, 3}; ConstTensor outputGateBias(TensorInfo( 4, outputGateBiasDimensions.data(), DataType::Float32), outputGateBiasData); LstmInputParams params; params.m_InputToForgetWeights = &inputToForgetWeights; params.m_InputToCellWeights = &inputToCellWeights; params.m_InputToOutputWeights = &inputToOutputWeights; params.m_RecurrentToForgetWeights = &recurrentToForgetWeights; params.m_RecurrentToCellWeights = &recurrentToCellWeights; params.m_RecurrentToOutputWeights = &recurrentToOutputWeights; params.m_ForgetGateBias = &forgetGateBias; params.m_CellBias = &cellBias; params.m_OutputGateBias = &outputGateBias; TestLstmLayerVisitor visitor(descriptor, params); Network net; IConnectableLayer* const layer = net.AddLstmLayer(descriptor, params); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckNamedLstmLayerBasic) { const char* layerName = "LstmLayer"; LstmDescriptor descriptor; descriptor.m_ActivationFunc = 3; descriptor.m_ClippingThresProj = 0.5f; descriptor.m_ClippingThresCell = 0.3f; descriptor.m_CifgEnabled = true; // if this is true then we DON'T need to set the OptCifgParams std::vector inputToForgetWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToForgetWeights( TensorInfo(4, inputToForgetWeightsDimensions.data(), DataType::Float32), inputToForgetWeightsData); std::vector inputToCellWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToCellWeights( TensorInfo(4, inputToCellWeightsDimensions.data(), DataType::Float32), inputToCellWeightsData); std::vector inputToOutputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToOutputWeights( TensorInfo(4, inputToOutputWeightsDimensions.data(), DataType::Float32), inputToOutputWeightsData); std::vector recurrentToForgetWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToForgetWeights(TensorInfo( 4, recurrentToForgetWeightsDimensions.data(), DataType::Float32), recurrentToForgetWeightsData); std::vector recurrentToCellWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToCellWeights(TensorInfo( 4, recurrentToCellWeightsDimensions.data(), DataType::Float32), recurrentToCellWeightsData); std::vector recurrentToOutputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToOutputWeights(TensorInfo( 4, recurrentToOutputWeightsDimensions.data(), DataType::Float32), recurrentToOutputWeightsData); std::vector forgetGateBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector forgetGateBiasDimensions = {1, 1, 3, 3}; ConstTensor forgetGateBias(TensorInfo( 4, forgetGateBiasDimensions.data(), DataType::Float32), forgetGateBiasData); std::vector cellBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector cellBiasDimensions = {1, 1, 3, 3}; ConstTensor cellBias(TensorInfo( 4, cellBiasDimensions.data(), DataType::Float32), cellBiasData); std::vector outputGateBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector outputGateBiasDimensions = {1, 1, 3, 3}; ConstTensor outputGateBias(TensorInfo( 4, outputGateBiasDimensions.data(), DataType::Float32), outputGateBiasData); LstmInputParams params; params.m_InputToForgetWeights = &inputToForgetWeights; params.m_InputToCellWeights = &inputToCellWeights; params.m_InputToOutputWeights = &inputToOutputWeights; params.m_RecurrentToForgetWeights = &recurrentToForgetWeights; params.m_RecurrentToCellWeights = &recurrentToCellWeights; params.m_RecurrentToOutputWeights = &recurrentToOutputWeights; params.m_ForgetGateBias = &forgetGateBias; params.m_CellBias = &cellBias; params.m_OutputGateBias = &outputGateBias; TestLstmLayerVisitor visitor(descriptor, params, layerName); Network net; IConnectableLayer* const layer = net.AddLstmLayer(descriptor, params, layerName); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckLstmLayerCifgDisabled) { LstmDescriptor descriptor; descriptor.m_ActivationFunc = 3; descriptor.m_ClippingThresProj = 0.5f; descriptor.m_ClippingThresCell = 0.3f; descriptor.m_CifgEnabled = false; // if this is true then we DON'T need to set the OptCifgParams std::vector inputToForgetWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToForgetWeights( TensorInfo(4, inputToForgetWeightsDimensions.data(), DataType::Float32), inputToForgetWeightsData); std::vector inputToCellWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToCellWeights( TensorInfo(4, inputToCellWeightsDimensions.data(), DataType::Float32), inputToCellWeightsData); std::vector inputToOutputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToOutputWeights( TensorInfo(4, inputToOutputWeightsDimensions.data(), DataType::Float32), inputToOutputWeightsData); std::vector recurrentToForgetWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToForgetWeights(TensorInfo( 4, recurrentToForgetWeightsDimensions.data(), DataType::Float32), recurrentToForgetWeightsData); std::vector recurrentToCellWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToCellWeights(TensorInfo( 4, recurrentToCellWeightsDimensions.data(), DataType::Float32), recurrentToCellWeightsData); std::vector recurrentToOutputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToOutputWeights(TensorInfo( 4, recurrentToOutputWeightsDimensions.data(), DataType::Float32), recurrentToOutputWeightsData); std::vector forgetGateBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector forgetGateBiasDimensions = {1, 1, 3, 3}; ConstTensor forgetGateBias(TensorInfo( 4, forgetGateBiasDimensions.data(), DataType::Float32), forgetGateBiasData); std::vector cellBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector cellBiasDimensions = {1, 1, 3, 3}; ConstTensor cellBias(TensorInfo( 4, cellBiasDimensions.data(), DataType::Float32), cellBiasData); std::vector outputGateBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector outputGateBiasDimensions = {1, 1, 3, 3}; ConstTensor outputGateBias(TensorInfo( 4, outputGateBiasDimensions.data(), DataType::Float32), outputGateBiasData); std::vector inputToInputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToInputWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToInputWeights( TensorInfo(4, inputToInputWeightsDimensions.data(), DataType::Float32), inputToInputWeightsData); std::vector recurrentToInputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToInputWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToInputWeights(TensorInfo( 4, recurrentToInputWeightsDimensions.data(), DataType::Float32), recurrentToInputWeightsData); std::vector cellToInputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector cellToInputWeightsDimensions = {1, 1, 3, 3}; ConstTensor cellToInputWeights( TensorInfo(4, cellToInputWeightsDimensions.data(), DataType::Float32), cellToInputWeightsData); std::vector inputGateBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputGateBiasDimensions = {1, 1, 3, 3}; ConstTensor inputGateBias( TensorInfo(4, inputGateBiasDimensions.data(), DataType::Float32), inputGateBiasData); LstmInputParams params; params.m_InputToForgetWeights = &inputToForgetWeights; params.m_InputToCellWeights = &inputToCellWeights; params.m_InputToOutputWeights = &inputToOutputWeights; params.m_RecurrentToForgetWeights = &recurrentToForgetWeights; params.m_RecurrentToCellWeights = &recurrentToCellWeights; params.m_RecurrentToOutputWeights = &recurrentToOutputWeights; params.m_ForgetGateBias = &forgetGateBias; params.m_CellBias = &cellBias; params.m_OutputGateBias = &outputGateBias; params.m_InputToInputWeights = &inputToInputWeights; params.m_RecurrentToInputWeights = &recurrentToInputWeights; params.m_CellToInputWeights = &cellToInputWeights; params.m_InputGateBias = &inputGateBias; TestLstmLayerVisitor visitor(descriptor, params); Network net; IConnectableLayer* const layer = net.AddLstmLayer(descriptor, params); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckNamedLstmLayerCifgDisabled) { const char* layerName = "LstmLayer"; LstmDescriptor descriptor; descriptor.m_ActivationFunc = 3; descriptor.m_ClippingThresProj = 0.5f; descriptor.m_ClippingThresCell = 0.3f; descriptor.m_CifgEnabled = false; // if this is true then we DON'T need to set the OptCifgParams std::vector inputToForgetWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToForgetWeights( TensorInfo(4, inputToForgetWeightsDimensions.data(), DataType::Float32), inputToForgetWeightsData); std::vector inputToCellWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToCellWeights( TensorInfo(4, inputToCellWeightsDimensions.data(), DataType::Float32), inputToCellWeightsData); std::vector inputToOutputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToOutputWeights( TensorInfo(4, inputToOutputWeightsDimensions.data(), DataType::Float32), inputToOutputWeightsData); std::vector recurrentToForgetWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToForgetWeights(TensorInfo( 4, recurrentToForgetWeightsDimensions.data(), DataType::Float32), recurrentToForgetWeightsData); std::vector recurrentToCellWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToCellWeights(TensorInfo( 4, recurrentToCellWeightsDimensions.data(), DataType::Float32), recurrentToCellWeightsData); std::vector recurrentToOutputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToOutputWeights(TensorInfo( 4, recurrentToOutputWeightsDimensions.data(), DataType::Float32), recurrentToOutputWeightsData); std::vector forgetGateBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector forgetGateBiasDimensions = {1, 1, 3, 3}; ConstTensor forgetGateBias(TensorInfo( 4, forgetGateBiasDimensions.data(), DataType::Float32), forgetGateBiasData); std::vector cellBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector cellBiasDimensions = {1, 1, 3, 3}; ConstTensor cellBias(TensorInfo( 4, cellBiasDimensions.data(), DataType::Float32), cellBiasData); std::vector outputGateBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector outputGateBiasDimensions = {1, 1, 3, 3}; ConstTensor outputGateBias(TensorInfo( 4, outputGateBiasDimensions.data(), DataType::Float32), outputGateBiasData); std::vector inputToInputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToInputWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToInputWeights( TensorInfo(4, inputToInputWeightsDimensions.data(), DataType::Float32), inputToInputWeightsData); std::vector recurrentToInputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToInputWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToInputWeights(TensorInfo( 4, recurrentToInputWeightsDimensions.data(), DataType::Float32), recurrentToInputWeightsData); std::vector cellToInputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector cellToInputWeightsDimensions = {1, 1, 3, 3}; ConstTensor cellToInputWeights( TensorInfo(4, cellToInputWeightsDimensions.data(), DataType::Float32), cellToInputWeightsData); std::vector inputGateBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputGateBiasDimensions = {1, 1, 3, 3}; ConstTensor inputGateBias( TensorInfo(4, inputGateBiasDimensions.data(), DataType::Float32), inputGateBiasData); LstmInputParams params; params.m_InputToForgetWeights = &inputToForgetWeights; params.m_InputToCellWeights = &inputToCellWeights; params.m_InputToOutputWeights = &inputToOutputWeights; params.m_RecurrentToForgetWeights = &recurrentToForgetWeights; params.m_RecurrentToCellWeights = &recurrentToCellWeights; params.m_RecurrentToOutputWeights = &recurrentToOutputWeights; params.m_ForgetGateBias = &forgetGateBias; params.m_CellBias = &cellBias; params.m_OutputGateBias = &outputGateBias; params.m_InputToInputWeights = &inputToInputWeights; params.m_RecurrentToInputWeights = &recurrentToInputWeights; params.m_CellToInputWeights = &cellToInputWeights; params.m_InputGateBias = &inputGateBias; TestLstmLayerVisitor visitor(descriptor, params, layerName); Network net; IConnectableLayer *const layer = net.AddLstmLayer(descriptor, params, layerName); layer->Accept(visitor); } // TODO add one with peephole BOOST_AUTO_TEST_CASE(CheckLstmLayerPeephole) { LstmDescriptor descriptor; descriptor.m_ActivationFunc = 3; descriptor.m_ClippingThresProj = 0.5f; descriptor.m_ClippingThresCell = 0.3f; descriptor.m_CifgEnabled = true; // if this is true then we DON'T need to set the OptCifgParams descriptor.m_PeepholeEnabled = true; std::vector inputToForgetWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToForgetWeights( TensorInfo(4, inputToForgetWeightsDimensions.data(), DataType::Float32), inputToForgetWeightsData); std::vector inputToCellWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToCellWeights( TensorInfo(4, inputToCellWeightsDimensions.data(), DataType::Float32), inputToCellWeightsData); std::vector inputToOutputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToOutputWeights( TensorInfo(4, inputToOutputWeightsDimensions.data(), DataType::Float32), inputToOutputWeightsData); std::vector recurrentToForgetWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToForgetWeights(TensorInfo( 4, recurrentToForgetWeightsDimensions.data(), DataType::Float32), recurrentToForgetWeightsData); std::vector recurrentToCellWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToCellWeights(TensorInfo( 4, recurrentToCellWeightsDimensions.data(), DataType::Float32), recurrentToCellWeightsData); std::vector recurrentToOutputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToOutputWeights(TensorInfo( 4, recurrentToOutputWeightsDimensions.data(), DataType::Float32), recurrentToOutputWeightsData); std::vector forgetGateBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector forgetGateBiasDimensions = {1, 1, 3, 3}; ConstTensor forgetGateBias(TensorInfo( 4, forgetGateBiasDimensions.data(), DataType::Float32), forgetGateBiasData); std::vector cellBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector cellBiasDimensions = {1, 1, 3, 3}; ConstTensor cellBias(TensorInfo( 4, cellBiasDimensions.data(), DataType::Float32), cellBiasData); std::vector outputGateBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector outputGateBiasDimensions = {1, 1, 3, 3}; ConstTensor outputGateBias(TensorInfo( 4, outputGateBiasDimensions.data(), DataType::Float32), outputGateBiasData); std::vector cellToForgetWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector cellToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor cellToForgetWeights( TensorInfo(4, cellToForgetWeightsDimensions.data(), DataType::Float32), cellToForgetWeightsData); std::vector cellToOutputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector cellToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor cellToOutputWeights( TensorInfo(4, cellToOutputWeightsDimensions.data(), DataType::Float32), cellToOutputWeightsData); LstmInputParams params; params.m_InputToForgetWeights = &inputToForgetWeights; params.m_InputToCellWeights = &inputToCellWeights; params.m_InputToOutputWeights = &inputToOutputWeights; params.m_RecurrentToForgetWeights = &recurrentToForgetWeights; params.m_RecurrentToCellWeights = &recurrentToCellWeights; params.m_RecurrentToOutputWeights = &recurrentToOutputWeights; params.m_ForgetGateBias = &forgetGateBias; params.m_CellBias = &cellBias; params.m_OutputGateBias = &outputGateBias; params.m_CellToForgetWeights = &cellToForgetWeights; params.m_CellToOutputWeights = &cellToOutputWeights; TestLstmLayerVisitor visitor(descriptor, params); Network net; IConnectableLayer *const layer = net.AddLstmLayer(descriptor, params); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckNamedLstmLayerPeephole) { const char* layerName = "LstmLayer"; LstmDescriptor descriptor; descriptor.m_ActivationFunc = 3; descriptor.m_ClippingThresProj = 0.5f; descriptor.m_ClippingThresCell = 0.3f; descriptor.m_CifgEnabled = true; // if this is true then we DON'T need to set the OptCifgParams descriptor.m_PeepholeEnabled = true; std::vector inputToForgetWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToForgetWeights( TensorInfo(4, inputToForgetWeightsDimensions.data(), DataType::Float32), inputToForgetWeightsData); std::vector inputToCellWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToCellWeights( TensorInfo(4, inputToCellWeightsDimensions.data(), DataType::Float32), inputToCellWeightsData); std::vector inputToOutputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToOutputWeights( TensorInfo(4, inputToOutputWeightsDimensions.data(), DataType::Float32), inputToOutputWeightsData); std::vector recurrentToForgetWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToForgetWeights(TensorInfo( 4, recurrentToForgetWeightsDimensions.data(), DataType::Float32), recurrentToForgetWeightsData); std::vector recurrentToCellWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToCellWeights(TensorInfo( 4, recurrentToCellWeightsDimensions.data(), DataType::Float32), recurrentToCellWeightsData); std::vector recurrentToOutputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToOutputWeights(TensorInfo( 4, recurrentToOutputWeightsDimensions.data(), DataType::Float32), recurrentToOutputWeightsData); std::vector forgetGateBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector forgetGateBiasDimensions = {1, 1, 3, 3}; ConstTensor forgetGateBias(TensorInfo( 4, forgetGateBiasDimensions.data(), DataType::Float32), forgetGateBiasData); std::vector cellBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector cellBiasDimensions = {1, 1, 3, 3}; ConstTensor cellBias(TensorInfo( 4, cellBiasDimensions.data(), DataType::Float32), cellBiasData); std::vector outputGateBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector outputGateBiasDimensions = {1, 1, 3, 3}; ConstTensor outputGateBias(TensorInfo( 4, outputGateBiasDimensions.data(), DataType::Float32), outputGateBiasData); std::vector cellToForgetWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector cellToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor cellToForgetWeights( TensorInfo(4, cellToForgetWeightsDimensions.data(), DataType::Float32), cellToForgetWeightsData); std::vector cellToOutputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector cellToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor cellToOutputWeights( TensorInfo(4, cellToOutputWeightsDimensions.data(), DataType::Float32), cellToOutputWeightsData); LstmInputParams params; params.m_InputToForgetWeights = &inputToForgetWeights; params.m_InputToCellWeights = &inputToCellWeights; params.m_InputToOutputWeights = &inputToOutputWeights; params.m_RecurrentToForgetWeights = &recurrentToForgetWeights; params.m_RecurrentToCellWeights = &recurrentToCellWeights; params.m_RecurrentToOutputWeights = &recurrentToOutputWeights; params.m_ForgetGateBias = &forgetGateBias; params.m_CellBias = &cellBias; params.m_OutputGateBias = &outputGateBias; params.m_CellToForgetWeights = &cellToForgetWeights; params.m_CellToOutputWeights = &cellToOutputWeights; TestLstmLayerVisitor visitor(descriptor, params, layerName); Network net; IConnectableLayer *const layer = net.AddLstmLayer(descriptor, params, layerName); layer->Accept(visitor); } // TODO add one with projection BOOST_AUTO_TEST_CASE(CheckLstmLayerProjection) { LstmDescriptor descriptor; descriptor.m_ActivationFunc = 3; descriptor.m_ClippingThresProj = 0.5f; descriptor.m_ClippingThresCell = 0.3f; descriptor.m_CifgEnabled = true; // if this is true then we DON'T need to set the OptCifgParams descriptor.m_ProjectionEnabled = true; std::vector inputToForgetWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToForgetWeights( TensorInfo(4, inputToForgetWeightsDimensions.data(), DataType::Float32), inputToForgetWeightsData); std::vector inputToCellWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToCellWeights( TensorInfo(4, inputToCellWeightsDimensions.data(), DataType::Float32), inputToCellWeightsData); std::vector inputToOutputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToOutputWeights( TensorInfo(4, inputToOutputWeightsDimensions.data(), DataType::Float32), inputToOutputWeightsData); std::vector recurrentToForgetWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToForgetWeights(TensorInfo( 4, recurrentToForgetWeightsDimensions.data(), DataType::Float32), recurrentToForgetWeightsData); std::vector recurrentToCellWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToCellWeights(TensorInfo( 4, recurrentToCellWeightsDimensions.data(), DataType::Float32), recurrentToCellWeightsData); std::vector recurrentToOutputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToOutputWeights(TensorInfo( 4, recurrentToOutputWeightsDimensions.data(), DataType::Float32), recurrentToOutputWeightsData); std::vector forgetGateBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector forgetGateBiasDimensions = {1, 1, 3, 3}; ConstTensor forgetGateBias(TensorInfo( 4, forgetGateBiasDimensions.data(), DataType::Float32), forgetGateBiasData); std::vector cellBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector cellBiasDimensions = {1, 1, 3, 3}; ConstTensor cellBias(TensorInfo( 4, cellBiasDimensions.data(), DataType::Float32), cellBiasData); std::vector outputGateBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector outputGateBiasDimensions = {1, 1, 3, 3}; ConstTensor outputGateBias(TensorInfo( 4, outputGateBiasDimensions.data(), DataType::Float32), outputGateBiasData); std::vector projectionBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector projectionBiasDimensions = {1, 1, 3, 3}; ConstTensor projectionBias( TensorInfo(4, projectionBiasDimensions.data(), DataType::Float32), projectionBiasData); std::vector projectionWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector projectionWeightsDimensions = {1, 1, 3, 3}; ConstTensor projectionWeights( TensorInfo(4, projectionWeightsDimensions.data(), DataType::Float32), projectionWeightsData); LstmInputParams params; params.m_InputToForgetWeights = &inputToForgetWeights; params.m_InputToCellWeights = &inputToCellWeights; params.m_InputToOutputWeights = &inputToOutputWeights; params.m_RecurrentToForgetWeights = &recurrentToForgetWeights; params.m_RecurrentToCellWeights = &recurrentToCellWeights; params.m_RecurrentToOutputWeights = &recurrentToOutputWeights; params.m_ForgetGateBias = &forgetGateBias; params.m_CellBias = &cellBias; params.m_OutputGateBias = &outputGateBias; params.m_ProjectionWeights = &projectionWeights; params.m_ProjectionBias = &projectionBias; TestLstmLayerVisitor visitor(descriptor, params); Network net; IConnectableLayer *const layer = net.AddLstmLayer(descriptor, params); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckNamedLstmLayerProjection) { const char* layerName = "LstmLayer"; LstmDescriptor descriptor; descriptor.m_ActivationFunc = 3; descriptor.m_ClippingThresProj = 0.5f; descriptor.m_ClippingThresCell = 0.3f; descriptor.m_CifgEnabled = true; // if this is true then we DON'T need to set the OptCifgParams descriptor.m_ProjectionEnabled = true; std::vector inputToForgetWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToForgetWeights( TensorInfo(4, inputToForgetWeightsDimensions.data(), DataType::Float32), inputToForgetWeightsData); std::vector inputToCellWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToCellWeights( TensorInfo(4, inputToCellWeightsDimensions.data(), DataType::Float32), inputToCellWeightsData); std::vector inputToOutputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector inputToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToOutputWeights( TensorInfo(4, inputToOutputWeightsDimensions.data(), DataType::Float32), inputToOutputWeightsData); std::vector recurrentToForgetWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToForgetWeights(TensorInfo( 4, recurrentToForgetWeightsDimensions.data(), DataType::Float32), recurrentToForgetWeightsData); std::vector recurrentToCellWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToCellWeights(TensorInfo( 4, recurrentToCellWeightsDimensions.data(), DataType::Float32), recurrentToCellWeightsData); std::vector recurrentToOutputWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector recurrentToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToOutputWeights(TensorInfo( 4, recurrentToOutputWeightsDimensions.data(), DataType::Float32), recurrentToOutputWeightsData); std::vector forgetGateBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector forgetGateBiasDimensions = {1, 1, 3, 3}; ConstTensor forgetGateBias(TensorInfo( 4, forgetGateBiasDimensions.data(), DataType::Float32), forgetGateBiasData); std::vector cellBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector cellBiasDimensions = {1, 1, 3, 3}; ConstTensor cellBias(TensorInfo( 4, cellBiasDimensions.data(), DataType::Float32), cellBiasData); std::vector outputGateBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector outputGateBiasDimensions = {1, 1, 3, 3}; ConstTensor outputGateBias(TensorInfo( 4, outputGateBiasDimensions.data(), DataType::Float32), outputGateBiasData); std::vector projectionBiasData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector projectionBiasDimensions = {1, 1, 3, 3}; ConstTensor projectionBias( TensorInfo(4, projectionBiasDimensions.data(), DataType::Float32), projectionBiasData); std::vector projectionWeightsData = {1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0}; std::vector projectionWeightsDimensions = {1, 1, 3, 3}; ConstTensor projectionWeights( TensorInfo(4, projectionWeightsDimensions.data(), DataType::Float32), projectionWeightsData); LstmInputParams params; params.m_InputToForgetWeights = &inputToForgetWeights; params.m_InputToCellWeights = &inputToCellWeights; params.m_InputToOutputWeights = &inputToOutputWeights; params.m_RecurrentToForgetWeights = &recurrentToForgetWeights; params.m_RecurrentToCellWeights = &recurrentToCellWeights; params.m_RecurrentToOutputWeights = &recurrentToOutputWeights; params.m_ForgetGateBias = &forgetGateBias; params.m_CellBias = &cellBias; params.m_OutputGateBias = &outputGateBias; params.m_ProjectionWeights = &projectionWeights; params.m_ProjectionBias = &projectionBias; TestLstmLayerVisitor visitor(descriptor, params, layerName); Network net; IConnectableLayer *const layer = net.AddLstmLayer(descriptor, params, layerName); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckQuantizedLstmLayer) { std::vector inputToInputWeightsData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector inputToInputWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToInputWeights( TensorInfo(4, inputToInputWeightsDimensions.data(), DataType::QuantisedAsymm8), inputToInputWeightsData); std::vector inputToForgetWeightsData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector inputToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToForgetWeights( TensorInfo(4, inputToForgetWeightsDimensions.data(), DataType::QuantisedAsymm8), inputToForgetWeightsData); std::vector inputToCellWeightsData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector inputToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToCellWeights( TensorInfo(4, inputToCellWeightsDimensions.data(), DataType::QuantisedAsymm8), inputToCellWeightsData); std::vector inputToOutputWeightsData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector inputToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToOutputWeights( TensorInfo(4, inputToOutputWeightsDimensions.data(), DataType::QuantisedAsymm8), inputToOutputWeightsData); std::vector recurrentToInputWeightsData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector recurrentToInputWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToInputWeights(TensorInfo( 4, recurrentToInputWeightsDimensions.data(), DataType::QuantisedAsymm8), recurrentToInputWeightsData); std::vector recurrentToForgetWeightsData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector recurrentToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToForgetWeights(TensorInfo( 4, recurrentToForgetWeightsDimensions.data(), DataType::QuantisedAsymm8), recurrentToForgetWeightsData); std::vector recurrentToCellWeightsData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector recurrentToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToCellWeights(TensorInfo( 4, recurrentToCellWeightsDimensions.data(), DataType::QuantisedAsymm8), recurrentToCellWeightsData); std::vector recurrentToOutputWeightsData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector recurrentToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToOutputWeights(TensorInfo( 4, recurrentToOutputWeightsDimensions.data(), DataType::QuantisedAsymm8), recurrentToOutputWeightsData); std::vector inputGateBiasData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector inputGateBiasDimensions = {1, 1, 3, 3}; ConstTensor inputGateBias( TensorInfo(4, inputGateBiasDimensions.data(), DataType::Signed32), inputGateBiasData); std::vector forgetGateBiasData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector forgetGateBiasDimensions = {1, 1, 3, 3}; ConstTensor forgetGateBias(TensorInfo( 4, forgetGateBiasDimensions.data(), DataType::Signed32), forgetGateBiasData); std::vector cellBiasData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector cellBiasDimensions = {1, 1, 3, 3}; ConstTensor cellBias(TensorInfo( 4, cellBiasDimensions.data(), DataType::Signed32), cellBiasData); std::vector outputGateBiasData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector outputGateBiasDimensions = {1, 1, 3, 3}; ConstTensor outputGateBias(TensorInfo( 4, outputGateBiasDimensions.data(), DataType::Signed32), outputGateBiasData); QuantizedLstmInputParams params; params.m_InputToInputWeights = &inputToInputWeights; params.m_InputToForgetWeights = &inputToForgetWeights; params.m_InputToCellWeights = &inputToCellWeights; params.m_InputToOutputWeights = &inputToOutputWeights; params.m_RecurrentToInputWeights = &recurrentToInputWeights; params.m_RecurrentToForgetWeights = &recurrentToForgetWeights; params.m_RecurrentToCellWeights = &recurrentToCellWeights; params.m_RecurrentToOutputWeights = &recurrentToOutputWeights; params.m_InputGateBias = &inputGateBias; params.m_ForgetGateBias = &forgetGateBias; params.m_CellBias = &cellBias; params.m_OutputGateBias = &outputGateBias; TestQuantizedLstmLayerVisitor visitor(params); Network net; IConnectableLayer* const layer = net.AddQuantizedLstmLayer(params); layer->Accept(visitor); } BOOST_AUTO_TEST_CASE(CheckNamedQuantizedLstmLayer) { const char* layerName = "LstmLayer"; std::vector inputToInputWeightsData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector inputToInputWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToInputWeights( TensorInfo(4, inputToInputWeightsDimensions.data(), DataType::QuantisedAsymm8), inputToInputWeightsData); std::vector inputToForgetWeightsData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector inputToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToForgetWeights( TensorInfo(4, inputToForgetWeightsDimensions.data(), DataType::QuantisedAsymm8), inputToForgetWeightsData); std::vector inputToCellWeightsData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector inputToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToCellWeights( TensorInfo(4, inputToCellWeightsDimensions.data(), DataType::QuantisedAsymm8), inputToCellWeightsData); std::vector inputToOutputWeightsData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector inputToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor inputToOutputWeights( TensorInfo(4, inputToOutputWeightsDimensions.data(), DataType::QuantisedAsymm8), inputToOutputWeightsData); std::vector recurrentToInputWeightsData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector recurrentToInputWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToInputWeights(TensorInfo( 4, recurrentToInputWeightsDimensions.data(), DataType::QuantisedAsymm8), recurrentToInputWeightsData); std::vector recurrentToForgetWeightsData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector recurrentToForgetWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToForgetWeights(TensorInfo( 4, recurrentToForgetWeightsDimensions.data(), DataType::QuantisedAsymm8), recurrentToForgetWeightsData); std::vector recurrentToCellWeightsData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector recurrentToCellWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToCellWeights(TensorInfo( 4, recurrentToCellWeightsDimensions.data(), DataType::QuantisedAsymm8), recurrentToCellWeightsData); std::vector recurrentToOutputWeightsData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector recurrentToOutputWeightsDimensions = {1, 1, 3, 3}; ConstTensor recurrentToOutputWeights(TensorInfo( 4, recurrentToOutputWeightsDimensions.data(), DataType::QuantisedAsymm8), recurrentToOutputWeightsData); std::vector inputGateBiasData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector inputGateBiasDimensions = {1, 1, 3, 3}; ConstTensor inputGateBias( TensorInfo(4, inputGateBiasDimensions.data(), DataType::Signed32), inputGateBiasData); std::vector forgetGateBiasData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector forgetGateBiasDimensions = {1, 1, 3, 3}; ConstTensor forgetGateBias(TensorInfo( 4, forgetGateBiasDimensions.data(), DataType::Signed32), forgetGateBiasData); std::vector cellBiasData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector cellBiasDimensions = {1, 1, 3, 3}; ConstTensor cellBias(TensorInfo( 4, cellBiasDimensions.data(), DataType::Signed32), cellBiasData); std::vector outputGateBiasData = {1, 2, 3, 4, 5, 6, 7, 8, 9}; std::vector outputGateBiasDimensions = {1, 1, 3, 3}; ConstTensor outputGateBias(TensorInfo( 4, outputGateBiasDimensions.data(), DataType::Signed32), outputGateBiasData); QuantizedLstmInputParams params; params.m_InputToInputWeights = &inputToInputWeights; params.m_InputToForgetWeights = &inputToForgetWeights; params.m_InputToCellWeights = &inputToCellWeights; params.m_InputToOutputWeights = &inputToOutputWeights; params.m_RecurrentToInputWeights = &recurrentToInputWeights; params.m_RecurrentToForgetWeights = &recurrentToForgetWeights; params.m_RecurrentToCellWeights = &recurrentToCellWeights; params.m_RecurrentToOutputWeights = &recurrentToOutputWeights; params.m_InputGateBias = &inputGateBias; params.m_ForgetGateBias = &forgetGateBias; params.m_CellBias = &cellBias; params.m_OutputGateBias = &outputGateBias; TestQuantizedLstmLayerVisitor visitor(params, layerName); Network net; IConnectableLayer* const layer = net.AddQuantizedLstmLayer(params, layerName); layer->Accept(visitor); } BOOST_AUTO_TEST_SUITE_END() } // namespace armnn