aboutsummaryrefslogtreecommitdiff
path: root/src/armnn/test/optimizations/AddBroadcastReshapeLayerTests.cpp
diff options
context:
space:
mode:
Diffstat (limited to 'src/armnn/test/optimizations/AddBroadcastReshapeLayerTests.cpp')
-rw-r--r--src/armnn/test/optimizations/AddBroadcastReshapeLayerTests.cpp4
1 files changed, 2 insertions, 2 deletions
diff --git a/src/armnn/test/optimizations/AddBroadcastReshapeLayerTests.cpp b/src/armnn/test/optimizations/AddBroadcastReshapeLayerTests.cpp
index 36a4507fc3..7573005518 100644
--- a/src/armnn/test/optimizations/AddBroadcastReshapeLayerTests.cpp
+++ b/src/armnn/test/optimizations/AddBroadcastReshapeLayerTests.cpp
@@ -290,7 +290,7 @@ TEST_CASE("ReshapeParentConstLayerTest")
{
Graph graph;
const TensorInfo info0({ 1, 2, 3, 5 }, DataType::QAsymmU8);
- const TensorInfo info1({ 5 }, DataType::QAsymmU8);
+ const TensorInfo info1({ 5 }, DataType::QAsymmU8, 0.0f, 0, true);
const TensorInfo outputInfo({ 1, 2, 3, 5 }, DataType::QAsymmU8);
auto input = graph.AddLayer<InputLayer>(0, "input");
@@ -346,7 +346,7 @@ TEST_CASE("ReshapeParentConstAddLayerMultipleConnectionsTest")
// What we'll do is have two sequential add layers both using the same const tensor.
Graph graph;
const TensorInfo inputInfo({ 1, 512 }, DataType::Float32);
- const TensorInfo constantTermInfo({ 1 }, DataType::Float32);
+ const TensorInfo constantTermInfo({ 1 }, DataType::Float32, 0.0f, 0, true);
const TensorInfo outputInfo({ 1, 512 }, DataType::Float32);
auto input = graph.AddLayer<InputLayer>(0, "input");