7 #include "../TfLiteParser.hpp" 16 explicit GetBufferFixture()
21 "operator_codes": [ { "builtin_code": "CONV_2D" } ], 25 "shape": [ 1, 3, 3, 1 ], 28 "name": "inputTensor", 37 "shape": [ 1, 1, 1, 1 ], 40 "name": "outputTensor", 49 "shape": [ 1, 3, 3, 1 ], 52 "name": "filterTensor", 68 "builtin_options_type": "Conv2DOptions", 73 "fused_activation_function": "NONE" 75 "custom_options_format": "FLEXBUFFERS" 82 { "data": [ 2,1,0, 6,2,1, 4,1,2 ], }, 91 std::vector<int32_t> bufferValues,
size_t bufferIndex)
93 for(
long unsigned int i=0; i<bufferValues.size(); i++)
95 CHECK_EQ(TfLiteParserImpl::GetBuffer(model, bufferIndex)->data[i], bufferValues[i]);
104 m_GraphBinary.size());
105 std::vector<int32_t> bufferValues = {2,1,0,6,2,1,4,1,2};
106 CheckBufferContents(model, bufferValues, 2);
113 m_GraphBinary.size());
114 CHECK(TfLiteParserImpl::GetBuffer(model, 0)->data.empty());
115 CHECK(TfLiteParserImpl::GetBuffer(model, 1)->data.empty());
116 CHECK(!TfLiteParserImpl::GetBuffer(model, 2)->data.empty());
117 CHECK(TfLiteParserImpl::GetBuffer(model, 3)->data.empty());
124 m_GraphBinary.size());
TEST_SUITE("TensorflowLiteParser_GetBuffer")
bool ReadStringToBinary()
TEST_CASE_FIXTURE(ClContextControlFixture, "CopyBetweenNeonAndGpu")
std::unique_ptr< onnx::ModelProto > ModelPtr
Base class for all ArmNN exceptions so that users can filter to just those.