From 9b22517ba0cd6f767123583ce56e864f50e9d758 Mon Sep 17 00:00:00 2001 From: Jeremy Johnson Date: Tue, 14 Dec 2021 16:34:47 +0000 Subject: Add python package support Move tosa_serializer into its own namespace Fix up for pre-commit black/flake8 Remove import dependency on reference model Signed-off-by: Jeremy Johnson Change-Id: I8693fb7c00d224142a66dcb19eac74ac77c6224b --- python/serializer/__init__.py | 3 + python/serializer/tosa_serializer.py | 802 ++++++++++++++++++++++++++++++++++ python/tosa_serializer.py | 814 ----------------------------------- 3 files changed, 805 insertions(+), 814 deletions(-) create mode 100644 python/serializer/__init__.py create mode 100644 python/serializer/tosa_serializer.py delete mode 100644 python/tosa_serializer.py diff --git a/python/serializer/__init__.py b/python/serializer/__init__.py new file mode 100644 index 0000000..39e9ecc --- /dev/null +++ b/python/serializer/__init__.py @@ -0,0 +1,3 @@ +"""Namespace.""" +# Copyright (c) 2021-2022 Arm Limited. +# SPDX-License-Identifier: Apache-2.0 diff --git a/python/serializer/tosa_serializer.py b/python/serializer/tosa_serializer.py new file mode 100644 index 0000000..b29f963 --- /dev/null +++ b/python/serializer/tosa_serializer.py @@ -0,0 +1,802 @@ +# Copyright (c) 2020-2022, ARM Limited. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +import os +import json +import flatbuffers +import numpy as np +import struct +from enum import IntEnum, unique +from tosa import ( + TosaGraph, + TosaBasicBlock, + TosaTensor, + TosaOperator, + Version, +) +import tosa.DType as TosaDType +import tosa.Op as TosaOp + +# Keep version number in sync with the version default value with schema/tosa.fbs +TOSA_VERSION_MAJOR = 0 +TOSA_VERSION_MINOR = 24 +TOSA_VERSION_PATCH = 0 +TOSA_VERSION_DRAFT = True +TOSA_VERSION = [ + TOSA_VERSION_MAJOR, + TOSA_VERSION_MINOR, + TOSA_VERSION_PATCH, + TOSA_VERSION_DRAFT, +] +# With the way flatc generates its python types, there is no programatic way +# to get string names for the integer types. Manually maintain a string table +# here. +DType = TosaDType.DType() +DTypeNames = [ + "UNKNOWN", + "BOOL", + "UINT8", + "INT4", + "INT8", + "INT16", + "INT32", + "INT48", + "FLOAT", +] + +ByteMask = np.uint64(0xFF) + + +def dtype_str_to_val(name): + + for i in range(len(DTypeNames)): + if name.casefold() == DTypeNames[i].casefold(): + return i + raise Exception("Unable to parse DType name {}".format(name)) + + +class TosaSerializerUnion: + """This class handles encapsulating and serializing union types into flatbuffers""" + + def __init__(self): + + # A tuple of the start and end functions. + # Set by the options constructors below + self.optFcns = None + + # The type from the tosa.Options enumeration. + # Set by the options constructors below. + self.utype = None + + # Each of these lists is a tuple of the add function and the + # value being added. Set by the options constructors below. + self.ints = [] + self.bools = [] + self.floats = [] + self.strings = [] + self.intvecs = [] + self.fpvecs = [] + + def serialize(self, builder): + + # We have to build strings and vectors first + strList = [] + intVecList = [] + fpVecList = [] + + for fcn, val in self.strings: + strList.append((fcn, builder.CreateString(val))) + + for fcn, val in self.intvecs: + intVecList.append((fcn, TosaSerializer.serializeInt32Vec(builder, val))) + + for fcn, val in self.fpvecs: + fpVecList.append((fcn, TosaSerializer.serializeFpVec(builder, val))) + + startFcn, endFcn = self.optFcns + + # Then serialize the options object from the list of primitives and + # other serialized values + startFcn(builder) + for fcn, val in self.ints: + fcn(builder, val) + + for fcn, val in self.bools: + fcn(builder, val) + + for fcn, val in self.floats: + fcn(builder, val) + + for fcn, val in strList: + fcn(builder, val) + + for fcn, val in intVecList: + fcn(builder, val) + + for fcn, val in fpVecList: + fcn(builder, val) + + return endFcn(builder) + + +class TosaSerializerAttribute(TosaSerializerUnion): + """This class handles encapsulating all of the enumerated types for attributes""" + + def __init__(self): + super().__init__() + + def PoolAttribute(self, kernel, stride, padding): + from tosa import PoolAttribute as a, Attribute + + self.utype = Attribute.Attribute().PoolAttribute + + self.optFcns = (a.Start, a.End) + self.intvecs.append((a.AddPadding, padding)) + self.intvecs.append((a.AddKernel, kernel)) + self.intvecs.append((a.AddStride, stride)) + + def ConvAttribute(self, padding, stride, dilation): + from tosa import ConvAttribute as a, Attribute + + self.utype = Attribute.Attribute().ConvAttribute + self.optFcns = (a.Start, a.End) + + self.intvecs.append((a.AddPadding, padding)) + self.intvecs.append((a.AddStride, stride)) + self.intvecs.append((a.AddDilation, dilation)) + + def TransposeConvAttribute(self, outpad, stride, dilation, output_shape): + from tosa import TransposeConvAttribute as a, Attribute + + self.utype = Attribute.Attribute().TransposeConvAttribute + self.optFcns = (a.Start, a.End) + + self.intvecs.append((a.AddOutpad, outpad)) + self.intvecs.append((a.AddStride, stride)) + self.intvecs.append((a.AddDilation, dilation)) + self.intvecs.append((a.AddOutputShape, output_shape)) + + def PadAttribute(self, padding, pad_const_int, pad_const_fp): + from tosa import PadAttribute as a, Attribute + + self.utype = Attribute.Attribute().PadAttribute + self.optFcns = (a.Start, a.End) + + self.intvecs.append((a.AddPadding, padding)) + self.ints.append((a.AddPadConstInt, pad_const_int)) + self.floats.append((a.AddPadConstFp, pad_const_fp)) + + def AxisAttribute(self, axis): + from tosa import AxisAttribute as a, Attribute + + self.utype = Attribute.Attribute().AxisAttribute + self.optFcns = (a.Start, a.End) + + self.ints.append((a.AddAxis, axis)) + + def ReshapeAttribute(self, shape): + from tosa import ReshapeAttribute as a, Attribute + + self.utype = Attribute.Attribute().ReshapeAttribute + self.optFcns = (a.Start, a.End) + + self.intvecs.append((a.AddShape, shape)) + + def SliceAttribute(self, begin, size): + from tosa import SliceAttribute as a, Attribute + + self.utype = Attribute.Attribute().SliceAttribute + self.optFcns = (a.Start, a.End) + + self.intvecs.append((a.AddBegin, begin)) + self.intvecs.append((a.AddSize, size)) + + def TileAttribute(self, multiples): + from tosa import TileAttribute as a, Attribute + + self.utype = Attribute.Attribute().TileAttribute + self.optFcns = (a.Start, a.End) + + self.intvecs.append((a.AddMultiples, multiples)) + + def ResizeAttribute( + self, output_size, stride, offset, shift, stride_fp, offset_fp, mode + ): + from tosa import ResizeAttribute as a, Attribute + + self.utype = Attribute.Attribute().ResizeAttribute + self.optFcns = (a.Start, a.End) + + self.intvecs.append((a.AddOutputSize, output_size)) + self.intvecs.append((a.AddStride, stride)) + self.intvecs.append((a.AddOffset, offset)) + self.ints.append((a.AddShift, shift)) + self.fpvecs.append((a.AddStrideFp, stride_fp)) + self.fpvecs.append((a.AddOffsetFp, offset_fp)) + self.ints.append((a.AddMode, mode)) + + def ClampAttribute(self, minint, maxint, minfp, maxfp): + from tosa import ClampAttribute as a, Attribute + + self.utype = Attribute.Attribute().ClampAttribute + self.optFcns = (a.Start, a.End) + + self.ints.append((a.AddMinInt, minint)) + self.ints.append((a.AddMaxInt, maxint)) + + self.ints.append((a.AddMinFp, minfp)) + self.ints.append((a.AddMaxFp, maxfp)) + + def RescaleAttribute( + self, input_zp, output_zp, multiplier, shift, scale32, double_round, per_channel + ): + from tosa import RescaleAttribute as a, Attribute + + self.utype = Attribute.Attribute().RescaleAttribute + self.optFcns = (a.Start, a.End) + + self.ints.append((a.AddInputZp, input_zp)) + self.ints.append((a.AddOutputZp, output_zp)) + self.intvecs.append((a.AddMultiplier, multiplier)) + self.intvecs.append((a.AddShift, shift)) + self.bools.append((a.AddScale32, scale32)) + self.bools.append((a.AddDoubleRound, double_round)) + self.bools.append((a.AddPerChannel, per_channel)) + + def MulAttribute(self, shift): + from tosa import MulAttribute as a, Attribute + + self.utype = Attribute.Attribute().MulAttribute + self.optFcns = (a.Start, a.End) + + self.ints.append((a.AddShift, shift)) + + def ArithmeticRightShiftAttribute(self, round): + from tosa import ArithmeticRightShiftAttribute as a, Attribute + + self.utype = Attribute.Attribute().ArithmeticRightShiftAttribute + self.optFcns = ( + a.Start, + a.End, + ) + + self.bools.append((a.AddRound, round)) + + def CondIfAttribute(self, then_branch, else_branch): + from tosa import CondIfAttribute as a, Attribute + + self.utype = Attribute.Attribute().CondIfAttribute + self.optFcns = (a.Start, a.End) + + self.strings.append((a.AddThenBranch, then_branch)) + self.strings.append((a.AddElseBranch, else_branch)) + + def WhileLoopAttribute(self, cond_branch, body_branch): + from tosa import WhileLoopAttribute as a, Attribute + + self.utype = Attribute.Attribute().WhileLoopAttribute + self.optFcns = (a.Start, a.End) + + self.strings.append((a.AddCondBranch, cond_branch)) + self.strings.append((a.AddBodyBranch, body_branch)) + + def TransposeAttribute(self, perm): + from tosa import TransposeAttribute as a, Attribute + + self.utype = Attribute.Attribute().TransposeAttribute + self.optFcns = (a.Start, a.End) + + self.intvecs.append((a.AddPerm, perm)) + + def TableAttribute(self, table): + from tosa import TableAttribute as a, Attribute + + self.utype = Attribute.Attribute().TableAttribute + self.optFcns = (a.Start, a.End) + + self.intvecs.append((a.AddTable, table)) + + +class TosaSerializerQuantInfo(TosaSerializerUnion): + """This class handles encapsulating all of the enumerated types for quantinfo""" + + def __init__(self): + super().__init__() + + def ConvQuantInfo(self, input_zp, weight_zp): + from tosa import ConvQuantInfo as q, QuantInfo + + self.utype = QuantInfo.QuantInfo().ConvQuantInfo + self.optFcns = (q.Start, q.End) + self.ints.append((q.AddInputZp, input_zp)) + self.ints.append((q.AddWeightZp, weight_zp)) + + def UnaryQuantInfo(self, input_zp, output_zp): + from tosa import UnaryQuantInfo as q, QuantInfo + + self.utype = QuantInfo.QuantInfo().UnaryQuantInfo + self.optFcns = (q.Start, q.End) + self.ints.append((q.AddInputZp, input_zp)) + self.ints.append((q.AddOutputZp, output_zp)) + + def MatMulQuantInfo(self, a_zp, b_zp): + from tosa import MatMulQuantInfo as q, QuantInfo + + self.utype = QuantInfo.QuantInfo().MatMulQuantInfo + self.optFcns = (q.Start, q.End) + self.ints.append((q.AddAZp, a_zp)) + self.ints.append((q.AddBZp, b_zp)) + + def PadQuantInfo(self, input_zp): + from tosa import PadQuantInfo as q, QuantInfo + + self.utype = QuantInfo.QuantInfo().PadQuantInfo + self.optFcns = (q.Start, q.End) + self.ints.append((q.AddInputZp, input_zp)) + + +class TosaSerializerTensor: + def __init__( + self, + name, + shape, + dtype, + data=None, + placeholderFilename=None, + ): + self.name = name + + if isinstance(shape, np.ndarray): + shape = shape.astype(int).tolist() + shape = list(map(int, shape)) + + self.shape = shape + self.dtype = dtype + + if isinstance(data, np.ndarray): + data = data.flatten().astype(int).tolist() + data = list(map(int, data)) + self.data = data + elif isinstance(data, list): + data = list(map(int, data)) + self.data = data + else: + self.data = None + + # Filename for placeholder tensors. These get generated by the test generation + # process and are written to disk, but are considered input tensors by the + # network so they do not appear in the TOSA serialiazation. However, if we + # want to form a unit test around these input tensors, we can get the filename + # from here. + self.placeholderFilename = placeholderFilename + + def __str__(self): + str = "TosaSerializerTensor name: {} shape: {} dtype: {}".format( + self.name, + self.shape, + DTypeNames[self.dtype], + ) + return str + + def setDtype(self, dtype): + self.dtype = dtype + + def serialize(self, builder): + fb_name = builder.CreateString(self.name) + fb_shapes = TosaSerializer.serializeInt32Vec(builder, self.shape) + if self.data: + u8_data = list() + # little endianess + if self.dtype == DType.BOOL: + for val in self.data: + val_u8 = np.uint8(val) + u8_data.append(val_u8) + elif self.dtype == DType.INT4: + in_size = len(self.data) + out_size = (in_size + 1) // 2 + for i in range(out_size): + val_0 = self.data[2 * i] + if (2 * i + 1) < in_size: + val_1 = self.data[2 * i + 1] + else: + val_1 = 0 + val_i8 = (val_0 & 0xF) | ((val_1 & 0xF) << 4) + val_u8 = np.uint8(val_i8) + u8_data.append(val_u8) + elif self.dtype == DType.INT8: + for val in self.data: + val_u8 = np.uint8(val) + u8_data.append(val_u8) + elif self.dtype == DType.INT16: + for val in self.data: + val_u16 = np.uint16(val) + b0 = val_u16 & ByteMask + b1 = (val_u16 >> np.uint16(8)) & ByteMask + u8_data.extend([b0, b1]) + elif self.dtype == DType.INT32: + for val in self.data: + val_u32 = np.uint32(val) + b0 = val_u32 & ByteMask + b1 = (val_u32 >> np.uint32(8)) & ByteMask + b2 = (val_u32 >> np.uint32(16)) & ByteMask + b3 = (val_u32 >> np.uint32(24)) & ByteMask + u8_data.extend([b0, b1, b2, b3]) + elif self.dtype == DType.INT48: + for val in self.data: + val_u64 = np.uint64(val) + b0 = val_u64 & ByteMask + b1 = (val_u64 >> np.uint64(8)) & ByteMask + b2 = (val_u64 >> np.uint64(16)) & ByteMask + b3 = (val_u64 >> np.uint64(24)) & ByteMask + b4 = (val_u64 >> np.uint64(32)) & ByteMask + b5 = (val_u64 >> np.uint64(40)) & ByteMask + u8_data.extend([b0, b1, b2, b3, b4, b5]) + elif self.dtype == DType.FLOAT: + for val in self.data: + b = struct.pack("!f", val) + u8_data.extend([b[3], b[2], b[1], b[0]]) + else: + raise Exception( + "unsupported data type {}".format(DTypeNames[self.dtype]) + ) + fb_data = TosaSerializer.serializeUint8Vec(builder, u8_data) + + TosaTensor.Start(builder) + TosaTensor.AddName(builder, fb_name) + TosaTensor.AddShape(builder, fb_shapes) + TosaTensor.AddType(builder, self.dtype) + if self.data: + TosaTensor.AddData(builder, fb_data) + + return TosaTensor.End(builder) + + +class TosaSerializerOperator: + def __init__(self, op, inputs, outputs, attributes=None, quantInfo=None): + self.op = op + self.attributes = attributes + self.inputs = TosaSerializer.toList(inputs) + self.outputs = TosaSerializer.toList(outputs) + self.quantInfo = quantInfo + + def __str__(self): + str = "Op {}\n----\n".format(self.op) + + for i in self.inputs: + str = str + " Input: {}\n".format(i) + for o in self.outputs: + str = str + " Output: {}\n".format(o) + + return str + + def serialize(self, builder): + fb_inputs = TosaSerializer.serializeStrVec( + builder, self.inputs, TosaOperator.StartInputsVector + ) + fb_outputs = TosaSerializer.serializeStrVec( + builder, self.outputs, TosaOperator.StartOutputsVector + ) + # Need to serialize quant_info and attributes enums still + if self.attributes is not None: + fb_attributes = self.attributes.serialize(builder) + + if self.quantInfo is not None: + fb_qinfo = self.quantInfo.serialize(builder) + + TosaOperator.Start(builder) + TosaOperator.AddOp(builder, self.op) + TosaOperator.AddInputs(builder, fb_inputs) + TosaOperator.AddOutputs(builder, fb_outputs) + if self.attributes is not None: + TosaOperator.AddAttributeType(builder, self.attributes.utype) + TosaOperator.AddAttribute(builder, fb_attributes) + if self.quantInfo is not None: + TosaOperator.AddQuantInfoType(builder, self.quantInfo.utype) + TosaOperator.AddQuantInfo(builder, fb_qinfo) + + return TosaOperator.End(builder) + + +class TosaSerializerBasicBlock: + def __init__(self, name): + self.name = name + self.operators = [] + + # Dict assures uniqueness, but allows us to look up by name + self.tensors = dict() + + self.inputs = [] + self.outputs = [] + + def addTensor( + self, + name, + shape, + dtype, + data=None, + placeholderFilename=None, + ): + if name not in self.tensors: + self.tensors[name] = TosaSerializerTensor( + name, shape, dtype, data, placeholderFilename + ) + + return self.tensors[name] + + def addInput(self, name): + self.inputs.append(name) + + def addOutput(self, name): + self.outputs.append(name) + + def addOperator(self, op, inputs, outputs, attributes=None, quant_info=None): + self.operators.append( + TosaSerializerOperator(op, inputs, outputs, attributes, quant_info) + ) + + def serialize(self, builder): + fb_name = builder.CreateString(self.name) + fbv_inputs = TosaSerializer.serializeStrVec( + builder, list(self.inputs), TosaBasicBlock.StartInputsVector + ) + fbv_outputs = TosaSerializer.serializeStrVec( + builder, list(self.outputs), TosaBasicBlock.StartOutputsVector + ) + fbv_tensors = TosaSerializer.serializeObjVec( + builder, + list(self.tensors.values()), + TosaBasicBlock.StartTensorsVector, + ) + fbv_operators = TosaSerializer.serializeObjVec( + builder, self.operators, TosaBasicBlock.StartOperatorsVector + ) + + TosaBasicBlock.Start(builder) + TosaBasicBlock.AddName(builder, fb_name) + TosaBasicBlock.AddInputs(builder, fbv_inputs) + TosaBasicBlock.AddOutputs(builder, fbv_outputs) + TosaBasicBlock.AddTensors(builder, fbv_tensors) + TosaBasicBlock.AddOperators(builder, fbv_operators) + return TosaBasicBlock.End(builder) + + +@unique +class TensorDir(IntEnum): + PLACEHOLDER = 0 + CONST = 1 + INTERMEDIATE = 2 + RESULT = 3 + + +class TosaSerializer: + def __init__(self, pathPrefix): + + # Get the global TOSA version if not already defined + + self.builder = flatbuffers.Builder(0) + + self.basicBlocks = [] + self.startBasicBlock("main") + self.pathPrefix = pathPrefix + + # Indicies used for adding/naming tensors + self.currInputIdx = 0 + self.currConstIdx = 0 + self.currLayerIdx = 1 + self.currResultIdx = 0 + + # Is this an illegal test that is expected to fail? + self.expectedReturnCode = 0 + self.expectedFailure = False + self.expectedFailureDesc = "" + + def __str__(self): + str = "" + for bb in self.basicBlocks: + str = str + bb.__str__() + return str + + def addPlaceholder(self, shape, dtype, vals): + if not self.currBasicBlock: + raise Exception("addTensor called without valid basic block") + + name = "input-{}".format(self.currInputIdx) + filename = "{}.npy".format(name) + self.currInputIdx = self.currInputIdx + 1 + + tens = self.currBasicBlock.addTensor(name, shape, dtype, None, filename) + # This is always an input to the block + self.currBasicBlock.addInput(name) + + if vals is not None: + np.save(os.path.join(self.pathPrefix, filename), vals, False) + + return tens + + def addConst(self, shape, dtype, vals): + if not self.currBasicBlock: + raise Exception("addTensor called without valid basic block") + + name = "const-{}".format(self.currInputIdx) + self.currInputIdx = self.currInputIdx + 1 + + tens = self.currBasicBlock.addTensor(name, shape, dtype, vals) + # Add the operator now + self.currBasicBlock.addOperator(TosaOp.Op().CONST, [], name) + + return tens + + def addIntermediate(self, shape, dtype): + + if not self.currBasicBlock: + raise Exception("addTensor called without valid basic block") + + name = "layer-{}".format(self.currLayerIdx) + self.currLayerIdx = self.currLayerIdx + 1 + + tens = self.currBasicBlock.addTensor(name, shape, dtype, None) + + return tens + + def addInputTensor(self, tensor): + self.currBasicBlock.addTensor(tensor.name, tensor.shape, tensor.dtype) + self.currBasicBlock.addInput(tensor.name) + + def addOutputTensor(self, tensor): + self.currBasicBlock.addOutput(tensor.name) + + def addOutput(self, shape, dtype): + if not self.currBasicBlock: + raise Exception("addTensor called without valid basic block") + + name = "result-{}".format(self.currResultIdx) + self.currResultIdx = self.currResultIdx + 1 + + tens = self.currBasicBlock.addTensor(name, shape, dtype, None) + self.currBasicBlock.addOutput(name) + return tens + + def addOperator(self, op, inputs, outputs, attributes=None, quant_info=None): + + if op == TosaOp.Op().CONST: + raise Exception("Use addConstTensor() to add CONST ops") + + return self.currBasicBlock.addOperator( + op, inputs, outputs, attributes, quant_info + ) + + def setExpectedReturnCode(self, val, fail, desc=""): + + self.expectedReturnCode = val + self.expectedFailureDesc = desc + self.expectedFailure = fail + + def serialize(self): + + builder = self.builder + + Version.Start(builder) + Version.Add_major(builder, TOSA_VERSION[0]) + Version.Add_minor(builder, TOSA_VERSION[1]) + Version.Add_patch(builder, TOSA_VERSION[2]) + Version.Add_draft(builder, TOSA_VERSION[3]) + version = Version.End(builder) + + fbv_bb = TosaSerializer.serializeObjVec( + builder, self.basicBlocks, TosaGraph.StartBlocksVector + ) + + TosaGraph.Start(builder) + TosaGraph.AddVersion(builder, version) + TosaGraph.AddBlocks(builder, fbv_bb) + graph = TosaGraph.End(builder) + + self.builder.Finish(graph) + return self.builder.Output() + + def writeJson(self, tosa_filename): + """Write a json test file so that it is fairly easy to pick up the test + and generate commands for third party tool""" + test_desc = dict() + + test_desc["tosa_file"] = tosa_filename + ifm_name = [] + ifm_file = [] + ofm_name = [] + ofm_file = [] + + for b in self.basicBlocks: + if b.name == "main": + for i in b.inputs: + ifm_name.append(i) + ifm_file.append(b.tensors[i].placeholderFilename) + for o in b.outputs: + ofm_name.append(o) + # Make up an OFM filename here. One isn't generated until the + # reference tool is run, so any name is a good name + ofm_file.append("ref-{}.npy".format(o)) + + test_desc["ifm_name"] = ifm_name + test_desc["ifm_file"] = ifm_file + test_desc["ofm_name"] = ofm_name + test_desc["ofm_file"] = ofm_file + test_desc["expected_return_code"] = self.expectedReturnCode + test_desc["expected_failure"] = self.expectedFailure + if self.expectedFailureDesc: + test_desc["expected_failure_desc"] = self.expectedFailureDesc + + return json.dumps(test_desc, indent=" ") + + def startBasicBlock(self, name): + self.currBasicBlock = TosaSerializerBasicBlock(name) + self.basicBlocks.append(self.currBasicBlock) + + @staticmethod + def serializeStrVec(builder, vec, start_fcn): + fb_strs = [builder.CreateString(i) for i in vec] + start_fcn(builder, len(fb_strs)) + for s in fb_strs[::-1]: + builder.PrependUOffsetTRelative(s) + return builder.EndVector() + + @staticmethod + def serializeUint8Vec(builder, vec): + builder.StartVector(1, len(vec), 8) + for v in vec[::-1]: + builder.PrependUint8(v) + try: + return builder.EndVector() + except TypeError: + return builder.EndVector(len(vec)) + + @staticmethod + def serializeInt32Vec(builder, vec): + builder.StartVector(4, len(vec), 4) + for v in vec[::-1]: + builder.PrependInt32(v) + try: + return builder.EndVector() + except TypeError: + return builder.EndVector(len(vec)) + + @staticmethod + def serializeFpVec(builder, vec): + builder.StartVector(4, len(vec), 4) + for v in vec[::-1]: + builder.PrependFloat32(v) + try: + return builder.EndVector() + except TypeError: + return builder.EndVector(len(vec)) + + @staticmethod + def serializeObjVec(builder, vec, start_fcn): + serialized_vec = [] + for v in vec[::-1]: + serialized_vec.append(v.serialize(builder)) + + start_fcn(builder, len(vec)) + for v in serialized_vec: + builder.PrependUOffsetTRelative(v) + try: + return builder.EndVector() + except TypeError: + return builder.EndVector(len(vec)) + + @staticmethod + def toList(val): + if isinstance(val, list): + return val + else: + return [val] diff --git a/python/tosa_serializer.py b/python/tosa_serializer.py deleted file mode 100644 index f294ba3..0000000 --- a/python/tosa_serializer.py +++ /dev/null @@ -1,814 +0,0 @@ -# Copyright (c) 2020-2021, ARM Limited. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -#!/usr/bin/env python3 - -import os -import sys -import json -import flatbuffers -import numpy as np -import struct -from enum import Enum, IntEnum, unique -from tosa import ( - TosaGraph, - TosaBasicBlock, - TosaTensor, - TosaOperator, - DType, - Op, - ResizeMode, - Version, -) -from tosa_ref_run import TosaReturnCode - -import tosa - -# Keep version number in sync with the version default value with schema/tosa.fbs -TOSA_VERSION_MAJOR = 0 -TOSA_VERSION_MINOR = 24 -TOSA_VERSION_PATCH = 0 -TOSA_VERSION_DRAFT = True -TOSA_VERSION = [TOSA_VERSION_MAJOR, - TOSA_VERSION_MINOR, - TOSA_VERSION_PATCH, - TOSA_VERSION_DRAFT] -# With the way flatc generates its python types, there is no programatic way -# to get string names for the integer types. Manually maintain a string table -# here. -DType = tosa.DType.DType() -DTypeNames = [ - "UNKNOWN", - "BOOL", - "UINT8", - "INT4", - "INT8", - "INT16", - "INT32", - "INT48", - "FLOAT", -] - -ByteMask = np.uint64(0xFF) - - -def dtype_str_to_val(name): - - for i in range(len(DTypeNames)): - if name.casefold() == DTypeNames[i].casefold(): - return i - raise Exception("Unable to parse DType name {}".format(name)) - - -class TosaSerializerUnion: - """This class handles encapsulating and serializing union types into flatbuffers""" - - def __init__(self): - - # A tuple of the start and end functions. Set by the options constructors below - self.optFcns = None - - # The type from the tosa.Options enumeration. Set by the options constructors below. - self.utype = None - - # Each of these lists is a tuple of the add function and the - # value being added. Set by the options constructors below. - self.ints = [] - self.bools = [] - self.floats = [] - self.strings = [] - self.intvecs = [] - self.fpvecs = [] - - def serialize(self, builder): - - # We have to build strings and vectors first - strList = [] - intVecList = [] - fpVecList = [] - - for fcn, val in self.strings: - strList.append((fcn, builder.CreateString(val))) - - for fcn, val in self.intvecs: - intVecList.append((fcn, TosaSerializer.serializeInt32Vec(builder, val))) - - for fcn, val in self.fpvecs: - fpVecList.append((fcn, TosaSerializer.serializeFpVec(builder, val))) - - startFcn, endFcn = self.optFcns - - # Then serialize the options object from the list of primitives and - # other serialized values - startFcn(builder) - for fcn, val in self.ints: - fcn(builder, val) - - for fcn, val in self.bools: - fcn(builder, val) - - for fcn, val in self.floats: - fcn(builder, val) - - for fcn, val in strList: - fcn(builder, val) - - for fcn, val in intVecList: - fcn(builder, val) - - for fcn, val in fpVecList: - fcn(builder, val) - - return endFcn(builder) - - -class TosaSerializerAttribute(TosaSerializerUnion): - """This class handles encapsulating all of the enumerated types for attributes""" - - def __init__(self): - super().__init__() - - def PoolAttribute(self, kernel, stride, padding): - from tosa import PoolAttribute as a, Attribute - - self.utype = Attribute.Attribute().PoolAttribute - - self.optFcns = (a.Start, a.End) - self.intvecs.append((a.AddPadding, padding)) - self.intvecs.append((a.AddKernel, kernel)) - self.intvecs.append((a.AddStride, stride)) - - def ConvAttribute(self, padding, stride, dilation): - from tosa import ConvAttribute as a, Attribute - - self.utype = Attribute.Attribute().ConvAttribute - self.optFcns = (a.Start, a.End) - - self.intvecs.append((a.AddPadding, padding)) - self.intvecs.append((a.AddStride, stride)) - self.intvecs.append((a.AddDilation, dilation)) - - def TransposeConvAttribute(self, outpad, stride, dilation, output_shape): - from tosa import TransposeConvAttribute as a, Attribute - - self.utype = Attribute.Attribute().TransposeConvAttribute - self.optFcns = (a.Start, a.End) - - self.intvecs.append((a.AddOutpad, outpad)) - self.intvecs.append((a.AddStride, stride)) - self.intvecs.append((a.AddDilation, dilation)) - self.intvecs.append((a.AddOutputShape, output_shape)) - - def PadAttribute(self, padding, pad_const_int, pad_const_fp): - from tosa import PadAttribute as a, Attribute - - self.utype = Attribute.Attribute().PadAttribute - self.optFcns = (a.Start, a.End) - - self.intvecs.append((a.AddPadding, padding)) - self.ints.append((a.AddPadConstInt, pad_const_int)) - self.floats.append((a.AddPadConstFp, pad_const_fp)) - - def AxisAttribute(self, axis): - from tosa import AxisAttribute as a, Attribute - - self.utype = Attribute.Attribute().AxisAttribute - self.optFcns = (a.Start, a.End) - - self.ints.append((a.AddAxis, axis)) - - def ReshapeAttribute(self, shape): - from tosa import ReshapeAttribute as a, Attribute - - self.utype = Attribute.Attribute().ReshapeAttribute - self.optFcns = (a.Start, a.End) - - self.intvecs.append((a.AddShape, shape)) - - def SliceAttribute(self, begin, size): - from tosa import SliceAttribute as a, Attribute - - self.utype = Attribute.Attribute().SliceAttribute - self.optFcns = (a.Start, a.End) - - self.intvecs.append((a.AddBegin, begin)) - self.intvecs.append((a.AddSize, size)) - - def TileAttribute(self, multiples): - from tosa import TileAttribute as a, Attribute - - self.utype = Attribute.Attribute().TileAttribute - self.optFcns = (a.Start, a.End) - - self.intvecs.append((a.AddMultiples, multiples)) - - def ResizeAttribute( - self, output_size, stride, offset, shift, stride_fp, offset_fp, mode - ): - from tosa import ResizeAttribute as a, Attribute - - self.utype = Attribute.Attribute().ResizeAttribute - self.optFcns = (a.Start, a.End) - - self.intvecs.append((a.AddOutputSize, output_size)) - self.intvecs.append((a.AddStride, stride)) - self.intvecs.append((a.AddOffset, offset)) - self.ints.append((a.AddShift, shift)) - self.fpvecs.append((a.AddStrideFp, stride_fp)) - self.fpvecs.append((a.AddOffsetFp, offset_fp)) - self.ints.append((a.AddMode, mode)) - - def ClampAttribute(self, minint, maxint, minfp, maxfp): - from tosa import ClampAttribute as a, Attribute - - self.utype = Attribute.Attribute().ClampAttribute - self.optFcns = (a.Start, a.End) - - self.ints.append((a.AddMinInt, minint)) - self.ints.append((a.AddMaxInt, maxint)) - - self.ints.append((a.AddMinFp, minfp)) - self.ints.append((a.AddMaxFp, maxfp)) - - def RescaleAttribute( - self, input_zp, output_zp, multiplier, shift, scale32, double_round, per_channel - ): - from tosa import RescaleAttribute as a, Attribute - - self.utype = Attribute.Attribute().RescaleAttribute - self.optFcns = (a.Start, a.End) - - self.ints.append((a.AddInputZp, input_zp)) - self.ints.append((a.AddOutputZp, output_zp)) - self.intvecs.append((a.AddMultiplier, multiplier)) - self.intvecs.append((a.AddShift, shift)) - self.bools.append((a.AddScale32, scale32)) - self.bools.append((a.AddDoubleRound, double_round)) - self.bools.append((a.AddPerChannel, per_channel)) - - def MulAttribute(self, shift): - from tosa import MulAttribute as a, Attribute - - self.utype = Attribute.Attribute().MulAttribute - self.optFcns = (a.Start, a.End) - - self.ints.append((a.AddShift, shift)) - - def ArithmeticRightShiftAttribute(self, round): - from tosa import ArithmeticRightShiftAttribute as a, Attribute - - self.utype = Attribute.Attribute().ArithmeticRightShiftAttribute - self.optFcns = ( - a.Start, - a.End, - ) - - self.bools.append((a.AddRound, round)) - - def CondIfAttribute(self, then_branch, else_branch): - from tosa import CondIfAttribute as a, Attribute - - self.utype = Attribute.Attribute().CondIfAttribute - self.optFcns = (a.Start, a.End) - - self.strings.append((a.AddThenBranch, then_branch)) - self.strings.append((a.AddElseBranch, else_branch)) - - def WhileLoopAttribute(self, cond_branch, body_branch): - from tosa import WhileLoopAttribute as a, Attribute - - self.utype = Attribute.Attribute().WhileLoopAttribute - self.optFcns = (a.Start, a.End) - - self.strings.append((a.AddCondBranch, cond_branch)) - self.strings.append((a.AddBodyBranch, body_branch)) - - def TransposeAttribute(self, perm): - from tosa import TransposeAttribute as a, Attribute - - self.utype = Attribute.Attribute().TransposeAttribute - self.optFcns = (a.Start, a.End) - - self.intvecs.append((a.AddPerm, perm)) - - def TableAttribute(self, table): - from tosa import TableAttribute as a, Attribute - - self.utype = Attribute.Attribute().TableAttribute - self.optFcns = (a.Start, a.End) - - self.intvecs.append((a.AddTable, table)) - -class TosaSerializerQuantInfo(TosaSerializerUnion): - """This class handles encapsulating all of the enumerated types for quantinfo types""" - - def __init__(self): - super().__init__() - - def ConvQuantInfo(self, input_zp, weight_zp): - from tosa import ConvQuantInfo as q, QuantInfo - - self.utype = QuantInfo.QuantInfo().ConvQuantInfo - self.optFcns = (q.Start, q.End) - self.ints.append((q.AddInputZp, input_zp)) - self.ints.append((q.AddWeightZp, weight_zp)) - - def UnaryQuantInfo(self, input_zp, output_zp): - from tosa import UnaryQuantInfo as q, QuantInfo - - self.utype = QuantInfo.QuantInfo().UnaryQuantInfo - self.optFcns = (q.Start, q.End) - self.ints.append((q.AddInputZp, input_zp)) - self.ints.append((q.AddOutputZp, output_zp)) - - def MatMulQuantInfo(self, a_zp, b_zp): - from tosa import MatMulQuantInfo as q, QuantInfo - - self.utype = QuantInfo.QuantInfo().MatMulQuantInfo - self.optFcns = (q.Start, q.End) - self.ints.append((q.AddAZp, a_zp)) - self.ints.append((q.AddBZp, b_zp)) - - def PadQuantInfo(self, input_zp): - from tosa import PadQuantInfo as q, QuantInfo - - self.utype = QuantInfo.QuantInfo().PadQuantInfo - self.optFcns = (q.Start, q.End) - self.ints.append((q.AddInputZp, input_zp)) - - -class TosaSerializerTensor: - def __init__( - self, - name, - shape, - dtype, - data=None, - placeholderFilename=None, - ): - self.name = name - - if isinstance(shape, np.ndarray): - shape = shape.astype(int).tolist() - shape = list(map(int, shape)) - - self.shape = shape - self.dtype = dtype - - if isinstance(data, np.ndarray): - data = data.flatten().astype(int).tolist() - data = list(map(int, data)) - self.data = data - elif isinstance(data, list): - data = list(map(int, data)) - self.data = data - else: - self.data = None - - # Filename for placeholder tensors. These get generated by the test generation - # process and are written to disk, but are considered input tensors by the network - # so they do not appear in the TOSA serialiazation. However, if we want to form a unit - # test around these input tensors, we can get the filename from here. - self.placeholderFilename = placeholderFilename - - def __str__(self): - str = "TosaSerializerTensor name: {} shape: {} dtype: {}".format( - self.name, - self.shape, - DTypeNames[self.dtype], - ) - return str - - def setDtype(self, dtype): - self.dtype = dtype - - def serialize(self, builder): - fb_name = builder.CreateString(self.name) - fb_shapes = TosaSerializer.serializeInt32Vec(builder, self.shape) - if self.data: - u8_data = list() - # little endianess - if self.dtype == DType.BOOL: - for val in self.data: - val_u8 = np.uint8(val) - u8_data.append(val_u8) - elif self.dtype == DType.INT4: - in_size = len(self.data) - out_size = (in_size + 1) // 2 - for i in range(out_size): - val_0 = self.data[2 * i] - if (2 * i + 1) < in_size: - val_1 = self.data[2 * i + 1] - else: - val_1 = 0 - val_i8 = (val_0 & 0xF) | ((val_1 & 0xF) << 4) - val_u8 = np.uint8(val_i8) - u8_data.append(val_u8) - elif self.dtype == DType.INT8: - for val in self.data: - val_u8 = np.uint8(val) - u8_data.append(val_u8) - elif self.dtype == DType.INT16: - for val in self.data: - val_u16 = np.uint16(val) - b0 = val_u16 & ByteMask - b1 = (val_u16 >> np.uint16(8)) & ByteMask - u8_data.extend([b0, b1]) - elif self.dtype == DType.INT32: - for val in self.data: - val_u32 = np.uint32(val) - b0 = val_u32 & ByteMask - b1 = (val_u32 >> np.uint32(8)) & ByteMask - b2 = (val_u32 >> np.uint32(16)) & ByteMask - b3 = (val_u32 >> np.uint32(24)) & ByteMask - u8_data.extend([b0, b1, b2, b3]) - elif self.dtype == DType.INT48: - for val in self.data: - val_u64 = np.uint64(val) - b0 = val_u64 & ByteMask - b1 = (val_u64 >> np.uint64(8)) & ByteMask - b2 = (val_u64 >> np.uint64(16)) & ByteMask - b3 = (val_u64 >> np.uint64(24)) & ByteMask - b4 = (val_u64 >> np.uint64(32)) & ByteMask - b5 = (val_u64 >> np.uint64(40)) & ByteMask - u8_data.extend([b0, b1, b2, b3, b4, b5]) - elif self.dtype == DType.FLOAT: - for val in self.data: - b = struct.pack("!f", val) - u8_data.extend([b[3], b[2], b[1], b[0]]) - else: - raise Exception( - "unsupported data type {}".format(DTypeNames[self.dtype]) - ) - fb_data = TosaSerializer.serializeUint8Vec(builder, u8_data) - - TosaTensor.Start(builder) - TosaTensor.AddName(builder, fb_name) - TosaTensor.AddShape(builder, fb_shapes) - TosaTensor.AddType(builder, self.dtype) - if self.data: - TosaTensor.AddData(builder, fb_data) - - return TosaTensor.End(builder) - - -class TosaSerializerOperator: - def __init__(self, op, inputs, outputs, attributes=None, quantInfo=None): - self.op = op - self.attributes = attributes - self.inputs = TosaSerializer.toList(inputs) - self.outputs = TosaSerializer.toList(outputs) - self.quantInfo = quantInfo - - def __str__(self): - str = "Op {}\n----\n".format(self.op) - - for i in self.inputs: - str = str + " Input: {}\n".format(i) - for o in self.outputs: - str = str + " Output: {}\n".format(o) - - return str - - def serialize(self, builder): - fb_inputs = TosaSerializer.serializeStrVec( - builder, self.inputs, TosaOperator.StartInputsVector - ) - fb_outputs = TosaSerializer.serializeStrVec( - builder, self.outputs, TosaOperator.StartOutputsVector - ) - # Need to serialize quant_info and attributes enums still - if self.attributes is not None: - fb_attributes = self.attributes.serialize(builder) - - if self.quantInfo is not None: - fb_qinfo = self.quantInfo.serialize(builder) - - TosaOperator.Start(builder) - TosaOperator.AddOp(builder, self.op) - TosaOperator.AddInputs(builder, fb_inputs) - TosaOperator.AddOutputs(builder, fb_outputs) - if self.attributes is not None: - TosaOperator.AddAttributeType(builder, self.attributes.utype) - TosaOperator.AddAttribute(builder, fb_attributes) - if self.quantInfo is not None: - TosaOperator.AddQuantInfoType(builder, self.quantInfo.utype) - TosaOperator.AddQuantInfo(builder, fb_qinfo) - - return TosaOperator.End(builder) - - -class TosaSerializerBasicBlock: - def __init__(self, name): - self.name = name - self.operators = [] - - # Dict assures uniqueness, but allows us to look up by name - self.tensors = dict() - - self.inputs = [] - self.outputs = [] - - def addTensor( - self, - name, - shape, - dtype, - data=None, - placeholderFilename=None, - ): - try: - # Someone already added this tensor. - tens = self.tensors[name] - except KeyError: - self.tensors[name] = TosaSerializerTensor( - name, shape, dtype, data, placeholderFilename - ) - - return self.tensors[name] - - def addInput(self, name): - self.inputs.append(name) - - def addOutput(self, name): - self.outputs.append(name) - - def addOperator(self, op, inputs, outputs, attributes=None, quant_info=None): - self.operators.append( - TosaSerializerOperator(op, inputs, outputs, attributes, quant_info) - ) - - def serialize(self, builder): - fb_name = builder.CreateString(self.name) - fbv_inputs = TosaSerializer.serializeStrVec( - builder, list(self.inputs), TosaBasicBlock.StartInputsVector - ) - fbv_outputs = TosaSerializer.serializeStrVec( - builder, list(self.outputs), TosaBasicBlock.StartOutputsVector - ) - fbv_tensors = TosaSerializer.serializeObjVec( - builder, - list(self.tensors.values()), - TosaBasicBlock.StartTensorsVector, - ) - fbv_operators = TosaSerializer.serializeObjVec( - builder, self.operators, TosaBasicBlock.StartOperatorsVector - ) - - TosaBasicBlock.Start(builder) - TosaBasicBlock.AddName(builder, fb_name) - TosaBasicBlock.AddInputs(builder, fbv_inputs) - TosaBasicBlock.AddOutputs(builder, fbv_outputs) - TosaBasicBlock.AddTensors(builder, fbv_tensors) - TosaBasicBlock.AddOperators(builder, fbv_operators) - return TosaBasicBlock.End(builder) - - -@unique -class TensorDir(IntEnum): - PLACEHOLDER = 0 - CONST = 1 - INTERMEDIATE = 2 - RESULT = 3 - - -class TosaSerializer: - def __init__(self, pathPrefix): - - # Get the global TOSA version if not already defined - - self.builder = flatbuffers.Builder(0) - - self.basicBlocks = [] - self.startBasicBlock("main") - self.pathPrefix = pathPrefix - - # Indicies used for adding/naming tensors - self.currInputIdx = 0 - self.currConstIdx = 0 - self.currLayerIdx = 1 - self.currResultIdx = 0 - - # Is this an illegal test that is expected to fail? - self.expectedReturnCode = TosaReturnCode.VALID - self.expectedFailure = False - self.expectedFailureDesc = "" - - def __str__(self): - str = "" - for bb in self.basicBlocks: - str = str + bb.__str__() - return str - - def addPlaceholder(self, shape, dtype, vals): - if not self.currBasicBlock: - raise Exception("addTensor called without valid basic block") - - name = "input-{}".format(self.currInputIdx) - filename = "{}.npy".format(name) - self.currInputIdx = self.currInputIdx + 1 - - tens = self.currBasicBlock.addTensor(name, shape, dtype, None, filename) - # This is always an input to the block - self.currBasicBlock.addInput(name) - - if vals is not None: - np.save(os.path.join(self.pathPrefix, filename), vals, False) - - return tens - - def addConst(self, shape, dtype, vals): - if not self.currBasicBlock: - raise Exception("addTensor called without valid basic block") - - name = "const-{}".format(self.currInputIdx) - filename = "{}.npy".format(name) - self.currInputIdx = self.currInputIdx + 1 - - tens = self.currBasicBlock.addTensor(name, shape, dtype, vals) - # Add the operator now - self.currBasicBlock.addOperator(tosa.Op.Op().CONST, [], name) - - return tens - - def addIntermediate(self, shape, dtype): - - if not self.currBasicBlock: - raise Exception("addTensor called without valid basic block") - - name = "layer-{}".format(self.currLayerIdx) - self.currLayerIdx = self.currLayerIdx + 1 - - tens = self.currBasicBlock.addTensor(name, shape, dtype, None) - - return tens - - def addInputTensor(self, tensor): - self.currBasicBlock.addTensor(tensor.name, tensor.shape, tensor.dtype) - self.currBasicBlock.addInput(tensor.name) - - def addOutputTensor(self, tensor): - self.currBasicBlock.addOutput(tensor.name) - - def addOutput(self, shape, dtype): - if not self.currBasicBlock: - raise Exception("addTensor called without valid basic block") - - name = "result-{}".format(self.currResultIdx) - self.currResultIdx = self.currResultIdx + 1 - - tens = self.currBasicBlock.addTensor(name, shape, dtype, None) - self.currBasicBlock.addOutput(name) - return tens - - def addOperator(self, op, inputs, outputs, attributes=None, quant_info=None): - - if op == tosa.Op.Op().CONST: - raise Exception("Use addConstTensor() to add CONST ops") - - return self.currBasicBlock.addOperator( - op, inputs, outputs, attributes, quant_info - ) - - def setExpectedReturnCode(self, val, desc=""): - - self.expectedReturnCode = val - self.expectedFailureDesc = desc - - if val == TosaReturnCode.VALID: - self.expectedFailure = False - else: - # Unpredictable or error results are considered expected failures - # for conformance - self.expectedFailure = True - - def serialize(self): - - builder = self.builder - - Version.Start(builder) - Version.Add_major(builder, TOSA_VERSION[0]) - Version.Add_minor(builder, TOSA_VERSION[1]) - Version.Add_patch(builder, TOSA_VERSION[2]) - Version.Add_draft(builder, TOSA_VERSION[3]) - version = Version.End(builder) - - fbv_bb = TosaSerializer.serializeObjVec( - builder, self.basicBlocks, TosaGraph.StartBlocksVector - ) - - TosaGraph.Start(builder) - TosaGraph.AddVersion(builder, version) - TosaGraph.AddBlocks(builder, fbv_bb) - graph = TosaGraph.End(builder) - - self.builder.Finish(graph) - return self.builder.Output() - - def writeJson(self, tosa_filename): - """Write a json test file so that it is fairly easy to pick up the test - and generate commands for third party tool""" - test_desc = dict() - - test_desc["tosa_file"] = tosa_filename - ifm_name = [] - ifm_file = [] - ofm_name = [] - ofm_file = [] - - for b in self.basicBlocks: - if b.name == "main": - for i in b.inputs: - ifm_name.append(i) - ifm_file.append(b.tensors[i].placeholderFilename) - for o in b.outputs: - ofm_name.append(o) - # Make up an OFM filename here. One isn't generated until the reference tool is - # run, so any name is a good name - ofm_file.append("ref-{}.npy".format(o)) - - test_desc["ifm_name"] = ifm_name - test_desc["ifm_file"] = ifm_file - test_desc["ofm_name"] = ofm_name - test_desc["ofm_file"] = ofm_file - test_desc["expected_return_code"] = self.expectedReturnCode - test_desc["expected_failure"] = self.expectedFailure - if self.expectedFailureDesc: - test_desc["expected_failure_desc"] = self.expectedFailureDesc - - return json.dumps(test_desc, indent=" ") - - def startBasicBlock(self, name): - self.currBasicBlock = TosaSerializerBasicBlock(name) - self.basicBlocks.append(self.currBasicBlock) - - @staticmethod - def serializeStrVec(builder, vec, start_fcn): - fb_strs = [builder.CreateString(i) for i in vec] - start_fcn(builder, len(fb_strs)) - for s in fb_strs[::-1]: - builder.PrependUOffsetTRelative(s) - return builder.EndVector() - - @staticmethod - def serializeUint8Vec(builder, vec): - builder.StartVector(1, len(vec), 8) - for v in vec[::-1]: - builder.PrependUint8(v) - try: - return builder.EndVector() - except TypeError: - return builder.EndVector(len(vec)) - - @staticmethod - def serializeInt32Vec(builder, vec): - builder.StartVector(4, len(vec), 4) - for v in vec[::-1]: - builder.PrependInt32(v) - try: - return builder.EndVector() - except TypeError: - return builder.EndVector(len(vec)) - - @staticmethod - def serializeFpVec(builder, vec): - builder.StartVector(4, len(vec), 4) - for v in vec[::-1]: - builder.PrependFloat32(v) - try: - return builder.EndVector() - except TypeError: - return builder.EndVector(len(vec)) - - @staticmethod - def serializeObjVec(builder, vec, start_fcn): - serialized_vec = [] - for v in vec[::-1]: - serialized_vec.append(v.serialize(builder)) - - start_fcn(builder, len(vec)) - for v in serialized_vec: - builder.PrependUOffsetTRelative(v) - try: - return builder.EndVector() - except TypeError: - return builder.EndVector(len(vec)) - - @staticmethod - def toList(val): - if isinstance(val, list): - return val - else: - return [val] - -- cgit v1.2.1