From 7519d50c993d60faa1ea09e56abfbf17cef23b49 Mon Sep 17 00:00:00 2001 From: James Peet Date: Mon, 19 Jul 2021 16:47:58 +0100 Subject: MLBEDSW-4892: Fix crash affecting biases without quantization. Remove quant_values attribute from Tensor class. It only needs a single values attribute, holding either quantized or unquantized values as appropriate. Change-Id: Ie96f80ac58061b6077e0f7048dc60209fdfbcafa Signed-off-by: James Peet --- ethosu/vela/high_level_command_to_npu_op.py | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) (limited to 'ethosu/vela/high_level_command_to_npu_op.py') diff --git a/ethosu/vela/high_level_command_to_npu_op.py b/ethosu/vela/high_level_command_to_npu_op.py index 80d0e476..9b76ec1f 100644 --- a/ethosu/vela/high_level_command_to_npu_op.py +++ b/ethosu/vela/high_level_command_to_npu_op.py @@ -437,8 +437,7 @@ def create_npu_elementwise_op(cmd: NpuStripe, arch: ArchitectureFeatures) -> Npu npu_op.ifm2.quantization = get_ifm_or_ifm2_quantization(ps, cmd.ifm2_tensor) if cmd.ifm2_tensor.shape == []: # scalar - assert cmd.ifm2_tensor.quant_values.size == 1 - npu_op.ifm2_scalar = cmd.ifm2_tensor.values.item(0) + npu_op.ifm2_scalar = cmd.ifm2_tensor.get_scalar() npu_op.ifm2.shape = NpuShape3D(height=0, width=0, depth=0) else: ifm2_blk = cmd.ifm2_box.get_block() -- cgit v1.2.1