From 78792223369fa34dacd0e69e189af035283da2ae Mon Sep 17 00:00:00 2001 From: Charles Xu Date: Wed, 13 May 2020 10:15:26 +0200 Subject: Add elementwise vector scalars support Write the constant scalars into flash. In case it's Dram or OffChipFlash, DMA the scalars from flash to sram. Signed-off-by: Charles Xu Change-Id: I42300a05dfe968d623b8aec8549644549e0f54b5 --- ethosu/vela/supported_operators.py | 7 ------- 1 file changed, 7 deletions(-) (limited to 'ethosu/vela/supported_operators.py') diff --git a/ethosu/vela/supported_operators.py b/ethosu/vela/supported_operators.py index e5271450..574b3a49 100644 --- a/ethosu/vela/supported_operators.py +++ b/ethosu/vela/supported_operators.py @@ -229,13 +229,6 @@ class SupportedOperators: if op.type in self.binary_elem_wise_main_ops: # if op type is unary, ifm2_tensor is None if len(ifm2_tensor.shape) > 2 and ifm2_tensor.shape[0] != 1: return False - - # check scalar size - if hasattr(ifm_tensor.values, "__len__") and len(ifm_tensor.values) > 1: - return False - if op.type in self.binary_elem_wise_main_ops: # same as above - if hasattr(ifm2_tensor.values, "__len__") and len(ifm2_tensor.values) > 1: - return False return True def check_memory_only_restrictions(self, op): -- cgit v1.2.1