aboutsummaryrefslogtreecommitdiff
path: root/ethosu/vela/weight_compressor.py
diff options
context:
space:
mode:
authorManupa Karunaratne <manupa.karunaratne@arm.com>2020-07-29 18:06:28 +0100
committerFredrik Knutsson <fredrik.knutsson.hunnebo@gmail.com>2020-08-06 06:48:19 +0000
commitbef228bb19ff01bdab8c0b7570ddc4e4f7cdeeb6 (patch)
tree01fc8d7f543bd694b30b49f468fb0262622c8a5b /ethosu/vela/weight_compressor.py
parenta0c3624899edc601525a589643c802469003f89d (diff)
downloadethos-u-vela-bef228bb19ff01bdab8c0b7570ddc4e4f7cdeeb6.tar.gz
[EXTAPI] exposing encode of biases to be consumed by an external API
*Renamed pack_bias_and_scale to encode_bias to be consumed externally *added unit test for the API Change-Id: I71829f3fcb390c475795848f0be3d132d3e158ee Signed-off-by: Manupa Karunaratne <manupa.karunaratne@arm.com>
Diffstat (limited to 'ethosu/vela/weight_compressor.py')
-rw-r--r--ethosu/vela/weight_compressor.py50
1 files changed, 28 insertions, 22 deletions
diff --git a/ethosu/vela/weight_compressor.py b/ethosu/vela/weight_compressor.py
index 687a0805..fa7d2d11 100644
--- a/ethosu/vela/weight_compressor.py
+++ b/ethosu/vela/weight_compressor.py
@@ -94,6 +94,33 @@ def encode_weights(
return encoded_stream
+@typecheck
+def encode_bias(bias: np.int64, scale: int, shift: int):
+ """
+ Public facing API to pack bias and scale values as required by the hardware
+ :param bias: 64bit signed number that includes 40bit signed bias
+ :param scale: 32bit scale value
+ :param shift: 6bit shift value
+ :return: packed 80bit [0(2-bits),shift(6-bits),scale(32-bits),bias(40-bits)]
+ """
+ assert -(1 << (40 - 1)) <= bias < (1 << (40 - 1)) # signed 40-bit range
+ assert 0 <= scale < (1 << 32) # unsigned 32-bit range
+ assert 0 <= shift < (1 << 6) # unsigned 6-bit range
+
+ data = bytearray(10)
+ data[0] = (bias >> (0 * 8)) & 0xFF
+ data[1] = (bias >> (1 * 8)) & 0xFF
+ data[2] = (bias >> (2 * 8)) & 0xFF
+ data[3] = (bias >> (3 * 8)) & 0xFF
+ data[4] = (bias >> (4 * 8)) & 0xFF
+ data[5] = (scale >> (0 * 8)) & 0xFF
+ data[6] = (scale >> (1 * 8)) & 0xFF
+ data[7] = (scale >> (2 * 8)) & 0xFF
+ data[8] = (scale >> (3 * 8)) & 0xFF
+ data[9] = shift & 0x3F
+ return data
+
+
def create_weight_compression_config(tens, npu_block_type, ofm_block_depth, ofm_depth_step, dilation):
# Note: for an ofm block only its depth is used in weight compression.
# And block depth > ofm depth gives same result as block depth == ofm depth
@@ -376,27 +403,6 @@ def calc_scales_and_pack_biases(tens, arch, ofm_depth_step, rescale_for_faf=Fals
# the operator should only have a single output
assert len(tens.consumer_list[0].outputs) == 1
-
- def pack_bias_and_scale(bias, scale, shift):
- bias = np.int64(bias)
- assert -(1 << (40 - 1)) <= bias < (1 << (40 - 1)) # signed 40-bit range
- assert 0 <= scale < (1 << 32) # unsigned 32-bit range
- assert 0 <= shift < (1 << 6) # unsigned 6-bit range
-
- # pack the 80 bit value = [0(2-bits),shift(6-bits),scale(32-bits),bias(40-bits)]
- data = bytearray(10)
- data[0] = (bias >> (0 * 8)) & 0xFF
- data[1] = (bias >> (1 * 8)) & 0xFF
- data[2] = (bias >> (2 * 8)) & 0xFF
- data[3] = (bias >> (3 * 8)) & 0xFF
- data[4] = (bias >> (4 * 8)) & 0xFF
- data[5] = (scale >> (0 * 8)) & 0xFF
- data[6] = (scale >> (1 * 8)) & 0xFF
- data[7] = (scale >> (2 * 8)) & 0xFF
- data[8] = (scale >> (3 * 8)) & 0xFF
- data[9] = shift & 0x3F
- return data
-
biases = tens.quant_values
first_consumer_op = tens.consumer_list[0]
@@ -470,7 +476,7 @@ def calc_scales_and_pack_biases(tens, arch, ofm_depth_step, rescale_for_faf=Fals
core_scales = quantised_scales[i + core : i + core + max_len : arch.ncores]
core_biases = biases[i + core : i + core + max_len : arch.ncores]
for j, core_bias in enumerate(core_biases):
- stream.extend(pack_bias_and_scale(core_bias, *core_scales[j]))
+ stream.extend(encode_bias(np.int64(core_bias), *core_scales[j]))
# Align to 16 for start for next substream
remainder = (len(stream)) % 16