diff options
author | Dmitrii Agibov <dmitrii.agibov@arm.com> | 2022-10-07 11:39:37 +0100 |
---|---|---|
committer | Dmitrii Agibov <dmitrii.agibov@arm.com> | 2022-10-07 11:40:21 +0100 |
commit | 3083f7ee68ce08147db08fca2474e5f4712fc8d7 (patch) | |
tree | c52e668c01a6a1041c08190e52a15944fd65b453 /src/mlia/resources | |
parent | bb7fb49484bb3687041061b2fdbbfae3959be54b (diff) | |
download | mlia-3083f7ee68ce08147db08fca2474e5f4712fc8d7.tar.gz |
MLIA-607 Update documentation and comments
Use "TensorFlow Lite" instead of "TFLite" in
documentation and comments
Change-Id: Ie4450d72fb2e5261d152d72ab8bd94c3da914c46
Diffstat (limited to 'src/mlia/resources')
9 files changed, 11 insertions, 11 deletions
diff --git a/src/mlia/resources/backend_configs/systems/corstone-300-vht/backend-config.json b/src/mlia/resources/backend_configs/systems/corstone-300-vht/backend-config.json index 5c44ebc..7bc12c7 100644 --- a/src/mlia/resources/backend_configs/systems/corstone-300-vht/backend-config.json +++ b/src/mlia/resources/backend_configs/systems/corstone-300-vht/backend-config.json @@ -16,7 +16,7 @@ "run": [ { "name": "--data", - "description": "Full file name for a custom model. Model must be in TFLite format compiled with Vela.", + "description": "Full file name for a custom model. Model must be in TensorFlow Lite format compiled with Vela.", "values": [], "alias": "input_file" }, @@ -52,7 +52,7 @@ "run": [ { "name": "--data", - "description": "Full file name for a custom model. Model must be in TFLite format compiled with Vela.", + "description": "Full file name for a custom model. Model must be in TensorFlow Lite format compiled with Vela.", "values": [], "alias": "input_file" }, diff --git a/src/mlia/resources/backend_configs/systems/corstone-300/backend-config.json b/src/mlia/resources/backend_configs/systems/corstone-300/backend-config.json index 41d2fd0..c27c6f5 100644 --- a/src/mlia/resources/backend_configs/systems/corstone-300/backend-config.json +++ b/src/mlia/resources/backend_configs/systems/corstone-300/backend-config.json @@ -16,7 +16,7 @@ "run": [ { "name": "--data", - "description": "Full file name for a custom model. Model must be in TFLite format compiled with Vela.", + "description": "Full file name for a custom model. Model must be in TensorFlow Lite format compiled with Vela.", "values": [], "alias": "input_file" }, @@ -52,7 +52,7 @@ "run": [ { "name": "--data", - "description": "Full file name for a custom model. Model must be in TFLite format compiled with Vela.", + "description": "Full file name for a custom model. Model must be in TensorFlow Lite format compiled with Vela.", "values": [], "alias": "input_file" }, diff --git a/src/mlia/resources/backend_configs/systems/corstone-310-vht/backend-config.json b/src/mlia/resources/backend_configs/systems/corstone-310-vht/backend-config.json index 3ea9a6a..bbcadfd 100644 --- a/src/mlia/resources/backend_configs/systems/corstone-310-vht/backend-config.json +++ b/src/mlia/resources/backend_configs/systems/corstone-310-vht/backend-config.json @@ -16,7 +16,7 @@ "run": [ { "name": "--data", - "description": "Full file name for a custom model. Model must be in TFLite format compiled with Vela.", + "description": "Full file name for a custom model. Model must be in TensorFlow Lite format compiled with Vela.", "values": [], "alias": "input_file" }, diff --git a/src/mlia/resources/backend_configs/systems/corstone-310/backend-config.json b/src/mlia/resources/backend_configs/systems/corstone-310/backend-config.json index d043a2d..e0b1d1d 100644 --- a/src/mlia/resources/backend_configs/systems/corstone-310/backend-config.json +++ b/src/mlia/resources/backend_configs/systems/corstone-310/backend-config.json @@ -16,7 +16,7 @@ "run": [ { "name": "--data", - "description": "Full file name for a custom model. Model must be in TFLite format compiled with Vela.", + "description": "Full file name for a custom model. Model must be in TensorFlow Lite format compiled with Vela.", "values": [], "alias": "input_file" }, diff --git a/src/mlia/resources/backends/applications/inference_runner-sse-300-22.05.01-ethos-U55-Shared_Sram-TA/backend-config.json b/src/mlia/resources/backends/applications/inference_runner-sse-300-22.05.01-ethos-U55-Shared_Sram-TA/backend-config.json index 7ee5e00..4856d27 100644 --- a/src/mlia/resources/backends/applications/inference_runner-sse-300-22.05.01-ethos-U55-Shared_Sram-TA/backend-config.json +++ b/src/mlia/resources/backends/applications/inference_runner-sse-300-22.05.01-ethos-U55-Shared_Sram-TA/backend-config.json @@ -1,7 +1,7 @@ [ { "name": "Generic Inference Runner: Ethos-U55/65 Shared SRAM", - "description": "This application allows running inferences using custom NN TFLite models on Ethos-U. No data pre-/post-processing is executed.", + "description": "This application allows running inferences using custom NN TensorFlow Lite models on Ethos-U. No data pre-/post-processing is executed.", "supported_systems": [ { "name": "Corstone-300: Cortex-M55+Ethos-U55" diff --git a/src/mlia/resources/backends/applications/inference_runner-sse-300-22.05.01-ethos-U55-Sram_Only-TA/backend-config.json b/src/mlia/resources/backends/applications/inference_runner-sse-300-22.05.01-ethos-U55-Sram_Only-TA/backend-config.json index 51ff429..3b512ff 100644 --- a/src/mlia/resources/backends/applications/inference_runner-sse-300-22.05.01-ethos-U55-Sram_Only-TA/backend-config.json +++ b/src/mlia/resources/backends/applications/inference_runner-sse-300-22.05.01-ethos-U55-Sram_Only-TA/backend-config.json @@ -1,7 +1,7 @@ [ { "name": "Generic Inference Runner: Ethos-U55 SRAM", - "description": "This application allows running inferences using custom NN TFLite models on Ethos-U. No data pre-/post-processing is executed.", + "description": "This application allows running inferences using custom NN TensorFlow Lite models on Ethos-U. No data pre-/post-processing is executed.", "supported_systems": [ { "name": "Corstone-300: Cortex-M55+Ethos-U55" diff --git a/src/mlia/resources/backends/applications/inference_runner-sse-300-22.05.01-ethos-U65-Dedicated_Sram-TA/backend-config.json b/src/mlia/resources/backends/applications/inference_runner-sse-300-22.05.01-ethos-U65-Dedicated_Sram-TA/backend-config.json index b59c85e..fc0569f 100644 --- a/src/mlia/resources/backends/applications/inference_runner-sse-300-22.05.01-ethos-U65-Dedicated_Sram-TA/backend-config.json +++ b/src/mlia/resources/backends/applications/inference_runner-sse-300-22.05.01-ethos-U65-Dedicated_Sram-TA/backend-config.json @@ -1,7 +1,7 @@ [ { "name": "Generic Inference Runner: Ethos-U65 Dedicated SRAM", - "description": "This application allows running inferences using custom NN TFLite models on Ethos-U. No data pre-/post-processing is executed.", + "description": "This application allows running inferences using custom NN TensorFlow Lite models on Ethos-U. No data pre-/post-processing is executed.", "supported_systems": [ { "name": "Corstone-300: Cortex-M55+Ethos-U65" diff --git a/src/mlia/resources/backends/applications/inference_runner-sse-310-22.05.01-ethos-U55-Shared_Sram-TA/backend-config.json b/src/mlia/resources/backends/applications/inference_runner-sse-310-22.05.01-ethos-U55-Shared_Sram-TA/backend-config.json index 69c5e60..0bdbd27 100644 --- a/src/mlia/resources/backends/applications/inference_runner-sse-310-22.05.01-ethos-U55-Shared_Sram-TA/backend-config.json +++ b/src/mlia/resources/backends/applications/inference_runner-sse-310-22.05.01-ethos-U55-Shared_Sram-TA/backend-config.json @@ -1,7 +1,7 @@ [ { "name": "Generic Inference Runner: Ethos-U55/65 Shared SRAM", - "description": "This application allows running inferences using custom NN TFLite models on Ethos-U. No data pre-/post-processing is executed.", + "description": "This application allows running inferences using custom NN TensorFlow Lite models on Ethos-U. No data pre-/post-processing is executed.", "supported_systems": [ { "name": "Corstone-310: Cortex-M85+Ethos-U55" diff --git a/src/mlia/resources/backends/applications/inference_runner-sse-310-22.05.01-ethos-U55-Sram_Only-TA/backend-config.json b/src/mlia/resources/backends/applications/inference_runner-sse-310-22.05.01-ethos-U55-Sram_Only-TA/backend-config.json index fbe4a16..a56e49d 100644 --- a/src/mlia/resources/backends/applications/inference_runner-sse-310-22.05.01-ethos-U55-Sram_Only-TA/backend-config.json +++ b/src/mlia/resources/backends/applications/inference_runner-sse-310-22.05.01-ethos-U55-Sram_Only-TA/backend-config.json @@ -1,7 +1,7 @@ [ { "name": "Generic Inference Runner: Ethos-U55 SRAM", - "description": "This application allows running inferences using custom NN TFLite models on Ethos-U. No data pre-/post-processing is executed.", + "description": "This application allows running inferences using custom NN TensorFlow Lite models on Ethos-U. No data pre-/post-processing is executed.", "supported_systems": [ { "name": "Corstone-310: Cortex-M85+Ethos-U55" |