aboutsummaryrefslogtreecommitdiff
path: root/examples/neon_copy_objects.cpp
diff options
context:
space:
mode:
authorMichalis Spyrou <michalis.spyrou@arm.com>2018-01-10 14:08:50 +0000
committerAnthony Barbier <anthony.barbier@arm.com>2018-11-02 16:43:10 +0000
commit2b5f0f2574551f59970bb9d710bafad2bc4bbd4a (patch)
treefd586f56b1285f0d6c52ecefc174eba0a9c8f157 /examples/neon_copy_objects.cpp
parent571b18a1fca4a5ed4dd24a38cb619f4de43ba3ed (diff)
downloadComputeLibrary-2b5f0f2574551f59970bb9d710bafad2bc4bbd4a.tar.gz
COMPMID-782 Port examples to the new format
Change-Id: Ib178a97c080ff650094d02ee49e2a0aa22376dd0 Reviewed-on: https://eu-gerrit-1.euhpc.arm.com/115717 Reviewed-by: Anthony Barbier <anthony.barbier@arm.com> Tested-by: Jenkins <bsgcomp@arm.com>
Diffstat (limited to 'examples/neon_copy_objects.cpp')
-rw-r--r--examples/neon_copy_objects.cpp208
1 files changed, 110 insertions, 98 deletions
diff --git a/examples/neon_copy_objects.cpp b/examples/neon_copy_objects.cpp
index 55d03dd8bf..9409cf366c 100644
--- a/examples/neon_copy_objects.cpp
+++ b/examples/neon_copy_objects.cpp
@@ -31,116 +31,128 @@
#include <iostream>
using namespace arm_compute;
+using namespace utils;
-void main_neon_copy_objects(int argc, char **argv)
+class NEONCopyObjectsExample : public Example
{
- ARM_COMPUTE_UNUSED(argc);
- ARM_COMPUTE_UNUSED(argv);
+public:
+ void do_setup(int argc, char **argv) override
+ {
+ ARM_COMPUTE_UNUSED(argc);
+ ARM_COMPUTE_UNUSED(argv);
- /** [Copy objects example] */
- constexpr unsigned int width = 4;
- constexpr unsigned int height = 3;
- constexpr unsigned int batch = 2;
+ /** [Copy objects example] */
+ constexpr unsigned int width = 4;
+ constexpr unsigned int height = 3;
+ constexpr unsigned int batch = 2;
- auto *src_data = new float[width * height * batch];
- auto *dst_data = new float[width * height * batch];
+ src_data = new float[width * height * batch];
+ dst_data = new float[width * height * batch];
- // Fill src_data with dummy values:
- for(unsigned int b = 0; b < batch; b++)
- {
- for(unsigned int h = 0; h < height; h++)
+ // Fill src_data with dummy values:
+ for(unsigned int b = 0; b < batch; b++)
{
- for(unsigned int w = 0; w < width; w++)
+ for(unsigned int h = 0; h < height; h++)
{
- src_data[b * (width * height) + h * width + w] = static_cast<float>(100 * b + 10 * h + w);
+ for(unsigned int w = 0; w < width; w++)
+ {
+ src_data[b * (width * height) + h * width + w] = static_cast<float>(100 * b + 10 * h + w);
+ }
}
}
- }
- Tensor input, output;
- NESoftmaxLayer softmax;
-
- // Initialize the tensors dimensions and type:
- const TensorShape shape(width, height, batch);
- input.allocator()->init(TensorInfo(shape, 1, DataType::F32));
- output.allocator()->init(TensorInfo(shape, 1, DataType::F32));
-
- // Configure softmax:
- softmax.configure(&input, &output);
-
- // Allocate the input / output tensors:
- input.allocator()->allocate();
- output.allocator()->allocate();
-
- // Fill the input tensor:
- // Simplest way: create an iterator to iterate through each element of the input tensor:
- Window input_window;
- input_window.use_tensor_dimensions(input.info()->tensor_shape());
- std::cout << " Dimensions of the input's iterator:\n";
- std::cout << " X = [start=" << input_window.x().start() << ", end=" << input_window.x().end() << ", step=" << input_window.x().step() << "]\n";
- std::cout << " Y = [start=" << input_window.y().start() << ", end=" << input_window.y().end() << ", step=" << input_window.y().step() << "]\n";
- std::cout << " Z = [start=" << input_window.z().start() << ", end=" << input_window.z().end() << ", step=" << input_window.z().step() << "]\n";
-
- // Create an iterator:
- Iterator input_it(&input, input_window);
-
- // Iterate through the elements of src_data and copy them one by one to the input tensor:
- // This is equivalent to:
- // for( unsigned int z = 0; z < batch; ++z)
- // {
- // for( unsigned int y = 0; y < height; ++y)
- // {
- // for( unsigned int x = 0; x < width; ++x)
- // {
- // *reinterpret_cast<float*>( input.buffer() + input.info()->offset_element_in_bytes(Coordinates(x,y,z))) = src_data[ z * (width*height) + y * width + x];
- // }
- // }
- // }
- // Except it works for an arbitrary number of dimensions
- execute_window_loop(input_window, [&](const Coordinates & id)
+ // Initialize the tensors dimensions and type:
+ const TensorShape shape(width, height, batch);
+ input.allocator()->init(TensorInfo(shape, 1, DataType::F32));
+ output.allocator()->init(TensorInfo(shape, 1, DataType::F32));
+
+ // Configure softmax:
+ softmax.configure(&input, &output);
+
+ // Allocate the input / output tensors:
+ input.allocator()->allocate();
+ output.allocator()->allocate();
+
+ // Fill the input tensor:
+ // Simplest way: create an iterator to iterate through each element of the input tensor:
+ Window input_window;
+ input_window.use_tensor_dimensions(input.info()->tensor_shape());
+ std::cout << " Dimensions of the input's iterator:\n";
+ std::cout << " X = [start=" << input_window.x().start() << ", end=" << input_window.x().end() << ", step=" << input_window.x().step() << "]\n";
+ std::cout << " Y = [start=" << input_window.y().start() << ", end=" << input_window.y().end() << ", step=" << input_window.y().step() << "]\n";
+ std::cout << " Z = [start=" << input_window.z().start() << ", end=" << input_window.z().end() << ", step=" << input_window.z().step() << "]\n";
+
+ // Create an iterator:
+ Iterator input_it(&input, input_window);
+
+ // Iterate through the elements of src_data and copy them one by one to the input tensor:
+ // This is equivalent to:
+ // for( unsigned int z = 0; z < batch; ++z)
+ // {
+ // for( unsigned int y = 0; y < height; ++y)
+ // {
+ // for( unsigned int x = 0; x < width; ++x)
+ // {
+ // *reinterpret_cast<float*>( input.buffer() + input.info()->offset_element_in_bytes(Coordinates(x,y,z))) = src_data[ z * (width*height) + y * width + x];
+ // }
+ // }
+ // }
+ // Except it works for an arbitrary number of dimensions
+ execute_window_loop(input_window, [&](const Coordinates & id)
+ {
+ std::cout << "Setting item [" << id.x() << "," << id.y() << "," << id.z() << "]\n";
+ *reinterpret_cast<float *>(input_it.ptr()) = src_data[id.z() * (width * height) + id.y() * width + id.x()];
+ },
+ input_it);
+
+ // More efficient way: create an iterator to iterate through each row (instead of each element) of the output tensor:
+ Window output_window;
+ output_window.use_tensor_dimensions(output.info()->tensor_shape(), /* first_dimension =*/Window::DimY); // Iterate through the rows (not each element)
+ std::cout << " Dimensions of the output's iterator:\n";
+ std::cout << " X = [start=" << output_window.x().start() << ", end=" << output_window.x().end() << ", step=" << output_window.x().step() << "]\n";
+ std::cout << " Y = [start=" << output_window.y().start() << ", end=" << output_window.y().end() << ", step=" << output_window.y().step() << "]\n";
+ std::cout << " Z = [start=" << output_window.z().start() << ", end=" << output_window.z().end() << ", step=" << output_window.z().step() << "]\n";
+
+ // Create an iterator:
+ Iterator output_it(&output, output_window);
+
+ // Iterate through the rows of the output tensor and copy them to dst_data:
+ // This is equivalent to:
+ // for( unsigned int z = 0; z < batch; ++z)
+ // {
+ // for( unsigned int y = 0; y < height; ++y)
+ // {
+ // memcpy( dst_data + z * (width*height) + y * width, input.buffer() + input.info()->offset_element_in_bytes(Coordinates(0,y,z)), width * sizeof(float));
+ // }
+ // }
+ // Except it works for an arbitrary number of dimensions
+ execute_window_loop(output_window, [&](const Coordinates & id)
+ {
+ std::cout << "Copying one row starting from [" << id.x() << "," << id.y() << "," << id.z() << "]\n";
+ // Copy one whole row:
+ memcpy(dst_data + id.z() * (width * height) + id.y() * width, output_it.ptr(), width * sizeof(float));
+ },
+ output_it);
+
+ /** [Copy objects example] */
+ }
+ void do_run() override
{
- std::cout << "Setting item [" << id.x() << "," << id.y() << "," << id.z() << "]\n";
- *reinterpret_cast<float *>(input_it.ptr()) = src_data[id.z() * (width * height) + id.y() * width + id.x()];
- },
- input_it);
-
- // Run NEON softmax:
- softmax.run();
-
- // More efficient way: create an iterator to iterate through each row (instead of each element) of the output tensor:
- Window output_window;
- output_window.use_tensor_dimensions(output.info()->tensor_shape(), /* first_dimension =*/Window::DimY); // Iterate through the rows (not each element)
- std::cout << " Dimensions of the output's iterator:\n";
- std::cout << " X = [start=" << output_window.x().start() << ", end=" << output_window.x().end() << ", step=" << output_window.x().step() << "]\n";
- std::cout << " Y = [start=" << output_window.y().start() << ", end=" << output_window.y().end() << ", step=" << output_window.y().step() << "]\n";
- std::cout << " Z = [start=" << output_window.z().start() << ", end=" << output_window.z().end() << ", step=" << output_window.z().step() << "]\n";
-
- // Create an iterator:
- Iterator output_it(&output, output_window);
-
- // Iterate through the rows of the output tensor and copy them to dst_data:
- // This is equivalent to:
- // for( unsigned int z = 0; z < batch; ++z)
- // {
- // for( unsigned int y = 0; y < height; ++y)
- // {
- // memcpy( dst_data + z * (width*height) + y * width, input.buffer() + input.info()->offset_element_in_bytes(Coordinates(0,y,z)), width * sizeof(float));
- // }
- // }
- // Except it works for an arbitrary number of dimensions
- execute_window_loop(output_window, [&](const Coordinates & id)
+ // Run NEON softmax:
+ softmax.run();
+ }
+ void do_teardown() override
{
- std::cout << "Copying one row starting from [" << id.x() << "," << id.y() << "," << id.z() << "]\n";
- // Copy one whole row:
- memcpy(dst_data + id.z() * (width * height) + id.y() * width, output_it.ptr(), width * sizeof(float));
- },
- output_it);
-
- delete[] src_data;
- delete[] dst_data;
- /** [Copy objects example] */
-}
+ delete[] src_data;
+ delete[] dst_data;
+ }
+private:
+ Tensor input{}, output{};
+ float *src_data{};
+ float *dst_data{};
+ NESoftmaxLayer softmax{};
+};
/** Main program for the copy objects test
*
* @param[in] argc Number of arguments
@@ -148,5 +160,5 @@ void main_neon_copy_objects(int argc, char **argv)
*/
int main(int argc, char **argv)
{
- return utils::run_example(argc, argv, main_neon_copy_objects);
+ return utils::run_example<NEONCopyObjectsExample>(argc, argv);
}