// // Copyright © 2017 Arm Ltd. All rights reserved. // SPDX-License-Identifier: MIT // #include "MemImportLayer.hpp" #include "LayerCloneBase.hpp" #include #include #include #include namespace armnn { MemImportLayer::MemImportLayer(const char* name) : Layer(1, 1, LayerType::MemImport, name) { } MemImportLayer* MemImportLayer::Clone(Graph& graph) const { return CloneBase(graph, GetName()); } std::unique_ptr MemImportLayer::CreateWorkload(const IWorkloadFactory& factory) const { IgnoreUnused(factory); MemImportQueueDescriptor descriptor; //This is different from other workloads. Does not get created by the workload factory. return std::make_unique(descriptor, PrepInfoAndDesc(descriptor)); } void MemImportLayer::ValidateTensorShapesFromInputs(ShapeInferenceMethod shapeInferenceMethod) { IgnoreUnused(shapeInferenceMethod); VerifyLayerConnections(1, CHECK_LOCATION()); auto inferredShapes = InferOutputShapes({ GetInputSlot(0).GetConnection()->GetTensorInfo().GetShape() }); ARMNN_ASSERT(inferredShapes.size() == 1); ConditionalThrowIfNotEqual( "MemImportLayer: TensorShape set on OutputSlot[0] does not match the inferred shape.", GetOutputSlot(0).GetTensorInfo().GetShape(), inferredShapes[0]); } void MemImportLayer::Accept(ILayerVisitor& visitor) const { IgnoreUnused(visitor); throw armnn::Exception("MemImportLayer should not appear in an input graph"); } } // namespace armnn