1 //
2 // Copyright © 2017 Arm Ltd and Contributors. All rights reserved.
3 // SPDX-License-Identifier: MIT
4 //
5 #include "FakeQuantizationLayer.hpp"
6
7 #include "LayerCloneBase.hpp"
8
9 #include <armnn/TypesUtils.hpp>
10 #include <armnn/backends/WorkloadData.hpp>
11 #include <armnn/backends/WorkloadFactory.hpp>
12
13 namespace armnn
14 {
15
FakeQuantizationLayer(const FakeQuantizationDescriptor & param,const char * name)16 FakeQuantizationLayer::FakeQuantizationLayer(const FakeQuantizationDescriptor& param, const char* name)
17 : LayerWithParameters(1, 1, LayerType::FakeQuantization, param, name)
18 {
19 }
20
CreateWorkload(const IWorkloadFactory & factory) const21 std::unique_ptr<IWorkload> FakeQuantizationLayer::CreateWorkload(const IWorkloadFactory& factory) const
22 {
23 FakeQuantizationQueueDescriptor descriptor;
24 SetAdditionalInfo(descriptor);
25
26 return factory.CreateWorkload(LayerType::FakeQuantization, descriptor, PrepInfoAndDesc(descriptor) );
27 }
28
Clone(Graph & graph) const29 FakeQuantizationLayer* FakeQuantizationLayer::Clone(Graph& graph) const
30 {
31 return CloneBase<FakeQuantizationLayer>(graph, m_Param, GetName());
32 }
33
ValidateTensorShapesFromInputs()34 void FakeQuantizationLayer::ValidateTensorShapesFromInputs()
35 {
36 VerifyLayerConnections(1, CHECK_LOCATION());
37
38 const TensorShape& outputShape = GetOutputSlot(0).GetTensorInfo().GetShape();
39
40 VerifyShapeInferenceType(outputShape, m_ShapeInferenceMethod);
41
42 auto inferredShapes = InferOutputShapes({ GetInputSlot(0).GetConnection()->GetTensorInfo().GetShape() });
43
44 ARMNN_ASSERT(inferredShapes.size() == 1);
45
46 ValidateAndCopyShape(outputShape, inferredShapes[0], m_ShapeInferenceMethod, "FakeQuantizationLayer");
47 }
48
ExecuteStrategy(IStrategy & strategy) const49 void FakeQuantizationLayer::ExecuteStrategy(IStrategy& strategy) const
50 {
51 IgnoreUnused(strategy);
52 throw armnn::Exception("FakeQuantizationLayer should not appear in an input graph");
53 }
54
55 } // namespace armnn
56