xref: /aosp_15_r20/external/armnn/delegate/test/QuantizationTestHelper.hpp (revision 89c4ff92f2867872bb9e2354d150bf0c8c502810)
1*89c4ff92SAndroid Build Coastguard Worker //
2*89c4ff92SAndroid Build Coastguard Worker // Copyright © 2020, 2023 Arm Ltd and Contributors. All rights reserved.
3*89c4ff92SAndroid Build Coastguard Worker // SPDX-License-Identifier: MIT
4*89c4ff92SAndroid Build Coastguard Worker //
5*89c4ff92SAndroid Build Coastguard Worker 
6*89c4ff92SAndroid Build Coastguard Worker #pragma once
7*89c4ff92SAndroid Build Coastguard Worker 
8*89c4ff92SAndroid Build Coastguard Worker #include "TestUtils.hpp"
9*89c4ff92SAndroid Build Coastguard Worker 
10*89c4ff92SAndroid Build Coastguard Worker #include <armnn_delegate.hpp>
11*89c4ff92SAndroid Build Coastguard Worker #include <DelegateTestInterpreter.hpp>
12*89c4ff92SAndroid Build Coastguard Worker 
13*89c4ff92SAndroid Build Coastguard Worker #include <flatbuffers/flatbuffers.h>
14*89c4ff92SAndroid Build Coastguard Worker #include <tensorflow/lite/kernels/register.h>
15*89c4ff92SAndroid Build Coastguard Worker #include <tensorflow/lite/version.h>
16*89c4ff92SAndroid Build Coastguard Worker 
17*89c4ff92SAndroid Build Coastguard Worker #include <schema_generated.h>
18*89c4ff92SAndroid Build Coastguard Worker 
19*89c4ff92SAndroid Build Coastguard Worker #include <doctest/doctest.h>
20*89c4ff92SAndroid Build Coastguard Worker 
21*89c4ff92SAndroid Build Coastguard Worker namespace
22*89c4ff92SAndroid Build Coastguard Worker {
23*89c4ff92SAndroid Build Coastguard Worker 
CreateQuantizationTfLiteModel(tflite::BuiltinOperator quantizationOperatorCode,tflite::TensorType inputTensorType,tflite::TensorType outputTensorType,const std::vector<int32_t> & inputTensorShape,const std::vector<int32_t> & outputTensorShape,float quantScale=1.0f,int quantOffset=0)24*89c4ff92SAndroid Build Coastguard Worker std::vector<char> CreateQuantizationTfLiteModel(tflite::BuiltinOperator quantizationOperatorCode,
25*89c4ff92SAndroid Build Coastguard Worker                                                 tflite::TensorType inputTensorType,
26*89c4ff92SAndroid Build Coastguard Worker                                                 tflite::TensorType outputTensorType,
27*89c4ff92SAndroid Build Coastguard Worker                                                 const std::vector <int32_t>& inputTensorShape,
28*89c4ff92SAndroid Build Coastguard Worker                                                 const std::vector <int32_t>& outputTensorShape,
29*89c4ff92SAndroid Build Coastguard Worker                                                 float quantScale = 1.0f,
30*89c4ff92SAndroid Build Coastguard Worker                                                 int quantOffset  = 0)
31*89c4ff92SAndroid Build Coastguard Worker {
32*89c4ff92SAndroid Build Coastguard Worker     using namespace tflite;
33*89c4ff92SAndroid Build Coastguard Worker     flatbuffers::FlatBufferBuilder flatBufferBuilder;
34*89c4ff92SAndroid Build Coastguard Worker 
35*89c4ff92SAndroid Build Coastguard Worker     std::vector<flatbuffers::Offset<tflite::Buffer>> buffers;
36*89c4ff92SAndroid Build Coastguard Worker     buffers.push_back(CreateBuffer(flatBufferBuilder));
37*89c4ff92SAndroid Build Coastguard Worker     buffers.push_back(CreateBuffer(flatBufferBuilder));
38*89c4ff92SAndroid Build Coastguard Worker     buffers.push_back(CreateBuffer(flatBufferBuilder));
39*89c4ff92SAndroid Build Coastguard Worker 
40*89c4ff92SAndroid Build Coastguard Worker 
41*89c4ff92SAndroid Build Coastguard Worker     auto quantizationParameters =
42*89c4ff92SAndroid Build Coastguard Worker             CreateQuantizationParameters(flatBufferBuilder,
43*89c4ff92SAndroid Build Coastguard Worker                                          0,
44*89c4ff92SAndroid Build Coastguard Worker                                          0,
45*89c4ff92SAndroid Build Coastguard Worker                                          flatBufferBuilder.CreateVector<float>({ quantScale }),
46*89c4ff92SAndroid Build Coastguard Worker                                          flatBufferBuilder.CreateVector<int64_t>({ quantOffset }),
47*89c4ff92SAndroid Build Coastguard Worker                                          QuantizationDetails_CustomQuantization);
48*89c4ff92SAndroid Build Coastguard Worker 
49*89c4ff92SAndroid Build Coastguard Worker     std::array<flatbuffers::Offset<Tensor>, 2> tensors;
50*89c4ff92SAndroid Build Coastguard Worker     tensors[0] = CreateTensor(flatBufferBuilder,
51*89c4ff92SAndroid Build Coastguard Worker                               flatBufferBuilder.CreateVector<int32_t>(inputTensorShape.data(),
52*89c4ff92SAndroid Build Coastguard Worker                                                                       inputTensorShape.size()),
53*89c4ff92SAndroid Build Coastguard Worker                               inputTensorType,
54*89c4ff92SAndroid Build Coastguard Worker                               1,
55*89c4ff92SAndroid Build Coastguard Worker                               flatBufferBuilder.CreateString("input"),
56*89c4ff92SAndroid Build Coastguard Worker                               quantizationParameters);
57*89c4ff92SAndroid Build Coastguard Worker     tensors[1] = CreateTensor(flatBufferBuilder,
58*89c4ff92SAndroid Build Coastguard Worker                               flatBufferBuilder.CreateVector<int32_t>(outputTensorShape.data(),
59*89c4ff92SAndroid Build Coastguard Worker                                                                       outputTensorShape.size()),
60*89c4ff92SAndroid Build Coastguard Worker                               outputTensorType,
61*89c4ff92SAndroid Build Coastguard Worker                               2,
62*89c4ff92SAndroid Build Coastguard Worker                               flatBufferBuilder.CreateString("output"),
63*89c4ff92SAndroid Build Coastguard Worker                               quantizationParameters);
64*89c4ff92SAndroid Build Coastguard Worker 
65*89c4ff92SAndroid Build Coastguard Worker     // create operator
66*89c4ff92SAndroid Build Coastguard Worker     tflite::BuiltinOptions operatorBuiltinOptionsType = tflite::BuiltinOptions_NONE;
67*89c4ff92SAndroid Build Coastguard Worker     flatbuffers::Offset<void> operatorBuiltinOptions = 0;
68*89c4ff92SAndroid Build Coastguard Worker     switch (quantizationOperatorCode)
69*89c4ff92SAndroid Build Coastguard Worker     {
70*89c4ff92SAndroid Build Coastguard Worker         case BuiltinOperator_QUANTIZE:
71*89c4ff92SAndroid Build Coastguard Worker         {
72*89c4ff92SAndroid Build Coastguard Worker             operatorBuiltinOptionsType = BuiltinOptions_QuantizeOptions;
73*89c4ff92SAndroid Build Coastguard Worker             operatorBuiltinOptions = CreateQuantizeOptions(flatBufferBuilder).Union();
74*89c4ff92SAndroid Build Coastguard Worker             break;
75*89c4ff92SAndroid Build Coastguard Worker         }
76*89c4ff92SAndroid Build Coastguard Worker         case BuiltinOperator_DEQUANTIZE:
77*89c4ff92SAndroid Build Coastguard Worker         {
78*89c4ff92SAndroid Build Coastguard Worker             operatorBuiltinOptionsType = BuiltinOptions_DequantizeOptions;
79*89c4ff92SAndroid Build Coastguard Worker             operatorBuiltinOptions = CreateDequantizeOptions(flatBufferBuilder).Union();
80*89c4ff92SAndroid Build Coastguard Worker             break;
81*89c4ff92SAndroid Build Coastguard Worker         }
82*89c4ff92SAndroid Build Coastguard Worker         default:
83*89c4ff92SAndroid Build Coastguard Worker             break;
84*89c4ff92SAndroid Build Coastguard Worker     }
85*89c4ff92SAndroid Build Coastguard Worker 
86*89c4ff92SAndroid Build Coastguard Worker     const std::vector<int32_t> operatorInputs{0};
87*89c4ff92SAndroid Build Coastguard Worker     const std::vector<int32_t> operatorOutputs{1};
88*89c4ff92SAndroid Build Coastguard Worker     flatbuffers::Offset <Operator> quantizationOperator =
89*89c4ff92SAndroid Build Coastguard Worker             CreateOperator(flatBufferBuilder,
90*89c4ff92SAndroid Build Coastguard Worker                            0,
91*89c4ff92SAndroid Build Coastguard Worker                            flatBufferBuilder.CreateVector<int32_t>(operatorInputs.data(), operatorInputs.size()),
92*89c4ff92SAndroid Build Coastguard Worker                            flatBufferBuilder.CreateVector<int32_t>(operatorOutputs.data(), operatorOutputs.size()),
93*89c4ff92SAndroid Build Coastguard Worker                            operatorBuiltinOptionsType,
94*89c4ff92SAndroid Build Coastguard Worker                            operatorBuiltinOptions);
95*89c4ff92SAndroid Build Coastguard Worker 
96*89c4ff92SAndroid Build Coastguard Worker     const std::vector<int> subgraphInputs{0};
97*89c4ff92SAndroid Build Coastguard Worker     const std::vector<int> subgraphOutputs{1};
98*89c4ff92SAndroid Build Coastguard Worker     flatbuffers::Offset <SubGraph> subgraph =
99*89c4ff92SAndroid Build Coastguard Worker             CreateSubGraph(flatBufferBuilder,
100*89c4ff92SAndroid Build Coastguard Worker                            flatBufferBuilder.CreateVector(tensors.data(), tensors.size()),
101*89c4ff92SAndroid Build Coastguard Worker                            flatBufferBuilder.CreateVector<int32_t>(subgraphInputs.data(), subgraphInputs.size()),
102*89c4ff92SAndroid Build Coastguard Worker                            flatBufferBuilder.CreateVector<int32_t>(subgraphOutputs.data(), subgraphOutputs.size()),
103*89c4ff92SAndroid Build Coastguard Worker                            flatBufferBuilder.CreateVector(&quantizationOperator, 1));
104*89c4ff92SAndroid Build Coastguard Worker 
105*89c4ff92SAndroid Build Coastguard Worker     flatbuffers::Offset <flatbuffers::String> modelDescription =
106*89c4ff92SAndroid Build Coastguard Worker             flatBufferBuilder.CreateString("ArmnnDelegate: Quantization Operator Model");
107*89c4ff92SAndroid Build Coastguard Worker     flatbuffers::Offset <OperatorCode> operatorCode = CreateOperatorCode(flatBufferBuilder, quantizationOperatorCode);
108*89c4ff92SAndroid Build Coastguard Worker 
109*89c4ff92SAndroid Build Coastguard Worker     flatbuffers::Offset <Model> flatbufferModel =
110*89c4ff92SAndroid Build Coastguard Worker             CreateModel(flatBufferBuilder,
111*89c4ff92SAndroid Build Coastguard Worker                         TFLITE_SCHEMA_VERSION,
112*89c4ff92SAndroid Build Coastguard Worker                         flatBufferBuilder.CreateVector(&operatorCode, 1),
113*89c4ff92SAndroid Build Coastguard Worker                         flatBufferBuilder.CreateVector(&subgraph, 1),
114*89c4ff92SAndroid Build Coastguard Worker                         modelDescription,
115*89c4ff92SAndroid Build Coastguard Worker                         flatBufferBuilder.CreateVector(buffers.data(), buffers.size()));
116*89c4ff92SAndroid Build Coastguard Worker 
117*89c4ff92SAndroid Build Coastguard Worker     flatBufferBuilder.Finish(flatbufferModel, armnnDelegate::FILE_IDENTIFIER);
118*89c4ff92SAndroid Build Coastguard Worker 
119*89c4ff92SAndroid Build Coastguard Worker     return std::vector<char>(flatBufferBuilder.GetBufferPointer(),
120*89c4ff92SAndroid Build Coastguard Worker                              flatBufferBuilder.GetBufferPointer() + flatBufferBuilder.GetSize());
121*89c4ff92SAndroid Build Coastguard Worker }
122*89c4ff92SAndroid Build Coastguard Worker 
123*89c4ff92SAndroid Build Coastguard Worker template <typename InputT, typename OutputT>
QuantizationTest(tflite::BuiltinOperator quantizeOperatorCode,tflite::TensorType inputTensorType,tflite::TensorType outputTensorType,std::vector<armnn::BackendId> & backends,std::vector<int32_t> & inputShape,std::vector<int32_t> & outputShape,std::vector<InputT> & inputValues,std::vector<OutputT> & expectedOutputValues,float quantScale=1.0f,int quantOffset=0)124*89c4ff92SAndroid Build Coastguard Worker void QuantizationTest(tflite::BuiltinOperator quantizeOperatorCode,
125*89c4ff92SAndroid Build Coastguard Worker                       tflite::TensorType inputTensorType,
126*89c4ff92SAndroid Build Coastguard Worker                       tflite::TensorType outputTensorType,
127*89c4ff92SAndroid Build Coastguard Worker                       std::vector<armnn::BackendId>& backends,
128*89c4ff92SAndroid Build Coastguard Worker                       std::vector<int32_t>& inputShape,
129*89c4ff92SAndroid Build Coastguard Worker                       std::vector<int32_t>& outputShape,
130*89c4ff92SAndroid Build Coastguard Worker                       std::vector<InputT>&  inputValues,
131*89c4ff92SAndroid Build Coastguard Worker                       std::vector<OutputT>& expectedOutputValues,
132*89c4ff92SAndroid Build Coastguard Worker                       float quantScale = 1.0f,
133*89c4ff92SAndroid Build Coastguard Worker                       int quantOffset  = 0)
134*89c4ff92SAndroid Build Coastguard Worker {
135*89c4ff92SAndroid Build Coastguard Worker     using namespace delegateTestInterpreter;
136*89c4ff92SAndroid Build Coastguard Worker     std::vector<char> modelBuffer = CreateQuantizationTfLiteModel(quantizeOperatorCode,
137*89c4ff92SAndroid Build Coastguard Worker                                                                   inputTensorType,
138*89c4ff92SAndroid Build Coastguard Worker                                                                   outputTensorType,
139*89c4ff92SAndroid Build Coastguard Worker                                                                   inputShape,
140*89c4ff92SAndroid Build Coastguard Worker                                                                   outputShape,
141*89c4ff92SAndroid Build Coastguard Worker                                                                   quantScale,
142*89c4ff92SAndroid Build Coastguard Worker                                                                   quantOffset);
143*89c4ff92SAndroid Build Coastguard Worker 
144*89c4ff92SAndroid Build Coastguard Worker     // Setup interpreter with just TFLite Runtime.
145*89c4ff92SAndroid Build Coastguard Worker     auto tfLiteInterpreter = DelegateTestInterpreter(modelBuffer);
146*89c4ff92SAndroid Build Coastguard Worker     CHECK(tfLiteInterpreter.AllocateTensors() == kTfLiteOk);
147*89c4ff92SAndroid Build Coastguard Worker     CHECK(tfLiteInterpreter.FillInputTensor(inputValues, 0) == kTfLiteOk);
148*89c4ff92SAndroid Build Coastguard Worker     CHECK(tfLiteInterpreter.Invoke() == kTfLiteOk);
149*89c4ff92SAndroid Build Coastguard Worker     std::vector<OutputT> tfLiteOutputValues = tfLiteInterpreter.GetOutputResult<OutputT>(0);
150*89c4ff92SAndroid Build Coastguard Worker     std::vector<int32_t> tfLiteOutputShape  = tfLiteInterpreter.GetOutputShape(0);
151*89c4ff92SAndroid Build Coastguard Worker 
152*89c4ff92SAndroid Build Coastguard Worker     // Setup interpreter with Arm NN Delegate applied.
153*89c4ff92SAndroid Build Coastguard Worker     auto armnnInterpreter = DelegateTestInterpreter(modelBuffer, backends);
154*89c4ff92SAndroid Build Coastguard Worker     CHECK(armnnInterpreter.AllocateTensors() == kTfLiteOk);
155*89c4ff92SAndroid Build Coastguard Worker     CHECK(armnnInterpreter.FillInputTensor(inputValues, 0) == kTfLiteOk);
156*89c4ff92SAndroid Build Coastguard Worker     CHECK(armnnInterpreter.Invoke() == kTfLiteOk);
157*89c4ff92SAndroid Build Coastguard Worker     std::vector<OutputT> armnnOutputValues = armnnInterpreter.GetOutputResult<OutputT>(0);
158*89c4ff92SAndroid Build Coastguard Worker     std::vector<int32_t> armnnOutputShape  = armnnInterpreter.GetOutputShape(0);
159*89c4ff92SAndroid Build Coastguard Worker 
160*89c4ff92SAndroid Build Coastguard Worker     armnnDelegate::CompareOutputData<OutputT>(tfLiteOutputValues, armnnOutputValues, expectedOutputValues);
161*89c4ff92SAndroid Build Coastguard Worker     armnnDelegate::CompareOutputShape(tfLiteOutputShape, armnnOutputShape, outputShape);
162*89c4ff92SAndroid Build Coastguard Worker 
163*89c4ff92SAndroid Build Coastguard Worker     tfLiteInterpreter.Cleanup();
164*89c4ff92SAndroid Build Coastguard Worker     armnnInterpreter.Cleanup();
165*89c4ff92SAndroid Build Coastguard Worker }
166*89c4ff92SAndroid Build Coastguard Worker 
167*89c4ff92SAndroid Build Coastguard Worker } // anonymous namespace