1*89c4ff92SAndroid Build Coastguard Worker //
2*89c4ff92SAndroid Build Coastguard Worker // Copyright © 2020, 2023 Arm Ltd and Contributors. All rights reserved.
3*89c4ff92SAndroid Build Coastguard Worker // SPDX-License-Identifier: MIT
4*89c4ff92SAndroid Build Coastguard Worker //
5*89c4ff92SAndroid Build Coastguard Worker
6*89c4ff92SAndroid Build Coastguard Worker #pragma once
7*89c4ff92SAndroid Build Coastguard Worker
8*89c4ff92SAndroid Build Coastguard Worker #include "TestUtils.hpp"
9*89c4ff92SAndroid Build Coastguard Worker
10*89c4ff92SAndroid Build Coastguard Worker #include <armnn_delegate.hpp>
11*89c4ff92SAndroid Build Coastguard Worker #include <DelegateTestInterpreter.hpp>
12*89c4ff92SAndroid Build Coastguard Worker
13*89c4ff92SAndroid Build Coastguard Worker #include <flatbuffers/flatbuffers.h>
14*89c4ff92SAndroid Build Coastguard Worker #include <tensorflow/lite/kernels/register.h>
15*89c4ff92SAndroid Build Coastguard Worker #include <tensorflow/lite/version.h>
16*89c4ff92SAndroid Build Coastguard Worker
17*89c4ff92SAndroid Build Coastguard Worker #include <schema_generated.h>
18*89c4ff92SAndroid Build Coastguard Worker
19*89c4ff92SAndroid Build Coastguard Worker #include <doctest/doctest.h>
20*89c4ff92SAndroid Build Coastguard Worker
21*89c4ff92SAndroid Build Coastguard Worker namespace
22*89c4ff92SAndroid Build Coastguard Worker {
23*89c4ff92SAndroid Build Coastguard Worker
CreateQuantizationTfLiteModel(tflite::BuiltinOperator quantizationOperatorCode,tflite::TensorType inputTensorType,tflite::TensorType outputTensorType,const std::vector<int32_t> & inputTensorShape,const std::vector<int32_t> & outputTensorShape,float quantScale=1.0f,int quantOffset=0)24*89c4ff92SAndroid Build Coastguard Worker std::vector<char> CreateQuantizationTfLiteModel(tflite::BuiltinOperator quantizationOperatorCode,
25*89c4ff92SAndroid Build Coastguard Worker tflite::TensorType inputTensorType,
26*89c4ff92SAndroid Build Coastguard Worker tflite::TensorType outputTensorType,
27*89c4ff92SAndroid Build Coastguard Worker const std::vector <int32_t>& inputTensorShape,
28*89c4ff92SAndroid Build Coastguard Worker const std::vector <int32_t>& outputTensorShape,
29*89c4ff92SAndroid Build Coastguard Worker float quantScale = 1.0f,
30*89c4ff92SAndroid Build Coastguard Worker int quantOffset = 0)
31*89c4ff92SAndroid Build Coastguard Worker {
32*89c4ff92SAndroid Build Coastguard Worker using namespace tflite;
33*89c4ff92SAndroid Build Coastguard Worker flatbuffers::FlatBufferBuilder flatBufferBuilder;
34*89c4ff92SAndroid Build Coastguard Worker
35*89c4ff92SAndroid Build Coastguard Worker std::vector<flatbuffers::Offset<tflite::Buffer>> buffers;
36*89c4ff92SAndroid Build Coastguard Worker buffers.push_back(CreateBuffer(flatBufferBuilder));
37*89c4ff92SAndroid Build Coastguard Worker buffers.push_back(CreateBuffer(flatBufferBuilder));
38*89c4ff92SAndroid Build Coastguard Worker buffers.push_back(CreateBuffer(flatBufferBuilder));
39*89c4ff92SAndroid Build Coastguard Worker
40*89c4ff92SAndroid Build Coastguard Worker
41*89c4ff92SAndroid Build Coastguard Worker auto quantizationParameters =
42*89c4ff92SAndroid Build Coastguard Worker CreateQuantizationParameters(flatBufferBuilder,
43*89c4ff92SAndroid Build Coastguard Worker 0,
44*89c4ff92SAndroid Build Coastguard Worker 0,
45*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<float>({ quantScale }),
46*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<int64_t>({ quantOffset }),
47*89c4ff92SAndroid Build Coastguard Worker QuantizationDetails_CustomQuantization);
48*89c4ff92SAndroid Build Coastguard Worker
49*89c4ff92SAndroid Build Coastguard Worker std::array<flatbuffers::Offset<Tensor>, 2> tensors;
50*89c4ff92SAndroid Build Coastguard Worker tensors[0] = CreateTensor(flatBufferBuilder,
51*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<int32_t>(inputTensorShape.data(),
52*89c4ff92SAndroid Build Coastguard Worker inputTensorShape.size()),
53*89c4ff92SAndroid Build Coastguard Worker inputTensorType,
54*89c4ff92SAndroid Build Coastguard Worker 1,
55*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateString("input"),
56*89c4ff92SAndroid Build Coastguard Worker quantizationParameters);
57*89c4ff92SAndroid Build Coastguard Worker tensors[1] = CreateTensor(flatBufferBuilder,
58*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<int32_t>(outputTensorShape.data(),
59*89c4ff92SAndroid Build Coastguard Worker outputTensorShape.size()),
60*89c4ff92SAndroid Build Coastguard Worker outputTensorType,
61*89c4ff92SAndroid Build Coastguard Worker 2,
62*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateString("output"),
63*89c4ff92SAndroid Build Coastguard Worker quantizationParameters);
64*89c4ff92SAndroid Build Coastguard Worker
65*89c4ff92SAndroid Build Coastguard Worker // create operator
66*89c4ff92SAndroid Build Coastguard Worker tflite::BuiltinOptions operatorBuiltinOptionsType = tflite::BuiltinOptions_NONE;
67*89c4ff92SAndroid Build Coastguard Worker flatbuffers::Offset<void> operatorBuiltinOptions = 0;
68*89c4ff92SAndroid Build Coastguard Worker switch (quantizationOperatorCode)
69*89c4ff92SAndroid Build Coastguard Worker {
70*89c4ff92SAndroid Build Coastguard Worker case BuiltinOperator_QUANTIZE:
71*89c4ff92SAndroid Build Coastguard Worker {
72*89c4ff92SAndroid Build Coastguard Worker operatorBuiltinOptionsType = BuiltinOptions_QuantizeOptions;
73*89c4ff92SAndroid Build Coastguard Worker operatorBuiltinOptions = CreateQuantizeOptions(flatBufferBuilder).Union();
74*89c4ff92SAndroid Build Coastguard Worker break;
75*89c4ff92SAndroid Build Coastguard Worker }
76*89c4ff92SAndroid Build Coastguard Worker case BuiltinOperator_DEQUANTIZE:
77*89c4ff92SAndroid Build Coastguard Worker {
78*89c4ff92SAndroid Build Coastguard Worker operatorBuiltinOptionsType = BuiltinOptions_DequantizeOptions;
79*89c4ff92SAndroid Build Coastguard Worker operatorBuiltinOptions = CreateDequantizeOptions(flatBufferBuilder).Union();
80*89c4ff92SAndroid Build Coastguard Worker break;
81*89c4ff92SAndroid Build Coastguard Worker }
82*89c4ff92SAndroid Build Coastguard Worker default:
83*89c4ff92SAndroid Build Coastguard Worker break;
84*89c4ff92SAndroid Build Coastguard Worker }
85*89c4ff92SAndroid Build Coastguard Worker
86*89c4ff92SAndroid Build Coastguard Worker const std::vector<int32_t> operatorInputs{0};
87*89c4ff92SAndroid Build Coastguard Worker const std::vector<int32_t> operatorOutputs{1};
88*89c4ff92SAndroid Build Coastguard Worker flatbuffers::Offset <Operator> quantizationOperator =
89*89c4ff92SAndroid Build Coastguard Worker CreateOperator(flatBufferBuilder,
90*89c4ff92SAndroid Build Coastguard Worker 0,
91*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<int32_t>(operatorInputs.data(), operatorInputs.size()),
92*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<int32_t>(operatorOutputs.data(), operatorOutputs.size()),
93*89c4ff92SAndroid Build Coastguard Worker operatorBuiltinOptionsType,
94*89c4ff92SAndroid Build Coastguard Worker operatorBuiltinOptions);
95*89c4ff92SAndroid Build Coastguard Worker
96*89c4ff92SAndroid Build Coastguard Worker const std::vector<int> subgraphInputs{0};
97*89c4ff92SAndroid Build Coastguard Worker const std::vector<int> subgraphOutputs{1};
98*89c4ff92SAndroid Build Coastguard Worker flatbuffers::Offset <SubGraph> subgraph =
99*89c4ff92SAndroid Build Coastguard Worker CreateSubGraph(flatBufferBuilder,
100*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector(tensors.data(), tensors.size()),
101*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<int32_t>(subgraphInputs.data(), subgraphInputs.size()),
102*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<int32_t>(subgraphOutputs.data(), subgraphOutputs.size()),
103*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector(&quantizationOperator, 1));
104*89c4ff92SAndroid Build Coastguard Worker
105*89c4ff92SAndroid Build Coastguard Worker flatbuffers::Offset <flatbuffers::String> modelDescription =
106*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateString("ArmnnDelegate: Quantization Operator Model");
107*89c4ff92SAndroid Build Coastguard Worker flatbuffers::Offset <OperatorCode> operatorCode = CreateOperatorCode(flatBufferBuilder, quantizationOperatorCode);
108*89c4ff92SAndroid Build Coastguard Worker
109*89c4ff92SAndroid Build Coastguard Worker flatbuffers::Offset <Model> flatbufferModel =
110*89c4ff92SAndroid Build Coastguard Worker CreateModel(flatBufferBuilder,
111*89c4ff92SAndroid Build Coastguard Worker TFLITE_SCHEMA_VERSION,
112*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector(&operatorCode, 1),
113*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector(&subgraph, 1),
114*89c4ff92SAndroid Build Coastguard Worker modelDescription,
115*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector(buffers.data(), buffers.size()));
116*89c4ff92SAndroid Build Coastguard Worker
117*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.Finish(flatbufferModel, armnnDelegate::FILE_IDENTIFIER);
118*89c4ff92SAndroid Build Coastguard Worker
119*89c4ff92SAndroid Build Coastguard Worker return std::vector<char>(flatBufferBuilder.GetBufferPointer(),
120*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.GetBufferPointer() + flatBufferBuilder.GetSize());
121*89c4ff92SAndroid Build Coastguard Worker }
122*89c4ff92SAndroid Build Coastguard Worker
123*89c4ff92SAndroid Build Coastguard Worker template <typename InputT, typename OutputT>
QuantizationTest(tflite::BuiltinOperator quantizeOperatorCode,tflite::TensorType inputTensorType,tflite::TensorType outputTensorType,std::vector<armnn::BackendId> & backends,std::vector<int32_t> & inputShape,std::vector<int32_t> & outputShape,std::vector<InputT> & inputValues,std::vector<OutputT> & expectedOutputValues,float quantScale=1.0f,int quantOffset=0)124*89c4ff92SAndroid Build Coastguard Worker void QuantizationTest(tflite::BuiltinOperator quantizeOperatorCode,
125*89c4ff92SAndroid Build Coastguard Worker tflite::TensorType inputTensorType,
126*89c4ff92SAndroid Build Coastguard Worker tflite::TensorType outputTensorType,
127*89c4ff92SAndroid Build Coastguard Worker std::vector<armnn::BackendId>& backends,
128*89c4ff92SAndroid Build Coastguard Worker std::vector<int32_t>& inputShape,
129*89c4ff92SAndroid Build Coastguard Worker std::vector<int32_t>& outputShape,
130*89c4ff92SAndroid Build Coastguard Worker std::vector<InputT>& inputValues,
131*89c4ff92SAndroid Build Coastguard Worker std::vector<OutputT>& expectedOutputValues,
132*89c4ff92SAndroid Build Coastguard Worker float quantScale = 1.0f,
133*89c4ff92SAndroid Build Coastguard Worker int quantOffset = 0)
134*89c4ff92SAndroid Build Coastguard Worker {
135*89c4ff92SAndroid Build Coastguard Worker using namespace delegateTestInterpreter;
136*89c4ff92SAndroid Build Coastguard Worker std::vector<char> modelBuffer = CreateQuantizationTfLiteModel(quantizeOperatorCode,
137*89c4ff92SAndroid Build Coastguard Worker inputTensorType,
138*89c4ff92SAndroid Build Coastguard Worker outputTensorType,
139*89c4ff92SAndroid Build Coastguard Worker inputShape,
140*89c4ff92SAndroid Build Coastguard Worker outputShape,
141*89c4ff92SAndroid Build Coastguard Worker quantScale,
142*89c4ff92SAndroid Build Coastguard Worker quantOffset);
143*89c4ff92SAndroid Build Coastguard Worker
144*89c4ff92SAndroid Build Coastguard Worker // Setup interpreter with just TFLite Runtime.
145*89c4ff92SAndroid Build Coastguard Worker auto tfLiteInterpreter = DelegateTestInterpreter(modelBuffer);
146*89c4ff92SAndroid Build Coastguard Worker CHECK(tfLiteInterpreter.AllocateTensors() == kTfLiteOk);
147*89c4ff92SAndroid Build Coastguard Worker CHECK(tfLiteInterpreter.FillInputTensor(inputValues, 0) == kTfLiteOk);
148*89c4ff92SAndroid Build Coastguard Worker CHECK(tfLiteInterpreter.Invoke() == kTfLiteOk);
149*89c4ff92SAndroid Build Coastguard Worker std::vector<OutputT> tfLiteOutputValues = tfLiteInterpreter.GetOutputResult<OutputT>(0);
150*89c4ff92SAndroid Build Coastguard Worker std::vector<int32_t> tfLiteOutputShape = tfLiteInterpreter.GetOutputShape(0);
151*89c4ff92SAndroid Build Coastguard Worker
152*89c4ff92SAndroid Build Coastguard Worker // Setup interpreter with Arm NN Delegate applied.
153*89c4ff92SAndroid Build Coastguard Worker auto armnnInterpreter = DelegateTestInterpreter(modelBuffer, backends);
154*89c4ff92SAndroid Build Coastguard Worker CHECK(armnnInterpreter.AllocateTensors() == kTfLiteOk);
155*89c4ff92SAndroid Build Coastguard Worker CHECK(armnnInterpreter.FillInputTensor(inputValues, 0) == kTfLiteOk);
156*89c4ff92SAndroid Build Coastguard Worker CHECK(armnnInterpreter.Invoke() == kTfLiteOk);
157*89c4ff92SAndroid Build Coastguard Worker std::vector<OutputT> armnnOutputValues = armnnInterpreter.GetOutputResult<OutputT>(0);
158*89c4ff92SAndroid Build Coastguard Worker std::vector<int32_t> armnnOutputShape = armnnInterpreter.GetOutputShape(0);
159*89c4ff92SAndroid Build Coastguard Worker
160*89c4ff92SAndroid Build Coastguard Worker armnnDelegate::CompareOutputData<OutputT>(tfLiteOutputValues, armnnOutputValues, expectedOutputValues);
161*89c4ff92SAndroid Build Coastguard Worker armnnDelegate::CompareOutputShape(tfLiteOutputShape, armnnOutputShape, outputShape);
162*89c4ff92SAndroid Build Coastguard Worker
163*89c4ff92SAndroid Build Coastguard Worker tfLiteInterpreter.Cleanup();
164*89c4ff92SAndroid Build Coastguard Worker armnnInterpreter.Cleanup();
165*89c4ff92SAndroid Build Coastguard Worker }
166*89c4ff92SAndroid Build Coastguard Worker
167*89c4ff92SAndroid Build Coastguard Worker } // anonymous namespace