1*89c4ff92SAndroid Build Coastguard Worker //
2*89c4ff92SAndroid Build Coastguard Worker // Copyright © 2020, 2023 Arm Ltd and Contributors. All rights reserved.
3*89c4ff92SAndroid Build Coastguard Worker // SPDX-License-Identifier: MIT
4*89c4ff92SAndroid Build Coastguard Worker //
5*89c4ff92SAndroid Build Coastguard Worker
6*89c4ff92SAndroid Build Coastguard Worker #pragma once
7*89c4ff92SAndroid Build Coastguard Worker
8*89c4ff92SAndroid Build Coastguard Worker #include "TestUtils.hpp"
9*89c4ff92SAndroid Build Coastguard Worker
10*89c4ff92SAndroid Build Coastguard Worker #include <armnn_delegate.hpp>
11*89c4ff92SAndroid Build Coastguard Worker #include <DelegateTestInterpreter.hpp>
12*89c4ff92SAndroid Build Coastguard Worker
13*89c4ff92SAndroid Build Coastguard Worker #include <flatbuffers/flatbuffers.h>
14*89c4ff92SAndroid Build Coastguard Worker #include <tensorflow/lite/kernels/register.h>
15*89c4ff92SAndroid Build Coastguard Worker #include <tensorflow/lite/version.h>
16*89c4ff92SAndroid Build Coastguard Worker
17*89c4ff92SAndroid Build Coastguard Worker #include <schema_generated.h>
18*89c4ff92SAndroid Build Coastguard Worker
19*89c4ff92SAndroid Build Coastguard Worker #include <doctest/doctest.h>
20*89c4ff92SAndroid Build Coastguard Worker
21*89c4ff92SAndroid Build Coastguard Worker namespace
22*89c4ff92SAndroid Build Coastguard Worker {
23*89c4ff92SAndroid Build Coastguard Worker
24*89c4ff92SAndroid Build Coastguard Worker template <typename T>
CreateFullyConnectedTfLiteModel(tflite::TensorType tensorType,tflite::ActivationFunctionType activationType,const std::vector<int32_t> & inputTensorShape,const std::vector<int32_t> & weightsTensorShape,const std::vector<int32_t> & biasTensorShape,std::vector<int32_t> & outputTensorShape,std::vector<T> & weightsData,bool constantWeights=true,float quantScale=1.0f,int quantOffset=0,float outputQuantScale=2.0f,int outputQuantOffset=0)25*89c4ff92SAndroid Build Coastguard Worker std::vector<char> CreateFullyConnectedTfLiteModel(tflite::TensorType tensorType,
26*89c4ff92SAndroid Build Coastguard Worker tflite::ActivationFunctionType activationType,
27*89c4ff92SAndroid Build Coastguard Worker const std::vector <int32_t>& inputTensorShape,
28*89c4ff92SAndroid Build Coastguard Worker const std::vector <int32_t>& weightsTensorShape,
29*89c4ff92SAndroid Build Coastguard Worker const std::vector <int32_t>& biasTensorShape,
30*89c4ff92SAndroid Build Coastguard Worker std::vector <int32_t>& outputTensorShape,
31*89c4ff92SAndroid Build Coastguard Worker std::vector <T>& weightsData,
32*89c4ff92SAndroid Build Coastguard Worker bool constantWeights = true,
33*89c4ff92SAndroid Build Coastguard Worker float quantScale = 1.0f,
34*89c4ff92SAndroid Build Coastguard Worker int quantOffset = 0,
35*89c4ff92SAndroid Build Coastguard Worker float outputQuantScale = 2.0f,
36*89c4ff92SAndroid Build Coastguard Worker int outputQuantOffset = 0)
37*89c4ff92SAndroid Build Coastguard Worker {
38*89c4ff92SAndroid Build Coastguard Worker using namespace tflite;
39*89c4ff92SAndroid Build Coastguard Worker flatbuffers::FlatBufferBuilder flatBufferBuilder;
40*89c4ff92SAndroid Build Coastguard Worker std::array<flatbuffers::Offset<tflite::Buffer>, 5> buffers;
41*89c4ff92SAndroid Build Coastguard Worker buffers[0] = CreateBuffer(flatBufferBuilder);
42*89c4ff92SAndroid Build Coastguard Worker buffers[1] = CreateBuffer(flatBufferBuilder);
43*89c4ff92SAndroid Build Coastguard Worker
44*89c4ff92SAndroid Build Coastguard Worker auto biasTensorType = ::tflite::TensorType_FLOAT32;
45*89c4ff92SAndroid Build Coastguard Worker if (tensorType == ::tflite::TensorType_INT8)
46*89c4ff92SAndroid Build Coastguard Worker {
47*89c4ff92SAndroid Build Coastguard Worker biasTensorType = ::tflite::TensorType_INT32;
48*89c4ff92SAndroid Build Coastguard Worker }
49*89c4ff92SAndroid Build Coastguard Worker if (constantWeights)
50*89c4ff92SAndroid Build Coastguard Worker {
51*89c4ff92SAndroid Build Coastguard Worker buffers[2] = CreateBuffer(flatBufferBuilder,
52*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector(reinterpret_cast<const uint8_t*>(weightsData.data()),
53*89c4ff92SAndroid Build Coastguard Worker sizeof(T) * weightsData.size()));
54*89c4ff92SAndroid Build Coastguard Worker
55*89c4ff92SAndroid Build Coastguard Worker if (tensorType == ::tflite::TensorType_INT8)
56*89c4ff92SAndroid Build Coastguard Worker {
57*89c4ff92SAndroid Build Coastguard Worker std::vector<int32_t> biasData = { 10 };
58*89c4ff92SAndroid Build Coastguard Worker buffers[3] = CreateBuffer(flatBufferBuilder,
59*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector(reinterpret_cast<const uint8_t*>(biasData.data()),
60*89c4ff92SAndroid Build Coastguard Worker sizeof(int32_t) * biasData.size()));
61*89c4ff92SAndroid Build Coastguard Worker
62*89c4ff92SAndroid Build Coastguard Worker }
63*89c4ff92SAndroid Build Coastguard Worker else
64*89c4ff92SAndroid Build Coastguard Worker {
65*89c4ff92SAndroid Build Coastguard Worker std::vector<float> biasData = { 10 };
66*89c4ff92SAndroid Build Coastguard Worker buffers[3] = CreateBuffer(flatBufferBuilder,
67*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector(reinterpret_cast<const uint8_t*>(biasData.data()),
68*89c4ff92SAndroid Build Coastguard Worker sizeof(float) * biasData.size()));
69*89c4ff92SAndroid Build Coastguard Worker }
70*89c4ff92SAndroid Build Coastguard Worker }
71*89c4ff92SAndroid Build Coastguard Worker else
72*89c4ff92SAndroid Build Coastguard Worker {
73*89c4ff92SAndroid Build Coastguard Worker buffers[2] = CreateBuffer(flatBufferBuilder);
74*89c4ff92SAndroid Build Coastguard Worker buffers[3] = CreateBuffer(flatBufferBuilder);
75*89c4ff92SAndroid Build Coastguard Worker }
76*89c4ff92SAndroid Build Coastguard Worker buffers[4] = CreateBuffer(flatBufferBuilder);
77*89c4ff92SAndroid Build Coastguard Worker
78*89c4ff92SAndroid Build Coastguard Worker auto quantizationParameters =
79*89c4ff92SAndroid Build Coastguard Worker CreateQuantizationParameters(flatBufferBuilder,
80*89c4ff92SAndroid Build Coastguard Worker 0,
81*89c4ff92SAndroid Build Coastguard Worker 0,
82*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<float>({ quantScale }),
83*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<int64_t>({ quantOffset }));
84*89c4ff92SAndroid Build Coastguard Worker
85*89c4ff92SAndroid Build Coastguard Worker auto outputQuantizationParameters =
86*89c4ff92SAndroid Build Coastguard Worker CreateQuantizationParameters(flatBufferBuilder,
87*89c4ff92SAndroid Build Coastguard Worker 0,
88*89c4ff92SAndroid Build Coastguard Worker 0,
89*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<float>({ outputQuantScale }),
90*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<int64_t>({ outputQuantOffset }));
91*89c4ff92SAndroid Build Coastguard Worker
92*89c4ff92SAndroid Build Coastguard Worker std::array<flatbuffers::Offset<Tensor>, 4> tensors;
93*89c4ff92SAndroid Build Coastguard Worker tensors[0] = CreateTensor(flatBufferBuilder,
94*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<int32_t>(inputTensorShape.data(),
95*89c4ff92SAndroid Build Coastguard Worker inputTensorShape.size()),
96*89c4ff92SAndroid Build Coastguard Worker tensorType,
97*89c4ff92SAndroid Build Coastguard Worker 1,
98*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateString("input_0"),
99*89c4ff92SAndroid Build Coastguard Worker quantizationParameters);
100*89c4ff92SAndroid Build Coastguard Worker tensors[1] = CreateTensor(flatBufferBuilder,
101*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<int32_t>(weightsTensorShape.data(),
102*89c4ff92SAndroid Build Coastguard Worker weightsTensorShape.size()),
103*89c4ff92SAndroid Build Coastguard Worker tensorType,
104*89c4ff92SAndroid Build Coastguard Worker 2,
105*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateString("weights"),
106*89c4ff92SAndroid Build Coastguard Worker quantizationParameters);
107*89c4ff92SAndroid Build Coastguard Worker tensors[2] = CreateTensor(flatBufferBuilder,
108*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<int32_t>(biasTensorShape.data(),
109*89c4ff92SAndroid Build Coastguard Worker biasTensorShape.size()),
110*89c4ff92SAndroid Build Coastguard Worker biasTensorType,
111*89c4ff92SAndroid Build Coastguard Worker 3,
112*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateString("bias"),
113*89c4ff92SAndroid Build Coastguard Worker quantizationParameters);
114*89c4ff92SAndroid Build Coastguard Worker
115*89c4ff92SAndroid Build Coastguard Worker tensors[3] = CreateTensor(flatBufferBuilder,
116*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<int32_t>(outputTensorShape.data(),
117*89c4ff92SAndroid Build Coastguard Worker outputTensorShape.size()),
118*89c4ff92SAndroid Build Coastguard Worker tensorType,
119*89c4ff92SAndroid Build Coastguard Worker 4,
120*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateString("output"),
121*89c4ff92SAndroid Build Coastguard Worker outputQuantizationParameters);
122*89c4ff92SAndroid Build Coastguard Worker
123*89c4ff92SAndroid Build Coastguard Worker
124*89c4ff92SAndroid Build Coastguard Worker // create operator
125*89c4ff92SAndroid Build Coastguard Worker tflite::BuiltinOptions operatorBuiltinOptionsType = BuiltinOptions_FullyConnectedOptions;
126*89c4ff92SAndroid Build Coastguard Worker flatbuffers::Offset<void> operatorBuiltinOptions =
127*89c4ff92SAndroid Build Coastguard Worker CreateFullyConnectedOptions(flatBufferBuilder,
128*89c4ff92SAndroid Build Coastguard Worker activationType,
129*89c4ff92SAndroid Build Coastguard Worker FullyConnectedOptionsWeightsFormat_DEFAULT, false).Union();
130*89c4ff92SAndroid Build Coastguard Worker
131*89c4ff92SAndroid Build Coastguard Worker const std::vector<int> operatorInputs{0, 1, 2};
132*89c4ff92SAndroid Build Coastguard Worker const std::vector<int> operatorOutputs{3};
133*89c4ff92SAndroid Build Coastguard Worker flatbuffers::Offset <Operator> fullyConnectedOperator =
134*89c4ff92SAndroid Build Coastguard Worker CreateOperator(flatBufferBuilder,
135*89c4ff92SAndroid Build Coastguard Worker 0,
136*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<int32_t>(operatorInputs.data(), operatorInputs.size()),
137*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<int32_t>(operatorOutputs.data(), operatorOutputs.size()),
138*89c4ff92SAndroid Build Coastguard Worker operatorBuiltinOptionsType, operatorBuiltinOptions);
139*89c4ff92SAndroid Build Coastguard Worker
140*89c4ff92SAndroid Build Coastguard Worker const std::vector<int> subgraphInputs{0, 1, 2};
141*89c4ff92SAndroid Build Coastguard Worker const std::vector<int> subgraphOutputs{3};
142*89c4ff92SAndroid Build Coastguard Worker flatbuffers::Offset <SubGraph> subgraph =
143*89c4ff92SAndroid Build Coastguard Worker CreateSubGraph(flatBufferBuilder,
144*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector(tensors.data(), tensors.size()),
145*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<int32_t>(subgraphInputs.data(), subgraphInputs.size()),
146*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector<int32_t>(subgraphOutputs.data(), subgraphOutputs.size()),
147*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector(&fullyConnectedOperator, 1));
148*89c4ff92SAndroid Build Coastguard Worker
149*89c4ff92SAndroid Build Coastguard Worker flatbuffers::Offset <flatbuffers::String> modelDescription =
150*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateString("ArmnnDelegate: FullyConnected Operator Model");
151*89c4ff92SAndroid Build Coastguard Worker flatbuffers::Offset <OperatorCode> operatorCode = CreateOperatorCode(flatBufferBuilder,
152*89c4ff92SAndroid Build Coastguard Worker tflite::BuiltinOperator_FULLY_CONNECTED);
153*89c4ff92SAndroid Build Coastguard Worker
154*89c4ff92SAndroid Build Coastguard Worker flatbuffers::Offset <Model> flatbufferModel =
155*89c4ff92SAndroid Build Coastguard Worker CreateModel(flatBufferBuilder,
156*89c4ff92SAndroid Build Coastguard Worker TFLITE_SCHEMA_VERSION,
157*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector(&operatorCode, 1),
158*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector(&subgraph, 1),
159*89c4ff92SAndroid Build Coastguard Worker modelDescription,
160*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.CreateVector(buffers.data(), buffers.size()));
161*89c4ff92SAndroid Build Coastguard Worker
162*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.Finish(flatbufferModel, armnnDelegate::FILE_IDENTIFIER);
163*89c4ff92SAndroid Build Coastguard Worker
164*89c4ff92SAndroid Build Coastguard Worker return std::vector<char>(flatBufferBuilder.GetBufferPointer(),
165*89c4ff92SAndroid Build Coastguard Worker flatBufferBuilder.GetBufferPointer() + flatBufferBuilder.GetSize());
166*89c4ff92SAndroid Build Coastguard Worker }
167*89c4ff92SAndroid Build Coastguard Worker
168*89c4ff92SAndroid Build Coastguard Worker template <typename T>
FullyConnectedTest(std::vector<armnn::BackendId> & backends,tflite::TensorType tensorType,tflite::ActivationFunctionType activationType,const std::vector<int32_t> & inputTensorShape,const std::vector<int32_t> & weightsTensorShape,const std::vector<int32_t> & biasTensorShape,std::vector<int32_t> & outputTensorShape,std::vector<T> & inputValues,std::vector<T> & expectedOutputValues,std::vector<T> & weightsData,bool constantWeights=true,float quantScale=1.0f,int quantOffset=0)169*89c4ff92SAndroid Build Coastguard Worker void FullyConnectedTest(std::vector<armnn::BackendId>& backends,
170*89c4ff92SAndroid Build Coastguard Worker tflite::TensorType tensorType,
171*89c4ff92SAndroid Build Coastguard Worker tflite::ActivationFunctionType activationType,
172*89c4ff92SAndroid Build Coastguard Worker const std::vector <int32_t>& inputTensorShape,
173*89c4ff92SAndroid Build Coastguard Worker const std::vector <int32_t>& weightsTensorShape,
174*89c4ff92SAndroid Build Coastguard Worker const std::vector <int32_t>& biasTensorShape,
175*89c4ff92SAndroid Build Coastguard Worker std::vector <int32_t>& outputTensorShape,
176*89c4ff92SAndroid Build Coastguard Worker std::vector <T>& inputValues,
177*89c4ff92SAndroid Build Coastguard Worker std::vector <T>& expectedOutputValues,
178*89c4ff92SAndroid Build Coastguard Worker std::vector <T>& weightsData,
179*89c4ff92SAndroid Build Coastguard Worker bool constantWeights = true,
180*89c4ff92SAndroid Build Coastguard Worker float quantScale = 1.0f,
181*89c4ff92SAndroid Build Coastguard Worker int quantOffset = 0)
182*89c4ff92SAndroid Build Coastguard Worker {
183*89c4ff92SAndroid Build Coastguard Worker using namespace delegateTestInterpreter;
184*89c4ff92SAndroid Build Coastguard Worker
185*89c4ff92SAndroid Build Coastguard Worker std::vector<char> modelBuffer = CreateFullyConnectedTfLiteModel(tensorType,
186*89c4ff92SAndroid Build Coastguard Worker activationType,
187*89c4ff92SAndroid Build Coastguard Worker inputTensorShape,
188*89c4ff92SAndroid Build Coastguard Worker weightsTensorShape,
189*89c4ff92SAndroid Build Coastguard Worker biasTensorShape,
190*89c4ff92SAndroid Build Coastguard Worker outputTensorShape,
191*89c4ff92SAndroid Build Coastguard Worker weightsData,
192*89c4ff92SAndroid Build Coastguard Worker constantWeights,
193*89c4ff92SAndroid Build Coastguard Worker quantScale,
194*89c4ff92SAndroid Build Coastguard Worker quantOffset);
195*89c4ff92SAndroid Build Coastguard Worker
196*89c4ff92SAndroid Build Coastguard Worker // Setup interpreter with just TFLite Runtime.
197*89c4ff92SAndroid Build Coastguard Worker auto tfLiteInterpreter = DelegateTestInterpreter(modelBuffer);
198*89c4ff92SAndroid Build Coastguard Worker CHECK(tfLiteInterpreter.AllocateTensors() == kTfLiteOk);
199*89c4ff92SAndroid Build Coastguard Worker
200*89c4ff92SAndroid Build Coastguard Worker // Setup interpreter with Arm NN Delegate applied.
201*89c4ff92SAndroid Build Coastguard Worker auto armnnInterpreter = DelegateTestInterpreter(modelBuffer, backends);
202*89c4ff92SAndroid Build Coastguard Worker CHECK(armnnInterpreter.AllocateTensors() == kTfLiteOk);
203*89c4ff92SAndroid Build Coastguard Worker
204*89c4ff92SAndroid Build Coastguard Worker CHECK(tfLiteInterpreter.FillInputTensor<T>(inputValues, 0) == kTfLiteOk);
205*89c4ff92SAndroid Build Coastguard Worker CHECK(armnnInterpreter.FillInputTensor<T>(inputValues, 0) == kTfLiteOk);
206*89c4ff92SAndroid Build Coastguard Worker
207*89c4ff92SAndroid Build Coastguard Worker if (!constantWeights)
208*89c4ff92SAndroid Build Coastguard Worker {
209*89c4ff92SAndroid Build Coastguard Worker CHECK(tfLiteInterpreter.FillInputTensor<T>(weightsData, 1) == kTfLiteOk);
210*89c4ff92SAndroid Build Coastguard Worker CHECK(armnnInterpreter.FillInputTensor<T>(weightsData, 1) == kTfLiteOk);
211*89c4ff92SAndroid Build Coastguard Worker
212*89c4ff92SAndroid Build Coastguard Worker if (tensorType == ::tflite::TensorType_INT8)
213*89c4ff92SAndroid Build Coastguard Worker {
214*89c4ff92SAndroid Build Coastguard Worker std::vector <int32_t> biasData = {10};
215*89c4ff92SAndroid Build Coastguard Worker CHECK(tfLiteInterpreter.FillInputTensor<int32_t>(biasData, 2) == kTfLiteOk);
216*89c4ff92SAndroid Build Coastguard Worker CHECK(armnnInterpreter.FillInputTensor<int32_t>(biasData, 2) == kTfLiteOk);
217*89c4ff92SAndroid Build Coastguard Worker }
218*89c4ff92SAndroid Build Coastguard Worker else
219*89c4ff92SAndroid Build Coastguard Worker {
220*89c4ff92SAndroid Build Coastguard Worker std::vector<float> biasData = {10};
221*89c4ff92SAndroid Build Coastguard Worker CHECK(tfLiteInterpreter.FillInputTensor<float>(biasData, 2) == kTfLiteOk);
222*89c4ff92SAndroid Build Coastguard Worker CHECK(armnnInterpreter.FillInputTensor<float>(biasData, 2) == kTfLiteOk);
223*89c4ff92SAndroid Build Coastguard Worker }
224*89c4ff92SAndroid Build Coastguard Worker }
225*89c4ff92SAndroid Build Coastguard Worker
226*89c4ff92SAndroid Build Coastguard Worker CHECK(tfLiteInterpreter.Invoke() == kTfLiteOk);
227*89c4ff92SAndroid Build Coastguard Worker std::vector<T> tfLiteOutputValues = tfLiteInterpreter.GetOutputResult<T>(0);
228*89c4ff92SAndroid Build Coastguard Worker std::vector<int32_t> tfLiteOutputShape = tfLiteInterpreter.GetOutputShape(0);
229*89c4ff92SAndroid Build Coastguard Worker
230*89c4ff92SAndroid Build Coastguard Worker CHECK(armnnInterpreter.Invoke() == kTfLiteOk);
231*89c4ff92SAndroid Build Coastguard Worker std::vector<T> armnnOutputValues = armnnInterpreter.GetOutputResult<T>(0);
232*89c4ff92SAndroid Build Coastguard Worker std::vector<int32_t> armnnOutputShape = armnnInterpreter.GetOutputShape(0);
233*89c4ff92SAndroid Build Coastguard Worker
234*89c4ff92SAndroid Build Coastguard Worker armnnDelegate::CompareOutputData<T>(tfLiteOutputValues, armnnOutputValues, expectedOutputValues);
235*89c4ff92SAndroid Build Coastguard Worker armnnDelegate::CompareOutputShape(tfLiteOutputShape, armnnOutputShape, outputTensorShape);
236*89c4ff92SAndroid Build Coastguard Worker
237*89c4ff92SAndroid Build Coastguard Worker tfLiteInterpreter.Cleanup();
238*89c4ff92SAndroid Build Coastguard Worker armnnInterpreter.Cleanup();
239*89c4ff92SAndroid Build Coastguard Worker }
240*89c4ff92SAndroid Build Coastguard Worker
241*89c4ff92SAndroid Build Coastguard Worker } // anonymous namespace