/aosp_15_r20/external/tensorflow/tensorflow/lite/delegates/gpu/common/tasks/ |
H A D | elementwise_test_util.cc | 39 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in AbsTest() 66 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in CosTest() 92 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in CopyTest() 117 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in EluTest() 145 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ExpTest() 173 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in FloorTest() 203 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in FloorDivTest() 237 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in FloorModTest() 270 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in HardSwishTest() 297 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in LogTest() [all …]
|
H A D | add_test_util.cc | 43 op_def.src_tensors.push_back({T, storage, Layout::HWC}); in AddTwoEqualIntTensorsTest() 44 op_def.src_tensors.push_back({T, storage, Layout::HWC}); in AddTwoEqualIntTensorsTest() 47 src_0 = op_def.src_tensors[0]; in AddTwoEqualIntTensorsTest() 48 src_1 = op_def.src_tensors[1]; in AddTwoEqualIntTensorsTest() 87 op_def.src_tensors.push_back({T, storage, Layout::HWC}); in AddTwoEqualUintTensorsTest() 88 op_def.src_tensors.push_back({T, storage, Layout::HWC}); in AddTwoEqualUintTensorsTest() 91 src_0 = op_def.src_tensors[0]; in AddTwoEqualUintTensorsTest() 92 src_1 = op_def.src_tensors[1]; in AddTwoEqualUintTensorsTest() 130 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in AddTwoEqualTensorsTest() 131 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in AddTwoEqualTensorsTest() [all …]
|
H A D | convolution_transposed_3x3_thin.cc | 80 AddSrcTensor("src_tensor", op_def.src_tensors[0]); in GenerateConvolutionTransposedCode() 83 if (op_def.src_tensors.size() == 2) { in GenerateConvolutionTransposedCode() 86 desc.element_type = op_def.src_tensors[1].GetDataType(); in GenerateConvolutionTransposedCode() 118 if (op_def.src_tensors[0].SupportsZeroClamp(Axis::WIDTH, gpu_info) && in GenerateConvolutionTransposedCode() 119 op_def.src_tensors[0].SupportsZeroClamp(Axis::HEIGHT, gpu_info)) { in GenerateConvolutionTransposedCode() 124 } else if (op_def.src_tensors[0].IsLinear() && in GenerateConvolutionTransposedCode() 125 op_def.src_tensors[0].ReturnsZeroForNegOneRead(gpu_info)) { in GenerateConvolutionTransposedCode() 253 gpu_info, definition.src_tensors[0].GetDataType(), attr.bias); in CreateConvolutionTransposed3x3Thin() 263 new_def.src_tensors = { in CreateConvolutionTransposed3x3ThinDynamicWeights() 264 definition.src_tensors[0]}; // leaving only src_tensor def, weights defs in CreateConvolutionTransposed3x3ThinDynamicWeights() [all …]
|
H A D | add.cc | 31 for (int i = 0; i < op_def.src_tensors.size(); ++i) { in CreateUnequalAdd() 33 op.AddSrcTensor(tensor_name, op_def.src_tensors[i]); in CreateUnequalAdd() 43 for (int i = 0; i < op_def.src_tensors.size(); ++i) { in CreateUnequalAdd() 55 for (int i = 0; i < op_def.src_tensors.size(); ++i) { in CreateUnequalAdd() 75 for (int i = 1; i < definition.src_tensors.size(); ++i) { in CreateAdd() 78 if (definition.src_tensors[i].HasAxis(Axis::DEPTH)) { in CreateAdd() 82 if (definition.src_tensors[i].HasAxis(Axis::BATCH)) { in CreateAdd()
|
H A D | max_unpooling.cc | 34 op->AddSrcTensor("src_tensor", op_def.src_tensors[0]); in GetMaxUnpoolingKernelCode() 35 op->AddSrcTensor("src_indices", op_def.src_tensors[1]); in GetMaxUnpoolingKernelCode() 75 if (!op_def.src_tensors[0].SupportsZeroClamp(Axis::WIDTH, gpu_info) || in GetMaxUnpoolingKernelCode() 76 !op_def.src_tensors[1].SupportsZeroClamp(Axis::WIDTH, gpu_info)) { in GetMaxUnpoolingKernelCode() 81 if (!op_def.src_tensors[0].SupportsZeroClamp(Axis::HEIGHT, gpu_info) || in GetMaxUnpoolingKernelCode() 82 !op_def.src_tensors[1].SupportsZeroClamp(Axis::HEIGHT, gpu_info)) { in GetMaxUnpoolingKernelCode() 88 if (!op_def.src_tensors[0].SupportsZeroClamp(Axis::DEPTH, gpu_info) || in GetMaxUnpoolingKernelCode() 89 !op_def.src_tensors[1].SupportsZeroClamp(Axis::DEPTH, gpu_info)) { in GetMaxUnpoolingKernelCode()
|
H A D | concat_test_util.cc | 46 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatWidthTest() 47 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatWidthTest() 79 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatHeightTest() 80 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatHeightTest() 113 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatChannelsTest() 114 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatChannelsTest() 115 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatChannelsTest() 151 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatChannelsAlignedx4Test() 152 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ConcatChannelsAlignedx4Test()
|
H A D | concat_xy.cc | 31 std::vector<std::string> tensor_names(op_def.src_tensors.size()); in GetConcatKernelCode() 32 for (int i = 0; i < op_def.src_tensors.size(); ++i) { in GetConcatKernelCode() 50 if (op_def.src_tensors[0].HasAxis(axis) && axis != Axis::BATCH) { in GetConcatKernelCode() 93 for (int i = 0; i < op_def.src_tensors.size(); ++i) { in GetConcatKernelCode() 97 if (op_def.src_tensors[i].HasAxis(Axis::BATCH)) { in GetConcatKernelCode() 117 for (int i = 0; i < definition.src_tensors.size(); ++i) { in CreateConcatXY() 119 op.AddSrcTensor(name, definition.src_tensors[i]); in CreateConcatXY()
|
H A D | convolution_transposed.cc | 134 AddSrcTensor("src_tensor", op_def.src_tensors[0]); in GenerateConvolutionTransposedCode() 137 if (op_def.src_tensors.size() != 1) { in GenerateConvolutionTransposedCode() 142 desc.element_type = op_def.src_tensors[1].GetDataType(); in GenerateConvolutionTransposedCode() 149 AddSrcTensor(name, definition_.src_tensors[1 + i]); in GenerateConvolutionTransposedCode() 154 const auto& src_def = op_def.src_tensors[0]; in GenerateConvolutionTransposedCode() 581 if (definition_.src_tensors[0].HasAxis(Axis::DEPTH)) { in BindArguments() 615 gpu_info, definition.src_tensors[0].GetDataType(), attr.bias); in CreateConvolutionTransposed() 628 gpu_info, definition.src_tensors[0].GetDataType(), attr.bias); in CreateConvolutionTransposed3D() 638 new_def.src_tensors = { in CreateConvolutionTransposedDynamicWeights() 639 definition.src_tensors[0]}; // leaving only src_tensor def, weights defs in CreateConvolutionTransposedDynamicWeights() [all …]
|
H A D | winograd.cc | 55 const auto src_desc = op_def.src_tensors[0]; in GetKernelWinograd4x4To36() 152 const auto src_desc = op_def.src_tensors[0]; in GetKernelWinograd36To4x4() 255 desc.AddSrcTensor("src_tensor", definition.src_tensors[0]); in CreateWinograd4x4To36() 291 const auto& src_desc = op_def.src_tensors[0]; in GetWinograd4x4To36TileX6Code() 292 AddSrcTensor("src_tensor", op_def.src_tensors[0]); in GetWinograd4x4To36TileX6Code() 474 definition_.src_tensors[0].GetDataType(), in UploadBt() 475 definition_.src_tensors[0].GetStorageType(), bt_aligned); in UploadBt() 550 desc.AddSrcTensor("src_tensor", definition.src_tensors[0]); in CreateWinograd36To4x4() 554 definition.src_tensors[0].GetDataType(), in CreateWinograd36To4x4() 555 definition.src_tensors[0].GetStorageType(), biases); in CreateWinograd36To4x4() [all …]
|
H A D | elementwise.cc | 254 gpu_info, definition.src_tensors[0].GetDataType(), constant_tensor); in CreateElementwiseTwoInput() 281 TensorDescriptor const_tensor_desc = definition.src_tensors[0]; in CreateElementwiseTwoInput() 407 op.AddSrcTensor("src_tensor", definition.src_tensors[0]); in CreateElementwiseOneInputWithBroadcast() 413 input_shape, definition.src_tensors[0], output_shape), in CreateElementwiseOneInputWithBroadcast() 432 op.AddSrcTensor("src_tensor", definition.src_tensors[0]); in CreateElementwiseWithBroadcast() 438 input_shape, definition.src_tensors[0], output_shape), in CreateElementwiseWithBroadcast() 455 op.AddSrcTensor("src0_tensor", definition.src_tensors[0]); in CreateElementwiseTwoInputWithBroadcast() 456 op.AddSrcTensor("src1_tensor", definition.src_tensors[1]); in CreateElementwiseTwoInputWithBroadcast() 461 first_input_shape, definition.src_tensors[0], in CreateElementwiseTwoInputWithBroadcast() 466 definition.src_tensors[1], output_shape), in CreateElementwiseTwoInputWithBroadcast()
|
H A D | split.cc | 33 AddSrcTensor("src_tensor", definition_.src_tensors[0]); in GetSplitCode() 60 if (definition_.src_tensors[0].HasAxis(Axis::BATCH)) { in GetSplitCode() 69 if (definition_.src_tensors[0].HasAxis(Axis::DEPTH)) { in GetSplitCode() 84 if (definition_.src_tensors[0].HasAxis(axis)) { in GetSplitCode() 122 AddSrcTensor("src_tensor", definition_.src_tensors[0]); in GetSplitChannelsCode() 128 definition_.src_tensors[0].HasAxis(Axis::BATCH) ? ", B" : ""; in GetSplitChannelsCode() 132 if (definition_.src_tensors[0].HasAxis(Axis::BATCH)) { in GetSplitChannelsCode() 141 if (definition_.src_tensors[0].HasAxis(Axis::DEPTH)) { in GetSplitChannelsCode()
|
H A D | padding_test_util.cc | 44 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in PaddingAppendWidthTest() 74 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in PaddingPrependWidthTest() 104 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in PaddingAppendHeightTest() 133 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in PaddingPrependHeightTest() 162 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in PaddingAppendChannelsTest() 191 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in PaddingPrependChannelsTest() 220 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in PaddingPrependChannelsX4Test() 249 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in PaddingComplexTest() 282 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in PaddingReflectWidthTest() 312 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in PaddingReflectChannelsTest()
|
H A D | reduce_test_util.cc | 44 op_def.src_tensors.push_back({T, storage, Layout::HWC}); in ReduceSumChannelsIntTest() 47 src_0 = op_def.src_tensors[0]; in ReduceSumChannelsIntTest() 84 op_def.src_tensors.push_back({T, storage, Layout::HWC}); in ReduceProductChannelsUIntTest() 87 src_0 = op_def.src_tensors[0]; in ReduceProductChannelsUIntTest() 124 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in MeanHWTest() 151 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ReduceSumChannelsTest() 182 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ReduceProductChannelsTest() 214 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ReduceMaxChannelsTest() 242 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ReduceMinChannelsTest()
|
H A D | gather_test_util.cc | 44 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in GatherWidthIntTest() 45 op_def.src_tensors.push_back({DataType::INT32, storage, Layout::HWC}); in GatherWidthIntTest() 48 src_0 = op_def.src_tensors[0]; in GatherWidthIntTest() 49 src_1 = op_def.src_tensors[1]; in GatherWidthIntTest() 86 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in GatherWidthTest() 87 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in GatherWidthTest()
|
H A D | tile.cc | 57 if (op_def.src_tensors[0].HasAxis(Axis::DEPTH)) { in GetTileCode() 61 if (op_def.src_tensors[0].HasAxis(Axis::BATCH)) { in GetTileCode() 66 if (op_def.src_tensors[0].HasAxis(Axis::DEPTH)) { in GetTileCode() 69 if (op_def.src_tensors[0].HasAxis(Axis::BATCH)) { in GetTileCode() 101 op.AddSrcTensor("src_tensor", op_def.src_tensors[0]); in CreateTile()
|
H A D | convolution_transposed_3x3.cc | 66 auto src_desc = op_def.src_tensors[0]; in GenerateConvolutionTransposedCode() 68 AddDstTensor("dst_tensor", op_def.src_tensors[0]); in GenerateConvolutionTransposedCode() 70 if (op_def.src_tensors.size() == 2) { in GenerateConvolutionTransposedCode() 73 desc.element_type = op_def.src_tensors[1].GetDataType(); in GenerateConvolutionTransposedCode() 419 gpu_info, definition.src_tensors[0].GetDataType(), attr.bias); in CreateConvolutionTransposed3x3() 429 new_def.src_tensors = { in CreateConvolutionTransposed3x3DynamicWeights() 430 definition.src_tensors[0]}; // leaving only src_tensor def, weights defs in CreateConvolutionTransposed3x3DynamicWeights() 434 new_def.src_tensors.push_back( in CreateConvolutionTransposed3x3DynamicWeights() 441 gpu_info, definition.src_tensors[0].GetDataType(), attr.bias); in CreateConvolutionTransposed3x3DynamicWeights()
|
H A D | cast_test_util.cc | 40 op_def.src_tensors.push_back( in CastTests() 44 src_desc = op_def.src_tensors[0]; in CastTests() 75 op_def.src_tensors.push_back( in CastToBoolTests() 79 src_desc = op_def.src_tensors[0]; in CastToBoolTests() 110 op_def.src_tensors.push_back({DataType::BOOL, storage_src, Layout::HWC}); in CastFromBoolTests() 114 src_desc = op_def.src_tensors[0]; in CastFromBoolTests()
|
H A D | convolution_transposed_4x4.cc | 83 auto src_desc = op_def.src_tensors[0]; in GenerateConvolutionTransposedCode() 87 if (op_def.src_tensors.size() == 2) { in GenerateConvolutionTransposedCode() 90 desc.element_type = op_def.src_tensors[1].GetDataType(); in GenerateConvolutionTransposedCode() 409 gpu_info, definition.src_tensors[0].GetDataType(), attr.bias); in CreateConvolutionTransposed4x4() 419 new_def.src_tensors = { in CreateConvolutionTransposed4x4DynamicWeights() 420 definition.src_tensors[0]}; // leaving only src_tensor def, weights defs in CreateConvolutionTransposed4x4DynamicWeights() 424 new_def.src_tensors.push_back( in CreateConvolutionTransposed4x4DynamicWeights() 430 gpu_info, definition.src_tensors[0].GetDataType(), attr.bias); in CreateConvolutionTransposed4x4DynamicWeights()
|
H A D | resize_test_util.cc | 45 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ResizeBilinearAlignedTest() 77 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ResizeBilinearNonAlignedTest() 110 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ResizeBilinearWithoutHalfPixelTest() 142 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ResizeBilinearWithHalfPixelTest() 174 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ResizeNearestTest() 206 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ResizeNearestAlignCornersTest() 238 op_def.src_tensors.push_back({data_type, storage, Layout::HWC}); in ResizeNearestHalfPixelCentersTest()
|
H A D | concat_z.cc | 40 std::vector<std::string> tensor_names(op_def.src_tensors.size()); in GetConcatKernelCode() 41 for (int i = 0; i < op_def.src_tensors.size(); ++i) { in GetConcatKernelCode() 52 for (int i = 0; i < op_def.src_tensors.size(); ++i) { in GetConcatKernelCode() 140 for (int i = 0; i < definition.src_tensors.size(); ++i) { in CreateConcatZ() 142 op.AddSrcTensor(name, definition.src_tensors[i]); in CreateConcatZ()
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/delegates/gpu/common/selectors/ |
H A D | operation_selector.cc | 99 TensorDescriptor src_transformed_desc = op_def.src_tensors[0]; in WinogradFromNode() 102 TensorDescriptor dst_transformed_desc = op_def.src_tensors[0]; in WinogradFromNode() 114 winograd_up_def.src_tensors.push_back(op_def.src_tensors[0]); in WinogradFromNode() 125 conv_def.src_tensors.push_back(src_transformed_desc); in WinogradFromNode() 138 winograd_down_def.src_tensors.push_back(dst_transformed_desc); in WinogradFromNode() 170 conv_temp_def.src_tensors[1] = {op_def.src_tensors[1].GetDataType(), in AddDynamicConv() 228 converter_def.src_tensors.push_back(op_def.src_tensors[1]); in AddDynamicConv() 229 for (int i = 1; i < conv_def.src_tensors.size(); ++i) { in AddDynamicConv() 230 converter_def.dst_tensors.push_back(conv_def.src_tensors[i]); in AddDynamicConv() 340 op_def.src_tensors[0].GetDataType(), in GPUOperationFromNodePart0() [all …]
|
/aosp_15_r20/external/ComputeLibrary/src/dynamic_fusion/sketch/ |
H A D | ArgumentPack.h | 173 std::vector<T *> src_tensors{}; in get_src_tensors() 179 src_tensors.push_back(tensor); in get_src_tensors() 182 return src_tensors; in get_src_tensors() 190 std::vector<const T *> src_tensors{}; in get_const_src_tensors() 196 src_tensors.push_back(tensor); in get_const_src_tensors() 199 return src_tensors; in get_const_src_tensors()
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/delegates/gpu/common/task/ |
H A D | gpu_operation.cc | 97 return src_tensors[0].GetDataType(); in GetPrimaryDataType() 100 return src_tensors[0].GetStorageType(); in GetPrimaryStorageType() 104 for (const auto& src : src_tensors) { in IsBatchSupported() 200 const auto link_value_type = elem1.definition_.src_tensors[0].GetDataType(); in FuseSimpleElemWithSimpleElem() 326 definition_.src_tensors.push_back( in AddOperation() 327 operation->definition_.src_tensors[i + 1]); in AddOperation() 367 *resutl = &definition_.src_tensors[index]; in GetTensorDescriptor() 410 std::make_unique<TensorDescriptor>(definition_.src_tensors[0])); in AssembleCode() 501 if (definition.src_tensors.size() > 1 && in CreateGpuOperation() 503 const auto second_tensor_def = definition.src_tensors[1]; in CreateGpuOperation() [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/delegates/gpu/cl/testing/ |
H A D | performance_profiling.cc | 162 std::vector<TensorFloat32> src_tensors(graph_cl.inputs().size()); in RunSerializedTest() local 164 src_tensors[i].id = graph_cl.inputs()[i]->id; in RunSerializedTest() 165 src_tensors[i].shape = graph_cl.inputs()[i]->tensor.shape; in RunSerializedTest() 166 src_tensors[i].data.resize(src_tensors[i].shape.DimensionsProduct()); in RunSerializedTest() 167 for (int j = 0; j < src_tensors[i].data.size(); ++j) { in RunSerializedTest() 168 src_tensors[i].data[j] = std::sin(j); in RunSerializedTest() 173 src_tensors[i], env.queue())); in RunSerializedTest() 198 graph_cl.inputs()[i]->id, src_tensors[i], env_v2.queue())); in RunSerializedTest()
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/delegates/gpu/common/tasks/special/ |
H A D | fc_fc_add.cc | 106 AddSrcTensor("src_tensor_0", op_def.src_tensors[0]); in GetFCFCAddKernelCode() 107 AddSrcTensor("src_tensor_1", op_def.src_tensors[1]); in GetFCFCAddKernelCode() 263 gpu_info, definition.src_tensors[0].GetDataType(), attr0.bias); in CreateFCFCAdd() 268 gpu_info, definition.src_tensors[0].GetDataType(), attr1.bias); in CreateFCFCAdd() 287 gpu_info, definition.src_tensors[0].GetDataType(), attr0.bias); in CreateFCFCAdd() 292 gpu_info, definition.src_tensors[0].GetDataType(), attr1.bias); in CreateFCFCAdd() 372 op_def.src_tensors.push_back(it->second); in TryFCFCAdd() 376 op_def.src_tensors.push_back(it->second); in TryFCFCAdd()
|