/aosp_15_r20/external/tensorflow/tensorflow/python/distribute/v1/ |
H A D | all_reduce.py | 250 def build_ring_all_reduce(input_tensors, num_workers, num_subchunks, argument 273 if len(input_tensors) < 2: 275 input_tensors, shape = _flatten_tensors(input_tensors) 276 devices = [t.device for t in input_tensors] 280 input_tensors, devices, 293 def _build_ring_gather(input_tensors, devices, num_subchunks, argument 313 num_devices = len(input_tensors) 317 return input_tensors 318 shape = input_tensors[0].shape 328 splits, split_pad_len = _padded_split(input_tensors[d], num_chunks) [all …]
|
H A D | all_reduce_test.py | 89 input_tensors = [] 96 input_tensors.append(array_ops.identity(t8)) 97 return input_tensors, device_names 102 input_tensors, device_names = self._buildInput(1, 1) 104 output_tensors = ar._build_ring_gather(input_tensors, device_names, 1, 107 self.assertEqual(output_tensors, input_tensors) 109 input_tensors, device_names = self._buildInput(1, 4) 112 input_tensors, device_names, 2, pred_by_c_d, rank_by_c_d, math_ops.add) 115 self.assertEqual(len(output_tensors), len(input_tensors)) 116 num_chunks = 2 * len(input_tensors) [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/core/ops/ |
H A D | array_ops_test.cc | 310 op.input_tensors.resize(2); in TEST() 316 op.input_tensors[0] = &in_t; in TEST() 342 op.input_tensors.resize(3); in TEST() 343 op.input_tensors[2] = &axis_dim_t; in TEST() 450 op.input_tensors.resize(2); in TEST() 471 op.input_tensors[1] = &paddings_t; in TEST() 481 op.input_tensors.resize(3); in TEST() 502 op.input_tensors[1] = &paddings_t; in TEST() 511 op.input_tensors.resize(2); in TEST() 535 op.input_tensors[1] = &paddings_t; in TEST() [all …]
|
H A D | image_ops_test.cc | 35 op.input_tensors.resize(2); in TEST() 49 op.input_tensors[1] = &size_tensor; in TEST() 212 op.input_tensors.resize(2); in TEST() 234 op.input_tensors[1] = &size_tensor; in TEST() 246 op.input_tensors.resize(4); in TEST() 262 op.input_tensors[3] = &size_tensor; in TEST() 277 op.input_tensors.resize(2); in TEST() 288 op.input_tensors[1] = &size_tensor; in TEST() 294 op.input_tensors.resize(4); in TEST() 304 op.input_tensors[3] = &image_size; in TEST() [all …]
|
H A D | math_ops_test.cc | 294 op.input_tensors.resize(3); in TEST() 306 op.input_tensors[0] = &start_t; in TEST() 309 op.input_tensors[1] = &limit_t; in TEST() 313 op.input_tensors[2] = &delta_t; in TEST() 337 op.input_tensors.resize(3); in TEST() 347 op.input_tensors[2] = &num_t; in TEST() 357 op.input_tensors.resize(3); in TEST() 368 op.input_tensors[2] = &num_segments_t; in TEST() 379 op.input_tensors.resize(3); in TEST() 392 op.input_tensors.resize(4); in TEST() [all …]
|
H A D | data_flow_ops_test.cc | 148 op.input_tensors.push_back(nullptr); in TEST() 149 op.input_tensors.push_back(&tensor_5); in TEST() 152 op.input_tensors[0] = &tensor_2; in TEST() 153 op.input_tensors[1] = nullptr; in TEST() 157 op.input_tensors[1] = &tensor_5; in TEST() 242 op.input_tensors.push_back(nullptr); in TEST() 243 op.input_tensors.push_back(&n_tensor); in TEST() 274 op.input_tensors.push_back(nullptr); in TEST() 275 op.input_tensors.push_back(&n_tensor); in TEST()
|
H A D | random_ops_test.cc | 26 op.input_tensors.resize(2); in TEST() 34 op.input_tensors[1] = &num_samples; in TEST() 42 op.input_tensors.resize(2); in TEST() 49 op.input_tensors[0] = &shape; in TEST() 57 op.input_tensors.resize(2); in TEST() 64 op.input_tensors[0] = &shape; in TEST()
|
/aosp_15_r20/external/tensorflow/tensorflow/python/keras/ |
H A D | models.py | 110 tensor in tensor_map for tensor in nest.flatten(node.input_tensors)): 129 def _clone_functional_model(model, input_tensors=None, layer_fn=_clone_layer): argument 172 if input_tensors is not None: 174 input_tensors = nest.flatten(input_tensors) 175 for i, input_tensor in enumerate(input_tensors): 195 input_tensors, output_tensors, created_layers = ( 199 model = Model(input_tensors, output_tensors, name=model.name) 279 def _clone_sequential_model(model, input_tensors=None, layer_fn=_clone_layer): argument 321 if isinstance(layer, InputLayer) and input_tensors is not None: 332 if input_tensors is None: [all …]
|
/aosp_15_r20/packages/modules/NeuralNetworks/runtime/test/specs/AIDL_V2/ |
D | pack.mod.py | 16 def test(name, axis_value, input_tensors, output_tensor, inputs_data, output_data): argument 17 …model = Model().Operation("PACK", Int32Scalar("axis", axis_value), *input_tensors).To(output_tenso… 19 …quant8_asymm_dict = dict(zip([*input_tensors, output_tensor], [quant8_asymm_type] * (len(input_ten… 22 …uant8_asymm_signed_dict = dict(zip([*input_tensors, output_tensor], [quant8_asymm_signed_type] * (… 24 …Example((dict(zip(input_tensors, inputs_data)), {output_tensor: output_data}), model=model, name=n… 29 input_tensors=[Input("in0", ("TENSOR_FLOAT32", [2]))], 38 input_tensors=[Input("in0", ("TENSOR_FLOAT32", [2]))], 47 input_tensors=[Input("in0", ("TENSOR_FLOAT32", [3,4])), 60 input_tensors=[Input("in0", ("TENSOR_FLOAT32", [3,4])), 76 input_tensors=[Input("in0", ("TENSOR_FLOAT32", [3,4])),
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/ |
H A D | graph_info_test.cc | 132 EXPECT_EQ(generated_subgraphs[subgraph_index].input_tensors, in CheckPartitionSubgraphs() 133 expected_subgraphs[subgraph_index].input_tensors); in CheckPartitionSubgraphs() 177 expected_subgraph.input_tensors = {0}; in TEST() 195 expected_subgraph.input_tensors = {0}; in TEST() 215 expected_subgraph.input_tensors = {}; in TEST() 235 expected_subgraph.input_tensors = {0}; in TEST() 258 expected_subgraph0.input_tensors = {0}; in TEST() 263 expected_subgraph1.input_tensors = {1}; in TEST() 285 expected_subgraph0.input_tensors = {0}; in TEST() 290 expected_subgraph1.input_tensors = {1}; in TEST() [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/python/ |
H A D | lite.py | 989 input_tensors = [ 997 return graph_def, input_tensors, output_tensors 1000 def _validate_inputs(self, graph_def, input_tensors): argument 1023 for tensor in input_tensors: 1047 def _optimize_tf_model(self, graph_def, input_tensors, output_tensors, argument 1067 input_tensors, 1103 def convert(self, graph_def, input_tensors, output_tensors): argument 1121 self._validate_inputs(graph_def, input_tensors) 1137 input_tensors=input_tensors, 1193 graph_def, input_tensors, output_tensors = self._load_saved_model( [all …]
|
/aosp_15_r20/external/pytorch/test/cpp/aoti_inference/ |
H A D | test.cpp | 32 auto input_tensors = in test_aoti() local 49 auto actual_output_tensors = runner->run(input_tensors); in test_aoti() 99 auto input_tensors = in test_aoti_constants_update() local 128 auto actual_output_tensors = runner->run(input_tensors); in test_aoti_constants_update() 138 actual_output_tensors = runner->run(input_tensors); in test_aoti_constants_update() 145 actual_output_tensors = runner->run(input_tensors); in test_aoti_constants_update() 152 actual_output_tensors = runner->run(input_tensors); in test_aoti_constants_update() 156 actual_output_tensors = runner->run(input_tensors); in test_aoti_constants_update() 164 actual_output_tensors = runner->run(input_tensors); in test_aoti_constants_update() 188 auto input_tensors = in test_aoti_double_buffering() local [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/compiler/mlir/tfrt/benchmarks/ |
H A D | benchmark_mlir_function.cc | 52 llvm::SmallVector<Tensor> input_tensors; in GetInputTensors() local 57 input_tensors.emplace_back(spec.dtype, shape); in GetInputTensors() 62 input_tensors.back().flat<float>().setRandom(); in GetInputTensors() 65 input_tensors.back().flat<int64_t>().setRandom(); in GetInputTensors() 72 return input_tensors; in GetInputTensors() 105 llvm::SmallVector<Tensor> input_tensors = GetInputTensors(input_specs); in RunJitRtBenchmark() local 111 for (const Tensor& tensor : input_tensors) { in RunJitRtBenchmark() 179 llvm::SmallVector<Tensor> input_tensors = GetInputTensors(input_specs); in RunTfrtBenchmark() local 182 executor.Execute(function_name, input_tensors); in RunTfrtBenchmark() 205 llvm::SmallVector<Tensor> input_tensors = GetInputTensors(input_specs); in RunEigenBenchmark() local [all …]
|
/aosp_15_r20/external/armnn/python/pyarmnn/test/ |
H A D | test_runtime.py | 38 input_tensors = [const_tensor_pair] 51 yield preferred_backends, network, runtime, input_tensors, output_tensors 77 input_tensors = ann.make_input_tensors([input_binding_info], [input_tensor_data]) 87 yield runtime, net_id, input_tensors, output_tensors 207 input_tensors = random_runtime[3] 214 runtime.EnqueueWorkload(net_id, input_tensors, output_tensors) 221 input_tensors = [] 229 runtime.EnqueueWorkload(net_id, input_tensors, output_tensors) 243 input_tensors = mock_model_runtime[2] 249 runtime.EnqueueWorkload(net_id, input_tensors, output_tensors) [all …]
|
H A D | test_tensor_conversion.py | 49 input_tensors = ann.make_input_tensors(input_tensor_info, input_data) 50 assert len(input_tensors) == 1 52 for tensor, tensor_info in zip(input_tensors, input_tensor_info): 90 input_tensors = ann.make_input_tensors(input_tensor_info, input_data) 91 assert len(input_tensors) == 1 93 for tensor, tensor_info in zip(input_tensors, input_tensor_info):
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/experimental/mlir/testing/ |
H A D | mlir_convert.py | 30 def mlir_convert(options, graph_def, input_tensors, output_tensors, **kwargs): argument 49 input_arrays = [x[0] for x in input_tensors] 50 input_shapes = zip_test_utils.get_input_shapes_map(input_tensors) 76 def representative_dataset(input_tensors): argument 78 for _, shape, _ in input_tensors: 88 yield representative_dataset(input_tensors) 126 input_tensors, argument 156 for input_tensor in input_tensors: 161 input_types = ",".join([x[2] for x in input_tensors]) 176 ",".join([x[0] for x in input_tensors]),
|
/aosp_15_r20/external/pytorch/torch/_inductor/ |
H A D | autotune_process.py | 501 self, *input_tensors: torch.Tensor, output_tensor: torch.Tensor 511 *input_tensors: torch.Tensor, 518 *input_tensors: torch.Tensor, 527 assert len(input_tensors) == 0 528 input_tensors = tuple(x.to_tensor() for x in self.input_tensor_meta) 535 fn = self.make_run_fn(*input_tensors, output_tensor=output_tensor) 545 out = self.do_bench(fn, *input_tensors, output_tensor) 570 self, *input_tensors: torch.Tensor, output_tensor: Optional[torch.Tensor] = None 581 *input_tensors: torch.Tensor, 586 for tensor in [*input_tensors, output_tensor] [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/testing/ |
H A D | mlir_convert.py | 31 input_tensors, argument 84 def representative_dataset(input_tensors): argument 86 for name, shape, dtype in input_tensors: 95 yield representative_dataset(input_tensors) 133 input_tensors, argument 163 for input_tensor in input_tensors: 168 input_types = ",".join([x[2] for x in input_tensors]) 183 ",".join([x[0] for x in input_tensors]),
|
/aosp_15_r20/external/tensorflow/tensorflow/core/kernels/data/ |
H A D | zip_dataset_op.cc | 137 std::vector<Tensor> input_tensors; in Get() local 138 TF_RETURN_IF_ERROR(inputs_[i]->Get(ctx, index, &input_tensors)); in Get() 139 out_tensors->insert(out_tensors->end(), input_tensors.begin(), in Get() 140 input_tensors.end()); in Get() 194 std::vector<Tensor> input_tensors; in GetNextInternal() local 196 status.Update(input_impl->GetNext(&input_contexts_[i], &input_tensors, in GetNextInternal() 210 input_impls_[j]->GetNext(&input_contexts_[j], &input_tensors, in GetNextInternal() 215 out_tensors->insert(out_tensors->end(), input_tensors.begin(), in GetNextInternal() 216 input_tensors.end()); in GetNextInternal()
|
H A D | map_defun_op_test.cc | 44 std::vector<Tensor> input_tensors = arguments_; in GetInputTensors() local 45 input_tensors.insert(input_tensors.end(), captured_inputs_.begin(), in GetInputTensors() 47 return input_tensors; in GetInputTensors() 245 auto input_tensors = test_case.map_defun_op_params.GetInputTensors(); in TEST_P() local 247 for (auto& input : input_tensors) { in TEST_P() 274 auto input_tensors = test_case.map_defun_op_params.GetInputTensors(); in TEST_F() local 276 for (auto& input : input_tensors) { in TEST_F()
|
/aosp_15_r20/external/executorch/backends/qualcomm/runtime/ |
H A D | QnnExecuTorchBackend.cpp | 109 std::vector<std::shared_ptr<TensorWrapper>> input_tensors = in execute() local 116 input_tensor_structs.reserve(input_tensors.size()); in execute() 117 for (int i = 0; i < input_tensors.size(); ++i) { in execute() 119 args[i]->toTensor().mutable_data_ptr(), input_tensors[i]) != in execute() 122 input_tensors[i]->FillDataBuffer( in execute() 125 input_tensor_structs.push_back(input_tensors[i]->CloneTensorStruct()); in execute() 128 int output_index = input_tensors.size(); in execute()
|
/aosp_15_r20/external/tflite-support/tensorflow_lite_support/cc/task/text/nlclassifier/ |
H A D | bert_nl_classifier.cc | 68 const std::vector<TfLiteTensor*>& input_tensors, const std::string& input) { in Preprocess() argument 72 FindTensorByName(input_tensors, input_tensor_metadatas, kIdsTensorName); in Preprocess() 74 FindTensorByName(input_tensors, input_tensor_metadatas, kMaskTensorName); in Preprocess() 76 input_tensors, input_tensor_metadatas, kSegmentIdsTensorName); in Preprocess() 205 const auto& input_tensors = GetInputTensors(); in InitializeFromMetadata() local 206 const auto& ids_tensor = *FindTensorByName(input_tensors, input_tensor_metadatas, in InitializeFromMetadata() 208 const auto& mask_tensor = *FindTensorByName(input_tensors, input_tensor_metadatas, in InitializeFromMetadata() 210 const auto& segment_ids_tensor = *FindTensorByName(input_tensors, input_tensor_metadatas, in InitializeFromMetadata()
|
/aosp_15_r20/external/tensorflow/tensorflow/core/grappler/optimizers/data/ |
H A D | split_utils_test.cc | 71 const std::vector<Tensor>& input_tensors, in RunFunction() argument 77 for (int i = 0; i < input_tensors.size(); i++) { in RunFunction() 78 const Tensor& tensor = input_tensors[i]; in RunFunction() 107 const std::vector<Tensor>& input_tensors, in RunSplitFunctions() argument 113 for (int i = 0; i < input_tensors.size(); i++) { in RunSplitFunctions() 114 const Tensor& tensor = input_tensors[i]; in RunSplitFunctions() 127 for (int i = input_tensors.size() - num_captured_inputs; in RunSplitFunctions() 128 i < input_tensors.size(); i++) { in RunSplitFunctions() 155 const std::vector<Tensor>& input_tensors, in CheckSplitFunctions() argument 158 ASSERT_NO_FATAL_FAILURE(RunFunction(orig, input_tensors, &orig_outputs)); in CheckSplitFunctions() [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/core/kernels/mkl/ |
H A D | mkl_concat_op.cc | 469 OpInputList input_tensors; in Compute() local 470 GetMklInputList(context, "values", &input_tensors); in Compute() 471 const int N = input_tensors.size(); in Compute() 496 : input_tensors[0].shape(); in Compute() 503 s.IsMklTensor() ? s.GetTfShape() : input_tensors[i].shape(); in Compute() 533 if (input_tensors[i].NumElements() == 0) num_of_empty_inputs++; in Compute() 587 CallEigenVersion(context, input_tensors, input_mins, input_maxes, in Compute() 600 dst_dims = TFShapeToMklDnnDims(input_tensors[0].shape()); in Compute() 621 if (input_tensors[k].NumElements() == 0) continue; in Compute() 623 srcs[k].SetUsrMem(src_md, &input_tensors[k]); in Compute() [all …]
|
/aosp_15_r20/external/pytorch/benchmarks/operator_benchmark/common/ |
H A D | repeat_benchmark.py | 34 input_tensors = [torch.randn(*input_shape) for input_shape in input_shapes] 36 for input_tensor, repeat in zip(input_tensors, repeats): 39 return input_tensors, (total_num_elements * DTYPE_TO_BYTES["float"]) 42 input_tensors, total_bytes = generate_data_for_repeat() variable 52 for input_tensor, repeat in zip(input_tensors, repeats):
|