/aosp_15_r20/external/eigen/unsupported/test/ |
H A D | cxx11_tensor_assign.cpp | 238 Tensor<int, 1> new_tensor(5); in test_auto_resize() local 239 new_tensor.setRandom(); in test_auto_resize() 241 tensor1 = tensor2 = tensor3 = tensor4 = new_tensor; in test_auto_resize() 243 VERIFY_IS_EQUAL(tensor1.dimension(0), new_tensor.dimension(0)); in test_auto_resize() 244 VERIFY_IS_EQUAL(tensor2.dimension(0), new_tensor.dimension(0)); in test_auto_resize() 245 VERIFY_IS_EQUAL(tensor3.dimension(0), new_tensor.dimension(0)); in test_auto_resize() 246 VERIFY_IS_EQUAL(tensor4.dimension(0), new_tensor.dimension(0)); in test_auto_resize() 247 for (int i = 0; i < new_tensor.dimension(0); ++i) { in test_auto_resize() 248 VERIFY_IS_EQUAL(tensor1(i), new_tensor(i)); in test_auto_resize() 249 VERIFY_IS_EQUAL(tensor2(i), new_tensor(i)); in test_auto_resize() [all …]
|
/aosp_15_r20/external/executorch/extension/tensor/test/ |
H A D | tensor_ptr_test.cpp | 350 auto new_tensor = make_tensor_ptr(*tensor); in TEST_F() local 352 EXPECT_EQ(new_tensor->dim(), tensor->dim()); in TEST_F() 353 EXPECT_EQ(new_tensor->size(0), tensor->size(0)); in TEST_F() 354 EXPECT_EQ(new_tensor->size(1), tensor->size(1)); in TEST_F() 356 new_tensor->const_data_ptr<int32_t>(), tensor->const_data_ptr<int32_t>()); in TEST_F() 357 EXPECT_EQ(new_tensor->scalar_type(), exec_aten::ScalarType::Int); in TEST_F() 395 auto new_tensor = make_tensor_ptr(*tensor); in TEST_F() local 397 EXPECT_EQ(new_tensor->dim(), tensor->dim()); in TEST_F() 398 EXPECT_EQ(new_tensor->size(0), tensor->size(0)); in TEST_F() 399 EXPECT_EQ(new_tensor->size(1), tensor->size(1)); in TEST_F() [all …]
|
/aosp_15_r20/external/pytorch/torch/ |
H A D | _tensor.py | 110 new_tensor = self.clone() 111 if type(new_tensor) is not type(self): 150 new_tensor = torch._utils._rebuild_qtensor( 163 if type(new_tensor) is not type(self): 171 new_tensor = self.new_empty([]) 172 if type(new_tensor) is not type(self): 181 new_tensor.set_( 185 new_tensor = new_tensor.conj_physical() 187 new_tensor = new_tensor.neg() 189 new_tensor.requires_grad_() [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/delegates/ |
H A D | utils_test.cc | 59 TfLiteTensor* new_tensor = nullptr; in TEST() local 63 /**new_type**/ kTfLiteUInt8, &new_tensor, &new_tensor_index), in TEST() 66 EXPECT_NE(new_tensor, nullptr); in TEST() 67 EXPECT_NE(new_tensor->dims, nullptr); in TEST() 68 EXPECT_EQ(new_tensor->type, kTfLiteUInt8); in TEST() 69 EXPECT_EQ(new_tensor->allocation_type, kTfLiteArenaRw); in TEST()
|
H A D | utils.cc | 34 TfLiteTensor** new_tensor, in CreateNewTensorWithDifferentType() argument 38 *new_tensor = &context->tensors[*new_tensor_index]; in CreateNewTensorWithDifferentType() 39 (*new_tensor)->type = new_type; in CreateNewTensorWithDifferentType() 40 (*new_tensor)->allocation_type = kTfLiteArenaRw; in CreateNewTensorWithDifferentType() 46 if (context->ResizeTensor(context, *new_tensor, dims) != kTfLiteOk) { in CreateNewTensorWithDifferentType()
|
/aosp_15_r20/external/pytorch/torch/onnx/ |
H A D | symbolic_opset11.py | 1317 new_tensor = tensor 1320 new_tensor = symbolic_helper._reshape_helper( 1321 g, new_tensor, g.op("Constant", value_t=torch.tensor([1])) 1323 new_tensor_list.append(new_tensor) 1344 new_tensor = tensor 1347 new_tensor = symbolic_helper._reshape_helper( 1348 g, new_tensor, g.op("Constant", value_t=torch.tensor([1, 1])) 1351 new_tensor = symbolic_helper._unsqueeze_helper( 1352 g, new_tensor, axes_i=[0] 1354 new_tensor_list.append(new_tensor) [all …]
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/ |
H A D | AutogradComposite.cpp | 52 auto new_tensor = at::zeros_symint({other_storage_numel}, other.options()); in _new_zeros_with_same_feature_meta() local 53 return new_tensor.as_strided_symint(other_sizes, other_strides, other_storage_offset); in _new_zeros_with_same_feature_meta() 86 auto new_tensor = at::zeros_symint({storage_numel}, other.options()); in _new_zeros_with_same_feature_meta() local 87 return new_tensor.as_strided_symint(out_sizes, out_strides, other_storage_offset); in _new_zeros_with_same_feature_meta()
|
H A D | Resize.cpp | 307 auto new_tensor = at::empty(src_tensor.sizes(), src_tensor.options()); in resize_bytes_nocuda() local 308 new_tensor.copy_(src_tensor); in resize_bytes_nocuda() 310 std::move(new_tensor.storage().mutable_data_ptr())); in resize_bytes_nocuda() 312 new_tensor.storage().unsafeGetStorageImpl()->allocator()); in resize_bytes_nocuda() 313 storage.set_nbytes(new_tensor.storage().nbytes()); in resize_bytes_nocuda()
|
H A D | ComplexHelper.h | 26 auto new_tensor = detail::make_tensor<TensorImpl>( in view_tensor() local 28 auto * impl = new_tensor.unsafeGetTensorImpl(); in view_tensor() 30 return new_tensor; in view_tensor()
|
/aosp_15_r20/external/tensorflow/tensorflow/core/common_runtime/eager/ |
H A D | custom_device_op_handler.cc | 81 tensorflow::ImmediateExecutionTensorHandle* new_tensor; in Execute() local 83 previous, target_device, &new_tensor)); in Execute() 84 Status s = op->SetInput(i, new_tensor); in Execute() 85 new_tensor->Unref(); in Execute()
|
/aosp_15_r20/external/pytorch/torch/distributed/tensor/ |
H A D | placement_types.py | 316 new_tensor = shard_dim_alltoall( 324 …new_tensor = unpad_tensor(new_tensor, self.dim, old_dim_unpad_size) # type: ignore[possibly-undef… 331 …new_tensor = unpad_tensor(new_tensor, new_shard_dim, new_dim_unpad_size) # type: ignore[possibly-… 333 return new_tensor
|
/aosp_15_r20/external/tensorflow/tensorflow/core/framework/ |
H A D | op_kernel.cc | 484 std::unique_ptr<Tensor> new_tensor = forward_input( in forward_input_to_output_with_shape() local 487 if (new_tensor != nullptr) { in forward_input_to_output_with_shape() 489 outputs_[output_index] = TensorValue(new_tensor.release()); in forward_input_to_output_with_shape() 584 std::unique_ptr<Tensor> new_tensor = in forward_input_or_allocate_temp() local 587 if (new_tensor != nullptr) { in forward_input_or_allocate_temp() 588 *out_temp = std::move(*new_tensor); in forward_input_or_allocate_temp() 754 Tensor new_tensor( in allocate_tensor() local 760 if (!new_tensor.IsInitialized()) { in allocate_tensor() 768 params_->step_id, new_tensor); in allocate_tensor() 770 *out_tensor = std::move(new_tensor); in allocate_tensor() [all …]
|
/aosp_15_r20/external/executorch/backends/arm/_passes/ |
H A D | match_arg_ranks_pass.py | 77 new_tensor = self.exported_program.state_dict[buffer_name].reshape( 80 self.exported_program.state_dict[buffer_name] = new_tensor 82 new_tensor, static_shapes=True
|
H A D | cast_int64_pass.py | 28 new_tensor = self.exported_program.state_dict[buffer_name].to( 31 self.exported_program.state_dict[buffer_name] = new_tensor
|
/aosp_15_r20/external/pytorch/torch/csrc/utils/ |
H A D | tensor_new.cpp | 1470 auto new_tensor = internal_new_from_data( in tensor_ctor() local 1482 new_tensor, *names, /*validate_names=*/true); in tensor_ctor() 1484 new_tensor.detach_(); // ensure new_tensor a leaf node in tensor_ctor() 1485 new_tensor.set_requires_grad(args_requires_grad); in tensor_ctor() 1486 return new_tensor; in tensor_ctor() 1510 Tensor new_tensor( in new_tensor() function 1534 auto new_tensor = new_from_data_copy( in new_tensor() local 1539 new_tensor.detach_(); // ensure new_tensor a leaf node in new_tensor() 1540 new_tensor.set_requires_grad(args_requires_grad); in new_tensor() 1541 return new_tensor; in new_tensor()
|
/aosp_15_r20/external/pytorch/torch/_inductor/ |
H A D | inductor_prims.py | 44 new_tensor = input_tensor.clone().as_strided( 48 new_tensor.copy_(input_tensor) 49 return new_tensor
|
/aosp_15_r20/external/pytorch/test/ |
H A D | test_serialization.py | 451 def __init__(self, new_tensor): argument 452 self.new_tensor = new_tensor 455 return (self.new_tensor.storage(), 456 self.new_tensor.storage_offset(), 457 tuple(self.new_tensor.size()), 458 self.new_tensor.stride()) 4211 new_tensor = torch.load(f) 4213 self.assertIsInstance(new_tensor, TestWrapperSubclass) 4214 self.assertEqual(new_tensor.elem, my_tensor.elem) 4215 self.assertEqual(new_tensor.foo, foo_val) [all …]
|
/aosp_15_r20/external/ComputeLibrary/src/dynamic_fusion/runtime/gpu/cl/ |
H A D | ClWorkloadRuntime.cpp | 103 auto new_tensor = inserted_pair->second.get(); in add_aux_tensor() local 104 _tensors.emplace_back(new_tensor, tensor_info, aux_memory_info); in add_aux_tensor() 105 return new_tensor; in add_aux_tensor()
|
/aosp_15_r20/external/tensorflow/tensorflow/python/framework/ |
H A D | sparse_tensor_test.py | 101 new_tensor = source.with_values([5.0, 1.0]) 102 self.assertAllEqual(new_tensor.indices, source.indices) 103 self.assertAllEqual(new_tensor.values, [5.0, 1.0]) 104 self.assertAllEqual(new_tensor.dense_shape, source.dense_shape)
|
/aosp_15_r20/external/tensorflow/tensorflow/dtensor/mlir/expansions/ |
H A D | scatter_spmd_expander.cc | 133 mlir::Value new_tensor, in TensorScatterOpExpand() 141 op->getLoc(), new_tensor.getType(), new_tensor, new_indices, new_updates); in TensorScatterOpExpand()
|
/aosp_15_r20/external/tensorflow/tensorflow/compiler/xla/service/ |
H A D | indexed_array_analysis.h | 332 T* new_tensor = new T(std::forward<Args>(args)...); in Construct() local 333 owned_tensors_.push_back(std::unique_ptr<T>(new_tensor)); in Construct() 334 return new_tensor; in Construct()
|
/aosp_15_r20/external/executorch/extension/gguf_util/converters/ |
H A D | llama_converter.py | 67 new_tensor = tensor.data.reshape(reversed_shape) 68 state_dict[nn_tensor_name] = torch.from_numpy(new_tensor)
|
/aosp_15_r20/external/pytorch/torch/distributed/fsdp/ |
H A D | _fsdp_extensions.py | 101 new_tensor, param_extension = fsdp_extension.pre_flatten_transform(tensor) 103 return new_tensor, param_extension
|
/aosp_15_r20/external/executorch/examples/models/llama/experimental/ |
H A D | load_gguf_q4_0.py | 140 new_tensor = tensor.data.reshape(reversed_shape) 141 state_dict[model_key] = torch.from_numpy(new_tensor)
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/tools/optimize/ |
H A D | quantize_model.cc | 105 std::unique_ptr<TensorT> new_tensor(new TensorT); in DuplicateBiasesWithMultipleUses() local 106 new_tensor->name = in DuplicateBiasesWithMultipleUses() 108 new_tensor->shape = bias_tensor->shape; in DuplicateBiasesWithMultipleUses() 109 new_tensor->type = bias_tensor->type; in DuplicateBiasesWithMultipleUses() 111 new_tensor->quantization = in DuplicateBiasesWithMultipleUses() 113 new_tensor->quantization->scale.assign( in DuplicateBiasesWithMultipleUses() 116 new_tensor->quantization->zero_point.assign( in DuplicateBiasesWithMultipleUses() 124 new_tensor->buffer = model->buffers.size() - 1; in DuplicateBiasesWithMultipleUses() 125 subgraph->tensors.push_back(std::move(new_tensor)); in DuplicateBiasesWithMultipleUses()
|