Home
last modified time | relevance | path

Searched refs:mutable_data_ptr (Results 1 – 25 of 308) sorted by relevance

12345678910>>...13

/aosp_15_r20/external/pytorch/aten/src/ATen/native/cuda/
H A DUniqueCub.cu83 int* inv_loc_ptr = static_cast<int*>(inv_loc.mutable_data_ptr()); in compute_unique()
98 inv_loc_out.mutable_data_ptr<int64_t>(), in compute_unique()
109 inverse_indices.mutable_data_ptr<int64_t>()); in compute_unique()
120 …cuda::cub::unique(data, data_out.mutable_data_ptr<scalar_t>(), length.mutable_data_ptr<int64_t>(),… in compute_unique()
126 data_out.mutable_data_ptr<scalar_t>(), in compute_unique()
127 counts.mutable_data_ptr<int64_t>(), in compute_unique()
128 length.mutable_data_ptr<int64_t>(), in compute_unique()
167 sorted.mutable_data_ptr<scalar_t>(), in operator ()()
177 sorted.mutable_data_ptr<scalar_t>(), in operator ()()
179 sorted_indices.mutable_data_ptr<int64_t>(), in operator ()()
[all …]
H A DEmbeddingBag.cu194 indices.const_data_ptr<index_t>(), sorted_indices.mutable_data_ptr<index_t>(), in embedding_bag_backward_cuda_sum_avg()
195 range.const_data_ptr<index_t>(), orig_indices.mutable_data_ptr<index_t>(), in embedding_bag_backward_cuda_sum_avg()
209 auto count_data = count.mutable_data_ptr<index_t>(); in embedding_bag_backward_cuda_sum_avg()
297 grad_weight.mutable_data_ptr<scalar_t>(), stride, numBags, in embedding_bag_backward_cuda_max()
404 weight.const_data_ptr<scalar_t>(), output.mutable_data_ptr<scalar_t>(), in _embedding_bag_cuda()
405 offset2bag.mutable_data_ptr<index_t>(), numIndices, numBags, featureSize, in _embedding_bag_cuda()
406 weight.stride(0), weight.stride(1), bag_size.mutable_data_ptr<index_t>(), in _embedding_bag_cuda()
407 max_indices.mutable_data_ptr<index_t>(), in _embedding_bag_cuda()
413 weight.const_data_ptr<scalar_t>(), output.mutable_data_ptr<scalar_t>(), in _embedding_bag_cuda()
414 offset2bag.mutable_data_ptr<index_t>(), numIndices, numBags, featureSize, in _embedding_bag_cuda()
[all …]
H A DEmbeddingBackwardKernel.cu236 int64_t *num_of_segments_ptr = num_of_segments_tensor.mutable_data_ptr<int64_t>(); in embedding_backward_cuda_kernel()
247 segment_offsets.mutable_data_ptr<index_t>(), in embedding_backward_cuda_kernel()
261 partials_per_segment.mutable_data_ptr<index_t>(), in embedding_backward_cuda_kernel()
275 partials_per_segment_offset.mutable_data_ptr<index_t>(), in embedding_backward_cuda_kernel()
280 … int64_t *num_of_partial_segments_ptr = num_of_partial_segments_tensor.mutable_data_ptr<int64_t>(); in embedding_backward_cuda_kernel()
294 partial_segment_offset.mutable_data_ptr<index_t>(), in embedding_backward_cuda_kernel()
330 … num_of_partial_segments_ptr, grad_weight_per_segment.mutable_data_ptr<partial_weight_t>(), in embedding_backward_cuda_kernel()
341 grad_weight_per_segment.mutable_data_ptr<partial_weight_t>(), in embedding_backward_cuda_kernel()
351 grad_weight.mutable_data_ptr<scalar_t>(), in embedding_backward_cuda_kernel()
H A DMultiMarginLoss.cu203 out.mutable_data_ptr<scalar_t>(), in multi_margin_loss_cuda_out()
214 out.mutable_data_ptr<scalar_t>(), in multi_margin_loss_cuda_out()
236 out.mutable_data_ptr<scalar_t>(), in multi_margin_loss_cuda_out()
246 out.mutable_data_ptr<scalar_t>(), in multi_margin_loss_cuda_out()
259 tmp_output.mutable_data_ptr<scalar_t>(), in multi_margin_loss_cuda_out()
269 tmp_output.mutable_data_ptr<scalar_t>(), in multi_margin_loss_cuda_out()
337 grad_input.mutable_data_ptr<scalar_t>(), in multi_margin_loss_cuda_backward_out()
350 grad_input.mutable_data_ptr<scalar_t>(), in multi_margin_loss_cuda_backward_out()
372 grad_input.mutable_data_ptr<scalar_t>(), in multi_margin_loss_cuda_backward_out()
384 grad_input.mutable_data_ptr<scalar_t>(), in multi_margin_loss_cuda_backward_out()
H A DLegacyThrustHelpers.cu32 auto orig_data = device_ptr(orig_indices.mutable_data_ptr<int64_t>()); in index_put_with_sort_kernel_thrust_helper()
40 auto sorted_data = device_ptr(sorted_indices.mutable_data_ptr<int64_t>()); in index_put_with_sort_kernel_thrust_helper()
58 auto count_data = thrust::device_ptr<index_t>(count.mutable_data_ptr<index_t>()); in embedding_dense_backward_cuda_scan()
96 auto dummy_dev = thrust::device_ptr<index_t>(dummy.mutable_data_ptr<index_t>()); in embedding_backward_cuda_kernel_unique_by_key()
103 thrust::device_ptr<index_t>(segment_offsets.mutable_data_ptr<index_t>())); in embedding_backward_cuda_kernel_unique_by_key()
H A DMultiLabelMarginCriterion.cu234 output.mutable_data_ptr<scalar_t>(), in multilabel_margin_loss_forward_out_cuda_template()
237 is_target_.mutable_data_ptr<scalar_t>(), in multilabel_margin_loss_forward_out_cuda_template()
259 output_tmp.mutable_data_ptr<scalar_t>(), in multilabel_margin_loss_forward_out_cuda_template()
262 is_target_.mutable_data_ptr<scalar_t>(), in multilabel_margin_loss_forward_out_cuda_template()
285 output.mutable_data_ptr<scalar_t>(), in multilabel_margin_loss_forward_out_cuda_template()
288 is_target_.mutable_data_ptr<scalar_t>(), in multilabel_margin_loss_forward_out_cuda_template()
344 grad_input.mutable_data_ptr<scalar_t>(), in multilabel_margin_loss_backward_cuda_out_template()
373 grad_input.mutable_data_ptr<scalar_t>(), in multilabel_margin_loss_backward_cuda_out_template()
H A DWeightNorm.cu379 (w.mutable_data_ptr<scalar_t>(), in weight_norm_cuda()
380 norms.mutable_data_ptr<accscalar_t>(), in weight_norm_cuda()
410 (w.mutable_data_ptr<scalar_t>(), in weight_norm_cuda()
411 norms.mutable_data_ptr<accscalar_t>(), in weight_norm_cuda()
468 (grad_v.mutable_data_ptr<scalar_t>(), in weight_norm_backward_cuda()
469 grad_g.mutable_data_ptr<scalar_t>(), in weight_norm_backward_cuda()
501 (grad_v.mutable_data_ptr<scalar_t>(), in weight_norm_backward_cuda()
502 grad_g.mutable_data_ptr<scalar_t>(), in weight_norm_backward_cuda()
H A DEmbedding.cu285 grad_weight.mutable_data_ptr<scalar_t>(), in embedding_dense_backward_cuda()
302 indices.const_data_ptr<index_t>(), sorted_indices.mutable_data_ptr<index_t>(), in embedding_dense_backward_cuda()
303 range.const_data_ptr<index_t>(), orig_indices.mutable_data_ptr<index_t>(), in embedding_dense_backward_cuda()
317 auto count_data = count.mutable_data_ptr<index_t>(); in embedding_dense_backward_cuda()
365 unique_indices.mutable_data_ptr<index_t>(), in embedding_renorm_cuda_()
366 num_unique_indices.mutable_data_ptr<int64_t>(), in embedding_renorm_cuda_()
382 self.mutable_data_ptr<scalar_t>(), in embedding_renorm_cuda_()
H A DNormalization.cuh1451 …accscalar_t* staging_data_ptr = grid.y > 1 ? staging_data.mutable_data_ptr<accscalar_t>() : nullpt… in batch_norm_stats_channels_last_cuda_template()
1452 int* semaphores_ptr = grid.y > 1 ? semaphores.mutable_data_ptr<int>() : nullptr; in batch_norm_stats_channels_last_cuda_template()
1456 out_mean.mutable_data_ptr<accscalar_t>(), in batch_norm_stats_channels_last_cuda_template()
1457 out_invstd.mutable_data_ptr<accscalar_t>(), in batch_norm_stats_channels_last_cuda_template()
1497 output.mutable_data_ptr<scalar_t>(), in batch_norm_elemt_channels_last_cuda_template()
1518 output.mutable_data_ptr<scalar_t>(), in batch_norm_elemt_channels_last_cuda_template()
1566 …accscalar_t* staging_data_ptr = grid.y > 1 ? staging_data.mutable_data_ptr<accscalar_t>() : nullpt… in batch_norm_backward_reduce_cuda_channels_last_template()
1567 int* semaphores_ptr = grid.y > 1 ? semaphores.mutable_data_ptr<int>() : nullptr; in batch_norm_backward_reduce_cuda_channels_last_template()
1574 sumn_dy.mutable_data_ptr<accscalar_t>(), in batch_norm_backward_reduce_cuda_channels_last_template()
1575 sum_dy_xmu.mutable_data_ptr<accscalar_t>(), in batch_norm_backward_reduce_cuda_channels_last_template()
[all …]
H A DNaiveDilatedConvolution.cu246 /* y=*/grad_bias.mutable_data_ptr<scalar_t>(), \ in slow_conv_dilated_all_cuda_template()
286 columns.mutable_data_ptr<scalar_t>()); in slow_conv_dilated_all_cuda_template()
302 /* C=*/output_n.mutable_data_ptr<scalar_t>(), in slow_conv_dilated_all_cuda_template()
327 /* C=*/columns.mutable_data_ptr<scalar_t>(), in slow_conv_dilated_all_cuda_template()
342 grad_input_n.mutable_data_ptr<scalar_t>()); in slow_conv_dilated_all_cuda_template()
358 columns.mutable_data_ptr<scalar_t>()); in slow_conv_dilated_all_cuda_template()
376 /* C=*/grad_weight.mutable_data_ptr<scalar_t>(), in slow_conv_dilated_all_cuda_template()
/aosp_15_r20/external/executorch/backends/cadence/reference/operators/
H A Dquantize_per_tensor.cpp35 uint8_t* out_data = out.mutable_data_ptr<uint8_t>(); in quantize_per_tensor_out()
39 int8_t* out_data = out.mutable_data_ptr<int8_t>(); in quantize_per_tensor_out()
45 uint16_t* out_data = out.mutable_data_ptr<uint16_t>(); in quantize_per_tensor_out()
49 int16_t* out_data = out.mutable_data_ptr<int16_t>(); in quantize_per_tensor_out()
53 int32_t* out_data = out.mutable_data_ptr<int32_t>(); in quantize_per_tensor_out()
/aosp_15_r20/external/executorch/backends/cadence/hifi/operators/
H A Dquantize_per_tensor.cpp36 uint8_t* out_data = out.mutable_data_ptr<uint8_t>(); in quantize_per_tensor_out()
40 int8_t* out_data = out.mutable_data_ptr<int8_t>(); in quantize_per_tensor_out()
44 int16_t* out_data = out.mutable_data_ptr<int16_t>(); in quantize_per_tensor_out()
50 uint16_t* out_data = out.mutable_data_ptr<uint16_t>(); in quantize_per_tensor_out()
54 int32_t* out_data = out.mutable_data_ptr<int32_t>(); in quantize_per_tensor_out()
H A Dop_div.cpp94 float* a_data = a.mutable_data_ptr<float>(); in div_out()
95 float* b_data = b.mutable_data_ptr<float>(); in div_out()
96 float* out_data = out.mutable_data_ptr<float>(); in div_out()
215 float* a_data = a.mutable_data_ptr<float>(); in div_out_mode()
216 float* b_data = b.mutable_data_ptr<float>(); in div_out_mode()
217 float* out_data = out.mutable_data_ptr<float>(); in div_out_mode()
/aosp_15_r20/external/executorch/extension/aten_util/test/
H A Dmake_aten_functor_from_et_functor_test.cpp30 out.mutable_data_ptr<int32_t>()[0] += 1; in add_1_out()
39 out.mutable_data_ptr<int64_t>()[0] += s1.value(); in add_optional_scalar_out()
42 out.mutable_data_ptr<int64_t>()[0] += s2.value(); in add_optional_scalar_out()
52 out.mutable_data_ptr<int64_t>()[0] += in add_optional_tensor_out()
53 s1.value().mutable_data_ptr<int64_t>()[0]; in add_optional_tensor_out()
56 out.mutable_data_ptr<int64_t>()[0] += in add_optional_tensor_out()
57 s2.value().mutable_data_ptr<int64_t>()[0]; in add_optional_tensor_out()
66 out.mutable_data_ptr<int64_t>()[0] += a[i]; in sum_arrayref_scalar_out()
75 out.mutable_data_ptr<int32_t>()[0] += a[i].const_data_ptr<int32_t>()[0]; in sum_arrayref_tensor_out()
86 out.mutable_data_ptr<int32_t>()[0] += in sum_arrayref_optional_tensor_out()
[all …]
/aosp_15_r20/external/executorch/kernels/optimized/cpu/
H A Dop_div.cpp83 out.mutable_data_ptr<CTYPE>(), in opt_div_out()
92 out.mutable_data_ptr<CTYPE>(), in opt_div_out()
117 out.mutable_data_ptr<CTYPE>(), in opt_div_out()
150 out.mutable_data_ptr<CTYPE>(), in opt_div_out()
158 out.mutable_data_ptr<CTYPE>(), in opt_div_out()
229 out.mutable_data_ptr<CTYPE>(), in opt_div_scalar_out()
251 out.mutable_data_ptr<CTYPE_OUT>(); in opt_div_scalar_out()
H A Dop_sub.cpp123 out.mutable_data_ptr<CTYPE>(), in opt_sub_out()
131 out.mutable_data_ptr<CTYPE>(), in opt_sub_out()
160 out.mutable_data_ptr<CTYPE>(), in opt_sub_out()
197 out.mutable_data_ptr<CTYPE>(), in opt_sub_out()
205 out.mutable_data_ptr<CTYPE>(), in opt_sub_out()
287 out.mutable_data_ptr<CTYPE>(), in opt_sub_scalar_out()
308 CTYPE_OUT* out_data = out.mutable_data_ptr<CTYPE_OUT>(); in opt_sub_scalar_out()
H A Dop_mul.cpp101 out.mutable_data_ptr<CTYPE>(), in handle_last_dim_broadcast()
170 out.mutable_data_ptr<CTYPE>(), in handle_broadcast_mul()
209 out.mutable_data_ptr<CTYPE>(), in opt_mul_out()
235 out.mutable_data_ptr<CTYPE>(), in opt_mul_out()
312 out.mutable_data_ptr<CTYPE>(), in opt_mul_scalar_out()
330 CTYPE_OUT* out_data = out.mutable_data_ptr<CTYPE_OUT>(); in opt_mul_scalar_out()
/aosp_15_r20/external/executorch/runtime/executor/test/
H A Dtensor_parser_test.cpp155 ASSERT_NE(t.mutable_data_ptr<float>()[0], 0.5); in TEST_F()
156 t.mutable_data_ptr<float>()[0] = 0.5; in TEST_F()
158 t.mutable_data_ptr<float>()[0], in TEST_F()
168 ASSERT_NE(t2.mutable_data_ptr<float>()[0], 0.5); in TEST_F()
174 t.mutable_data_ptr<float>()[0] = 0.5; in TEST_F()
183 ASSERT_NE(t3.mutable_data_ptr<float>()[0], 0.5); in TEST_F()
/aosp_15_r20/external/pytorch/aten/src/ATen/native/ao_sparse/quantized/cpu/
H A Dqlinear_serialize.cpp26 vec.data(), vec.data() + vec.size(), t.mutable_data_ptr<UNDERLYING_DTYPE>()); in wrap_vector()
132 weight_values.mutable_data_ptr<uint8_t>(), in serialize()
173 w_scales_compact.mutable_data_ptr<float>()[0] = w_scales_data_ptr[0]; in serialize()
174 w_zero_points_compact.mutable_data_ptr<int8_t>()[0] = in serialize()
186 w_scales_compact.mutable_data_ptr<float>()); in serialize()
194 w_zero_points_compact.mutable_data_ptr<int8_t>(), in serialize()
/aosp_15_r20/external/executorch/backends/qualcomm/runtime/
H A DQnnExecuTorchBackend.cpp119 args[i]->toTensor().mutable_data_ptr(), input_tensors[i]) != in execute()
132 void* mutable_data_ptr = in execute() local
133 args[output_index]->toTensor().mutable_data_ptr(); in execute()
134 if (qnn_manager->RegisterMem(mutable_data_ptr, output_tensor) != in execute()
136 output_tensor->FillDataBuffer(mutable_data_ptr, false /* copy_data */); in execute()
/aosp_15_r20/external/executorch/runtime/core/exec_aten/testing_util/test/
H A Dtensor_factory_test.cpp127 memcpy(expected.mutable_data_ptr<int32_t>(), data, sizeof(data)); in TEST_F()
156 memcpy(expected.mutable_data_ptr<float>(), data, sizeof(data)); in TEST_F()
186 memcpy(expected.mutable_data_ptr<bool>(), data, sizeof(data)); in TEST_F()
213 t1.mutable_data_ptr<int32_t>()[0] = 99; in TEST_F()
319 memcpy(expected.mutable_data_ptr<int>(), data, sizeof(data)); in TEST_F()
356 memcpy(expected.mutable_data_ptr<float>(), data, sizeof(data)); in TEST_F()
393 memcpy(expected.mutable_data_ptr<bool>(), data, sizeof(data)); in TEST_F()
460 t1.mutable_data_ptr<int32_t>()[0] = 99; in TEST_F()
513 memcpy(expected.mutable_data_ptr<int>(), data, sizeof(data)); in TEST_F()
627 memcpy(expected.mutable_data_ptr<int32_t>(), data, sizeof(data)); in TEST_F()
[all …]
/aosp_15_r20/external/pytorch/aten/src/ATen/native/nested/
H A DNestedTensorMatmul.cpp100 int64_t* self_new_sizes_ptr = self_new_sizes.mutable_data_ptr<int64_t>(); in matmul_with_bmm_nested()
103 int64_t* self_new_strides_ptr = self_new_strides.mutable_data_ptr<int64_t>(); in matmul_with_bmm_nested()
105 int64_t* self_new_offsets_ptr = self_new_offsets.mutable_data_ptr<int64_t>(); in matmul_with_bmm_nested()
109 int64_t* mat2_new_sizes_ptr = mat2_new_sizes.mutable_data_ptr<int64_t>(); in matmul_with_bmm_nested()
112 int64_t* mat2_new_strides_ptr = mat2_new_strides.mutable_data_ptr<int64_t>(); in matmul_with_bmm_nested()
114 int64_t* mat2_new_offsets_ptr = mat2_new_offsets.mutable_data_ptr<int64_t>(); in matmul_with_bmm_nested()
159 int64_t* out_new_offsets_ptr = out_new_offsets.mutable_data_ptr<int64_t>(); in matmul_with_bmm_nested()
/aosp_15_r20/external/pytorch/aten/src/ATen/native/
H A DNaiveDilatedConvolution.cpp269 columns.mutable_data_ptr<scalar_t>(), in slow_conv_dilated_all_cpu_template()
312 /* C=*/output_n.mutable_data_ptr<scalar_t>(), in slow_conv_dilated_all_cpu_template()
327 /* C=*/output_n.mutable_data_ptr<scalar_t>(), in slow_conv_dilated_all_cpu_template()
378 /* C=*/columns.mutable_data_ptr<scalar_t>(), in slow_conv_dilated_all_cpu_template()
393 /* C=*/columns.mutable_data_ptr<scalar_t>(), in slow_conv_dilated_all_cpu_template()
424 columns.mutable_data_ptr<scalar_t>(), in slow_conv_dilated_all_cpu_template()
468 /* C=*/grad_weight.mutable_data_ptr<scalar_t>(), in slow_conv_dilated_all_cpu_template()
483 /* C=*/grad_weight.mutable_data_ptr<scalar_t>(), in slow_conv_dilated_all_cpu_template()
/aosp_15_r20/external/executorch/extension/kernel_util/test/
H A Dmake_boxed_from_unboxed_functor_test.cpp38 out.mutable_data_ptr<int32_t>()[0] = 1; in set_1_out()
50 out.mutable_data_ptr<int32_t>()[i] = sum; in add_tensor_out()
62 out.mutable_data_ptr<int32_t>()[0] += s1.value(); in add_optional_scalar_out()
65 out.mutable_data_ptr<int32_t>()[0] += s2.value(); in add_optional_scalar_out()
78 out.mutable_data_ptr<int32_t>()[j] += in add_optional_tensor_out()
/aosp_15_r20/external/executorch/kernels/portable/cpu/
H A Dop_linear_scratch_example.cpp89 scratch.mutable_data_ptr<scalar_t>() + (i * K + j); in linear_scratch_example()
109 scratch.mutable_data_ptr<scalar_t>() + (i * K + j); in linear_scratch_example()
110 scalar_t* out_ptr = out.mutable_data_ptr<scalar_t>() + (i * K + j); in linear_scratch_example()
111 scalar_t* bias_ptr = bias.value().mutable_data_ptr<scalar_t>() + j; in linear_scratch_example()

12345678910>>...13