Home
last modified time | relevance | path

Searched full:padding_idx (Results 1 – 25 of 94) sorted by relevance

1234

/aosp_15_r20/external/pytorch/torch/nn/modules/
H A Dsparse.py25padding_idx (int, optional): If specified, the entries at :attr:`padding_idx` do not contribute to…
26 … therefore, the embedding vector at :attr:`padding_idx` is not updated during training,
28 … the embedding vector at :attr:`padding_idx` will default to all zeros,
87 >>> # example with padding_idx
88 >>> embedding = nn.Embedding(10, 3, padding_idx=0)
97 >>> padding_idx = 0
98 >>> embedding = nn.Embedding(3, 3, padding_idx=padding_idx)
105 ... embedding.weight[padding_idx] = torch.ones(3)
116 "padding_idx",
125 padding_idx: Optional[int]
[all …]
/aosp_15_r20/external/pytorch/torch/csrc/api/src/nn/modules/
H A Dembedding.cpp23 if (options.padding_idx() != std::nullopt) { in reset()
24 if (*options.padding_idx() > 0) { in reset()
26 *options.padding_idx() < options.num_embeddings(), in reset()
27 "Padding_idx must be within num_embeddings"); in reset()
28 } else if (*options.padding_idx() < 0) { in reset()
30 *options.padding_idx() >= -options.num_embeddings(), in reset()
31 "Padding_idx must be within num_embedding"); in reset()
32 options.padding_idx(options.num_embeddings() + *options.padding_idx()); in reset()
53 if (options.padding_idx() != std::nullopt) { in reset_parameters()
55 weight[*options.padding_idx()].fill_(0); in reset_parameters()
[all …]
/aosp_15_r20/external/pytorch/test/nn/
H A Dtest_embedding.py129 padding_idx = 2
133 embeddings[padding_idx] = padding_vec
134 embedding_nn = nn.Embedding.from_pretrained(embeddings, padding_idx=padding_idx)
135 self.assertEqual(embedding_nn.weight[padding_idx], padding_vec)
138 padding_idx = 2
141 embeddings, padding_idx=padding_idx
179 embed_old = embed_old.from_pretrained(embeddings, padding_idx=2)
181 res_F = F.embedding(a, embeddings, padding_idx=2)
206 embed_old = embed_old.from_pretrained(embeddings, padding_idx=2)
208 res_F = F.embedding_bag(a, embeddings, padding_idx=2)
[all …]
/aosp_15_r20/external/pytorch/aten/src/ATen/native/
H A DEmbeddingBag.cpp81 bool is_fast_path_index_select(const Tensor& src, Tensor& output, index_t padding_idx) { in is_fast_path_index_select() argument
85 padding_idx < static_cast<index_t>(0); in is_fast_path_index_select()
91 …h_index_select_scale(const Tensor& src, const Tensor& scale, Tensor& output, index_t padding_idx) { in is_fast_path_index_select_scale() argument
95 scale.strides()[0] == 1 && padding_idx < static_cast<index_t>(0); in is_fast_path_index_select_scale()
99 …_path(const Tensor& src, const std::optional<Tensor>& scale, Tensor& output, index_t padding_idx) { in is_fast_path() argument
101 is_fast_path_index_select_scale(src, scale.value(), output, padding_idx) : in is_fast_path()
102 is_fast_path_index_select(src, output, padding_idx); in is_fast_path()
118 index_t padding_idx, in index_select_add() argument
138 // We can skip indices equal to padding_idx so they are not included in in index_select_add()
145 if (idx != padding_idx) { in index_select_add()
[all …]
H A DEmbedding.cpp38 c10::SymInt padding_idx, bool scale_grad_by_freq, bool sparse) { in embedding_symint() argument
58 c10::SymInt padding_idx, bool scale_grad_by_freq, bool sparse) { in embedding_backward_symint() argument
67 padding_idx.guard_int(__FILE__, __LINE__), in embedding_backward_symint()
71 grad, indices, std::move(num_weights), padding_idx, scale_grad_by_freq); in embedding_backward_symint()
77 int64_t padding_idx, bool scale_grad_by_freq) { in embedding_sparse_backward() argument
90 if (padding_idx != -1) { in embedding_sparse_backward()
91 c10::List<std::optional<Tensor>> c({indices != padding_idx}); in embedding_sparse_backward()
100 // check if all our grad come from padding_idx in embedding_sparse_backward()
114 int64_t padding_idx, bool scale_grad_by_freq) { in embedding_dense_backward_cpu() argument
155 if (indices_data[i] != padding_idx) { in embedding_dense_backward_cpu()
/aosp_15_r20/external/pytorch/torch/distributed/_shard/sharding_spec/chunk_sharding_spec_ops/
H A D_common.py95 padding_idx=None, argument
118 padding_idx: If specified, the entries at padding_idx do
120 vector at padding_idx is not updated during training,
122 Note that the embedding vector at padding_idx is
139 padding_idx=padding_idx,
145 padding_idx=padding_idx,
300 def _handle_row_wise_mask(gather_inp, padding_idx, weight, world_size, rank): argument
303 on the current rank. This function also adjust the ``padding_idx``
313 padding_idx: If specified, the entries at padding_idx do
315 vector at padding_idx is not updated during training,
[all …]
H A Dembedding.py100 padding_idx = kwargs.get("padding_idx")
109 input, world_size, weight, local_shard, max_norm, norm_type, padding_idx, pg
121 padding_idx,
187 input, world_size, weight, local_shard, max_norm, norm_type, padding_idx, pg argument
203 padding_idx: If specified, the entries at padding_idx do
205 vector at padding_idx is not updated during training,
229 padding_idx=padding_idx,
235 input, world_size, weight, local_shard, max_norm, norm_type, padding_idx, rank, pg argument
251 padding_idx: If specified, the entries at padding_idx do
253 vector at padding_idx is not updated during training,
[all …]
H A Dembedding_bag.py66 defined by the offset (excluding the row specified in ``padding_idx``).
107 padding_idx = kwargs.get("padding_idx")
127 padding_idx,
143 padding_idx,
249 padding_idx, argument
269 padding_idx: If specified, the entries at padding_idx do
271 vector at padding_idx is not updated during training,
273 Note that the embedding vector at padding_idx is
307 padding_idx=padding_idx,
322 padding_idx, argument
[all …]
/aosp_15_r20/external/pytorch/torch/csrc/api/include/torch/nn/options/
H A Dembedding.h16 /// 2).padding_idx(3).max_norm(2).norm_type(2.5).scale_grad_by_freq(true).sparse(true));
25 /// If specified, the entries at `padding_idx` do not contribute to the
26 /// gradient; therefore, the embedding vector at `padding_idx` is not updated
28 /// Embedding, the embedding vector at `padding_idx` will default to all
31 TORCH_ARG(std::optional<int64_t>, padding_idx) = std::nullopt;
55 /// If specified, the entries at `padding_idx` do not contribute to the
56 /// gradient; therefore, the embedding vector at `padding_idx` is not updated
58 TORCH_ARG(std::optional<int64_t>, padding_idx) = std::nullopt;
84 /// If specified, the entries at `padding_idx` do not contribute to the
85 /// gradient; therefore, the embedding vector at `padding_idx` is not updated
[all …]
/aosp_15_r20/external/pytorch/torch/csrc/api/include/torch/nn/functional/
H A Dembedding.h27 std::optional<int64_t> padding_idx, in embedding() argument
34 if (padding_idx != std::nullopt) { in embedding()
35 if (*padding_idx > 0) { in embedding()
37 *padding_idx < weight.size(0), in embedding()
38 "Padding_idx must be within num_embeddings"); in embedding()
39 } else if (*padding_idx < 0) { in embedding()
41 *padding_idx >= -weight.size(0), in embedding()
42 "Padding_idx must be within num_embedding"); in embedding()
43 padding_idx = weight.size(0) + *padding_idx; in embedding()
46 padding_idx = -1; in embedding()
[all …]
/aosp_15_r20/external/pytorch/test/distributed/_shard/sharded_tensor/ops/
H A Dtest_embedding_bag.py42 padding_idx=None, argument
53 padding_idx=padding_idx,
63 padding_idx=padding_idx,
145 padding_idx=padding_idx,
156 padding_idx=padding_idx,
205 self._run_sharded_embedding_bag(spec, [5, 5], 17, 14, "sum", padding_idx=6)
244 self._run_sharded_embedding_bag(spec, [4, 3], 16, 14, "sum", padding_idx=12)
245 self._run_sharded_embedding_bag(spec, [4, 3], 16, 14, "mean", padding_idx=12)
246 self._run_sharded_embedding_bag(spec, [4, 3], 16, 14, "max", padding_idx=12)
255 padding_idx=10,
[all …]
H A Dtest_embedding.py39 padding_idx=None, argument
48 padding_idx=padding_idx,
56 padding_idx=padding_idx,
103 padding_idx=padding_idx,
110 padding_idx=padding_idx,
125 self._run_sharded_embedding(spec, [34], 15, 14, padding_idx=10)
126 self._run_sharded_embedding(spec, [8, 6, 5, 4], 23, 13, padding_idx=12)
167 self._run_sharded_embedding(spec, [6, 7, 6], 64, 11, padding_idx=30)
180 self._run_sharded_embedding(spec, [8, 6, 5, 4], 21, 11, padding_idx=10)
/aosp_15_r20/external/pytorch/aten/src/ATen/native/cuda/
H A DEmbeddingBag.cu67 index_t padding_idx, int64_t numRows) { in EmbeddingBag_updateOutputKernel_max() argument
88 bool pad = (input[emb] == padding_idx); in EmbeddingBag_updateOutputKernel_max()
118 index_t padding_idx, int64_t numRows) { in EmbeddingBag_updateOutputKernel_sum_mean() argument
139 bool pad = (input[emb] == padding_idx); in EmbeddingBag_updateOutputKernel_sum_mean()
176 int64_t padding_idx) { in embedding_bag_backward_cuda_sum_avg() argument
235 count, num_weights, padding_idx, mode == EmbeddingBagMode::MEAN, offset2bag, in embedding_bag_backward_cuda_sum_avg()
243 index_t padding_idx, const index_t numel) { in EmbeddingBag_accGradParametersKernel_max() argument
258 if (word_idx >= 0 && word_idx != padding_idx) { in EmbeddingBag_accGradParametersKernel_max()
271 int64_t padding_idx) { in embedding_bag_backward_cuda_max() argument
298 padding_idx, grad_weight.numel()); in embedding_bag_backward_cuda_max()
[all …]
H A DEmbedding.cu54 int padding_idx) in embedding_backward_feature_kernel() argument
89 if(src_row < n && f < s && dst_row != padding_idx) in embedding_backward_feature_kernel()
99 if(dst_row != padding_idx && src_row < n) // Per-warp exit condition, safe with ballot_sync in embedding_backward_feature_kernel()
137 index_t* count, int64_t numel, int64_t stride, int padding_idx) { in embedding_backward_kernel() argument
158 && input[idx] != padding_idx) { in embedding_backward_kernel()
249 int64_t num_weights, int64_t padding_idx, in embedding_dense_backward_cuda() argument
288 static_cast<int>(padding_idx)); in embedding_dense_backward_cuda()
344 sorted_indices, count, num_weights, padding_idx); in embedding_dense_backward_cuda()
/aosp_15_r20/external/executorch/kernels/test/
H A Dop_embedding_test.cpp30 int64_t padding_idx, in op_embedding_out() argument
38 padding_idx, in op_embedding_out()
65 /*padding_idx=*/0, in test_dtype()
133 /*padding_idx=*/0, in TEST_F()
175 /*padding_idx=*/0, in TEST_F()
214 /*padding_idx=*/0, in TEST_F()
250 /*padding_idx=*/0, in TEST_F()
280 /*padding_idx=*/0, in TEST_F()
310 /*padding_idx=*/0, in TEST_F()
320 /*padding_idx=*/0, in TEST_F()
[all …]
/aosp_15_r20/external/pytorch/torch/ao/nn/qat/modules/
H A Dembedding_ops.py32 padding_idx=None, argument
46 padding_idx,
66 self.padding_idx,
98 mod.padding_idx,
113 self.padding_idx,
153 padding_idx=None, argument
169 padding_idx,
192 self.padding_idx,
227 mod.padding_idx,
244 self.padding_idx,
/aosp_15_r20/external/pytorch/torch/ao/nn/quantized/reference/modules/
H A Dsparse.py26 padding_idx: Optional[int] = None,
39 padding_idx,
58 self.padding_idx,
70 mod.padding_idx,
101 padding_idx: Optional[int] = None,
116 padding_idx,
143 self.padding_idx,
158 mod.padding_idx,
/aosp_15_r20/external/pytorch/torch/csrc/api/include/torch/nn/modules/
H A Dembedding.h28 /// 2).padding_idx(3).max_norm(2).norm_type(2.5).scale_grad_by_freq(true).sparse(true));
80 .padding_idx(options.padding_idx())
104 …max_norm(2).norm_type(2.5).scale_grad_by_freq(true).sparse(true).mode(torch::kSum).padding_idx(1));
165 .padding_idx(options.padding_idx()));
/aosp_15_r20/external/pytorch/torch/nn/utils/_expanded_weights/
H A Dembedding_expanded_weights.py28 ctx.padding_idx, ctx.scale_grad_by_freq = (
29 expanded_kwargs["padding_idx"],
38 padding_idx, scale_grad_by_freq, sparse = (
39 ctx.padding_idx,
70 padding_idx,
/aosp_15_r20/external/pytorch/aten/src/ATen/functorch/
H A DBatchRulesModules.cpp26 c10::SymInt padding_idx, bool scale_grad_by_freq, bool sparse) { in embedding_batch_rule() argument
29 …auto result = at::embedding_symint(weight, indices, std::move(padding_idx), scale_grad_by_freq, sp… in embedding_batch_rule()
35 …auto result = at::embedding_symint(weight_, indices, std::move(padding_idx), scale_grad_by_freq, s… in embedding_batch_rule()
49 …auto result = at::embedding_symint(weight_, indices_, std::move(padding_idx), scale_grad_by_freq, … in embedding_batch_rule()
57 c10::SymInt num_weights, c10::SymInt padding_idx, bool scale_grad_by_freq) { in embedding_dense_backward_batch_rule() argument
64 grad, indices, std::move(num_weights), std::move(padding_idx), scale_grad_by_freq); in embedding_dense_backward_batch_rule()
78 if (padding_idx >= 0) { in embedding_dense_backward_batch_rule()
79 result.select_symint(1, std::move(padding_idx)).fill_(0); in embedding_dense_backward_batch_rule()
/aosp_15_r20/external/executorch/backends/apple/mps/runtime/operations/
H A DIndexingOps.mm77 int padding_idx = graphNode->padding_idx();
79 if (padding_idx != -1) {
80 MPSGraphTensor* constantTensor = [_mpsGraph constantWithScalar:padding_idx
/aosp_15_r20/external/pytorch/test/distributed/_tensor/
H A Dtest_embedding_ops.py148 self._run_embedding_op_test(mesh, 1, [34], 15, 14, padding_idx=10)
149 self._run_embedding_op_test(mesh, 1, [8, 6, 5, 4], 23, 13, padding_idx=12)
159 mesh, 1, [8, 6, 5, 4], 23, 13, padding_idx=12, max_norm=2.0
168 self._run_embedding_op_test(mesh, 0, [34], 15, 14, padding_idx=10)
/aosp_15_r20/external/executorch/exir/passes/
H A D_quant_patterns_and_replacements.py943 padding_idx,
954 out = torch.ops.aten.embedding.default(weight, indicies, padding_idx)
964 _, # padding_idx only matters for training and not when running op for inference
985 padding_idx,
999 out = torch.ops.aten.embedding.default(weight, indices, padding_idx)
1010 _, # padding_idx only matters for training and not when running op for inference
/aosp_15_r20/external/pytorch/torch/nn/
H A Dfunctional.py2440 padding_idx: Optional[int] = None,
2456 entries in :attr:`weight` at the row specified by :attr:`padding_idx`
2462 :attr:`padding_idx` to all zeros on construction.
2468padding_idx (int, optional): If specified, the entries at :attr:`padding_idx` do not contribute to…
2469 … therefore, the embedding vector at :attr:`padding_idx` is not updated during training,
2504 >>> # example with padding_idx
2509 >>> F.embedding(input, embedding_matrix, padding_idx=0)
2521 padding_idx=padding_idx,
2527 if padding_idx is not None:
2528 if padding_idx > 0:
[all …]
/aosp_15_r20/external/executorch/kernels/portable/cpu/
H A Dop_embedding.cpp71 // embedding.out(Tensor weight, Tensor indices, int padding_idx=-1, bool
77 int64_t padding_idx, in embedding_out() argument
82 (void)padding_idx; in embedding_out()

1234