/aosp_15_r20/external/pytorch/aten/src/ATen/native/cuda/ |
H A D | RNN.cu | 34 const TensorArg& input_bias, const TensorArg& hidden_bias, in checkSizes() argument 43 checkSameSize(c, input_bias, hidden_bias); in checkSizes() 49 checkAllSameGPU(c, {input_gates, hidden_gates, input_bias, hidden_bias, prev_hidden}); in checkSizes() 371 const Tensor& input_bias, const Tensor& hidden_bias, in lstm_forward_impl() argument 384 auto hidden_biasI = tryGetTensorInfo<scalar_t, index_type>(hidden_bias); in lstm_forward_impl() 392 if (allContiguous({input_gates, hidden_gates, input_bias, hidden_bias, cx, hy, cy, workspace})) { in lstm_forward_impl() 444 const Tensor& input_bias, const Tensor& hidden_bias, in gru_forward_impl() argument 457 auto hidden_biasI = tryGetTensorInfo<scalar_t, index_type>(hidden_bias); in gru_forward_impl() 464 if (allContiguous({input_gates, hidden_gates, input_bias, hidden_bias, hx, hy, workspace})) { in gru_forward_impl() 523 const Tensor& hidden_bias = c10::value_or_else(hidden_bias_opt, [] {return Tensor();}); in _thnn_fused_lstm_cell_cuda() local [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/python/ops/v1_compat_tests/ |
H A D | gradient_checker_test.py | 236 hidden_bias = constant_op.constant( 240 name="hidden_bias") 254 inp, hidden_weight, hidden_bias, softmax_weight, softmax_bias 259 [features], # hidden_bias 266 nn_ops.xw_plus_b(inp, hidden_weight, hidden_bias), name="features") 297 self.assertLess(self._BuildAndTestMiniMNIST(2, "hidden_bias"), 1e-8)
|
/aosp_15_r20/external/tensorflow/tensorflow/python/ops/ |
H A D | gradient_checker_v2_test.py | 312 hidden_bias = constant_op.constant( 316 name="hidden_bias") 329 all_params = [inp, hidden_weight, hidden_bias, softmax_weight, softmax_bias] 332 def f(inp, hidden_weight, hidden_bias, softmax_weight, softmax_bias): argument 334 nn_ops.xw_plus_b(inp, hidden_weight, hidden_bias), name="features") 367 self.assertLess(self._BuildAndTestMiniMNIST(2, "hidden_bias"), 1e-8)
|
/aosp_15_r20/external/libtextclassifier/native/lang_id/common/ |
H A D | embedding-network-params.h | 241 // ** Access methods for repeated MatrixParams hidden_bias. 246 // Returns number of rows of proto.hidden_bias(i). 249 // Returns number of columns of proto.hidden_bias(i). 252 // Returns pointer to elements of proto.hidden_bias(i), in row-major order.
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/ |
H A D | RNN.cpp | 1554 const Tensor& hidden_bias = c10::value_or_else(hidden_bias_opt, [] {return Tensor();}); in _thnn_differentiable_lstm_cell_backward() local 1563 if (hidden_bias.defined()) { in _thnn_differentiable_lstm_cell_backward() 1564 gates = gates + hidden_bias; in _thnn_differentiable_lstm_cell_backward() 1606 const Tensor& hidden_bias = c10::value_or_else(hidden_bias_opt, [] {return Tensor();}); in _thnn_differentiable_gru_cell_backward() local 1613 if (hidden_bias.defined()){ in _thnn_differentiable_gru_cell_backward() 1614 h_g = h_g + hidden_bias; in _thnn_differentiable_gru_cell_backward()
|
H A D | native_functions.yaml | 7747 …ates, Tensor hidden_gates, Tensor cx, Tensor? input_bias=None, Tensor? hidden_bias=None) -> (Tenso… 7762 …, Tensor input_gates, Tensor hidden_gates, Tensor? input_bias, Tensor? hidden_bias, Tensor cx, Ten… 7764 …ates, Tensor hidden_gates, Tensor hx, Tensor? input_bias=None, Tensor? hidden_bias=None) -> (Tenso… 7774 …put_gates, Tensor hidden_gates, Tensor hx, Tensor? input_bias, Tensor? hidden_bias) -> (Tensor, Te…
|
/aosp_15_r20/external/pytorch/tools/autograd/ |
H A D | derivatives.yaml | 2861 …ates, Tensor hidden_gates, Tensor cx, Tensor? input_bias=None, Tensor? hidden_bias=None) -> (Tenso… 2863 …, hidden_bias: "GradMode::is_enabled() ? _thnn_differentiable_lstm_cell_backward(grads[0], grads[1… 2865 …ates, Tensor hidden_gates, Tensor hx, Tensor? input_bias=None, Tensor? hidden_bias=None) -> (Tenso… 2866 …hidden_bias: "grad.defined() ? (GradMode::is_enabled() ? _thnn_differentiable_gru_cell_backward(gr…
|
/aosp_15_r20/external/pytorch/torch/ |
H A D | _meta_registrations.py | 5866 hidden_bias, argument 5883 input_bias.shape == hidden_bias.shape, 5884 lambda: f"{input_bias.shape} != {hidden_bias.shape}", 5895 for x in [hidden_gates, input_bias, hidden_bias, prev_hidden] 5907 hidden_bias=None, argument 5909 rnn_cell_checkSizes(input_gates, hidden_gates, input_bias, hidden_bias, 4, cx)
|
/aosp_15_r20/external/pytorch/torch/csrc/inductor/aoti_torch/generated/ |
H A D | c_shim_cuda.h | 45 …, AtenTensorHandle cx, AtenTensorHandle* input_bias, AtenTensorHandle* hidden_bias, AtenTensorHand…
|