/aosp_15_r20/external/pytorch/test/functorch/ |
H A D | test_aotdispatch.py | 557 a.mul_(2) 599 a_.view(-1).mul_(2) 616 b.mul_(2) 664 a_view.mul_(2) 684 a.mul_(2) 812 a.mul_(2) 1055 out_ref[0].mul_(3) 1056 out_test[0].mul_(3) 1063 a.mul_(2) 1081 a.mul_(2) [all …]
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/ |
H A D | Loss.cpp | 301 loss.mul_(weight); in binary_cross_entropy_out_cpu() 354 grad_input.mul_(weight); in binary_cross_entropy_backward_out_cpu() 373 log_sigmoid_input.mul_(log_weight); in binary_cross_entropy_with_logits() 376 Tensor loss = (1 - target).mul_(input).sub_(log_sigmoid_input); in binary_cross_entropy_with_logits() 379 loss.mul_(weight); in binary_cross_entropy_with_logits() 406 at::mul_out(grad_input, target, z).mul_(-norm); in soft_margin_loss_backward_out() 408 grad_input.div_(z).mul_(grad_output); in soft_margin_loss_backward_out() 423 at::neg_out(output, input).mul_(target).exp_().log1p_(); in soft_margin_loss_out()
|
H A D | Dropout.cpp | 52 return input.mul_(noise); in multiply() 78 b = noise.add(-1).mul_(alpha * a).add_(alpha * a * p); in _dropout_impl() 79 noise.mul_(a); in _dropout_impl() 119 output = input.mul(mask).mul_(scale); in native_dropout_cpu()
|
/aosp_15_r20/external/libchrome-gestures/include/ |
H A D | accel_filter_interpreter.h | 52 CurveSegment() : x_(INFINITY), sqr_(0.0), mul_(1.0), int_(0.0) {} in CurveSegment() 54 : x_(x), sqr_(s), mul_(m), int_(b) {} in CurveSegment() 56 : x_(that.x_), sqr_(that.sqr_), mul_(that.mul_), int_(that.int_) {} in CurveSegment() 61 double mul_; // Slope of line (x multiplier) member
|
/aosp_15_r20/external/pytorch/test/autograd/ |
H A D | test_complex.py | 16 x2.mul_(2) 20 y0.mul_(2) 36 x0.mul_(2) 56 x0.mul_(2)
|
/aosp_15_r20/external/executorch/backends/mediatek/quantizer/ |
H A D | annotator.py | 135 torch.ops.aten.mul_.Scalar, 136 torch.ops.aten.mul_.Tensor, 235 torch.ops.aten.mul_.Scalar, 236 torch.ops.aten.mul_.Tensor,
|
/aosp_15_r20/external/libchrome-gestures/src/ |
H A D | accel_filter_interpreter_unittest.cc | 952 ASSERT_EQ(segs[0].mul_, 1.0); in TEST_F() 978 ASSERT_EQ(segs[0].mul_, linear_until_x / divisor); in TEST_F() 993 ASSERT_EQ(segs[1].mul_, 0.0); in TEST_F() 1010 ASSERT_EQ(segs[2].mul_, slope); in TEST_F() 1014 // return seg.mul_ + seg.int_ / speed;; in TEST_F() 1042 ASSERT_EQ(segs[0].mul_, 1.0); in TEST_F() 1068 ASSERT_EQ(segs[0].mul_, linear_until_x / divisor); in TEST_F() 1083 ASSERT_EQ(segs[1].mul_, 0.0); in TEST_F() 1100 ASSERT_EQ(segs[2].mul_, slope); in TEST_F()
|
/aosp_15_r20/external/OpenCL-CTS/test_conformance/subgroups/ |
H A D | subgroup_common_templates.h | 284 case ArithmeticOp::mul_: return a * b; in calculate() 310 case ArithmeticOp::mul_: { in calculate() 332 case ArithmeticOp::mul_: { in calculate() 351 case ArithmeticOp::mul_: return to_half(to_float(a) * to_float(b)); in calculate() 369 if (operation == ArithmeticOp::mul_ || operation == ArithmeticOp::add_) in generate_inputs() 384 if (operation == ArithmeticOp::mul_ in generate_inputs()
|
H A D | test_subgroup_non_uniform_arithmetic.cpp | 50 error |= rft.run_impl<T, SCIN_NU<T, ArithmeticOp::mul_>>( in run_functions_add_mul_max_min_for_type() 58 error |= rft.run_impl<T, SCEX_NU<T, ArithmeticOp::mul_>>( in run_functions_add_mul_max_min_for_type() 66 error |= rft.run_impl<T, RED_NU<T, ArithmeticOp::mul_>>( in run_functions_add_mul_max_min_for_type()
|
H A D | subhelpers.h | 257 mul_, enumerator 486 case ArithmeticOp::mul_: return (Ty)1; in identify_limits() 511 case ArithmeticOp::mul_: return (cl_int)1; 931 case ArithmeticOp::mul_: return (cl_float)1; 990 case ArithmeticOp::mul_: return (cl_double)1; 1077 case ArithmeticOp::mul_: return { 0x3c00 };
|
/aosp_15_r20/external/pytorch/test/inductor/ |
H A D | test_layout_optim.py | 207 y.mul_(2) 225 x.mul_(2) 242 x.mul_(2) 260 y.mul_(2)
|
/aosp_15_r20/external/pytorch/test/cpp/api/ |
H A D | grad_mode.cpp | 24 c.mul_(2); in TEST() 66 view_out.mul_( in TEST() 70 view_out.mul_(2); in TEST()
|
/aosp_15_r20/external/pytorch/torch/csrc/api/src/nn/modules/ |
H A D | _functions.cpp | 71 .mul_( in forward() 79 output.mul_(input); in forward() 115 .mul_(grad_output); in backward()
|
/aosp_15_r20/external/pytorch/torch/csrc/api/src/optim/ |
H A D | rmsprop.cpp | 116 square_avg.mul_(alpha).addcmul_(grad, grad, 1 - alpha); in step() 121 grad_avg.mul_(alpha).add_(grad, 1 - alpha); in step() 131 buf.mul_(options.momentum()).addcdiv_(grad, avg); in step()
|
H A D | adamw.cpp | 92 p.mul_(1 - options.lr() * options.weight_decay()); in step() 124 exp_avg.mul_(beta1).add_(grad, 1 - beta1); in step() 125 exp_avg_sq.mul_(beta2).addcmul_(grad, grad, 1 - beta2); in step()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/sparse/ |
H A D | SparseBlasImpl.cpp | 266 result.mul_(alpha); in _compressed_row_strided_addmm_out() 282 result.mul_(beta); in _compressed_row_strided_addmm_out() 287 mm.mul_(alpha); in _compressed_row_strided_addmm_out() 294 result.mul_(alpha); in _compressed_row_strided_addmm_out()
|
/aosp_15_r20/external/pytorch/torch/nn/modules/ |
H A D | _functions.py | 263 ctx.scale.mul_(ctx.alpha / ctx.size).add_(ctx.k) 266 output.mul_(input) 288 torch.pow(ctx.scale, -ctx.beta, out=grad_input).mul_(grad_output)
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/sparse/cuda/ |
H A D | SparseBlas.cpp | 59 result.mul_(beta); in sparse_sampled_addmm_out_sparse_csr_cuda() 151 result.mul_(beta); in addmm_out_sparse_compressed_cuda() 188 result.mul_(beta); in baddbmm_out_sparse_csr_cuda()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/functorch/ |
H A D | PyTorchOperatorHacks.cpp | 155 return input.mul_(noise); in multiply() 191 b = noise.add(-1).mul_(alpha * a).add_(alpha * a * p); in _dropout_impl() 192 noise.mul_(a); in _dropout_impl()
|
H A D | BatchRulesBinaryOps.cpp | 471 …VMAP_SUPPORT2(mul_, Tensor, SINGLE_ARG(binary_pointwise_inplace_batch_rule<TensorInplaceT, &Tensor… in TORCH_LIBRARY_IMPL() 472 …VMAP_SUPPORT2(mul_, Scalar, SINGLE_ARG(unary_inplace_batch_rule<ScalarInplaceT, &Tensor::mul_, con… in TORCH_LIBRARY_IMPL()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/test/ |
H A D | operators_test.cpp | 32 decltype(&ATEN_FN2(mul_, Tensor)), &ATEN_FN2(mul_, Tensor), in TEST()
|
/aosp_15_r20/external/pytorch/torch/optim/ |
H A D | adadelta.py | 286 square_avg.mul_(rho).addcmul_(grad, grad, value=1 - rho) 291 delta.div_(std).mul_(grad) 292 acc_delta.mul_(rho).addcmul_(delta, delta, value=1 - rho)
|
H A D | asgd.py | 247 param.mul_(1 - lambd * eta) 251 param.mul_(1 - lambd * eta_value) # decay term 256 ax.add_(param.sub(ax).mul_(mu))
|
H A D | _functional.py | 66 exp_avg_update_values = grad_values.sub(old_exp_avg_values).mul_(1 - beta1) 70 grad_values.pow(2).sub_(old_exp_avg_sq_values).mul_(1 - beta2)
|
/aosp_15_r20/external/pytorch/test/ |
H A D | test_autograd.py | 1263 tensor.mul_(4.0) 2061 a.mul_(2) 2068 a.mul_(2) 2069 a.mul_(2) 2109 var.mul_(2) 2117 var.mul_(2) 2129 view.mul_(2) 2157 b.mul_(2) 2175 b.mul_(2) 2176 b.mul_(2) [all …]
|