Lines Matching full:conj
223 result: handle_r_to_c(result.scalar_type(), self_t.conj() * self_p.sgn())
226 self: grad * -((-self * self + 1).rsqrt()).conj()
231 other: handle_r_to_c(other.scalar_type(), maybe_multiply(grad, alpha.conj()))
239 self: maybe_multiply(grad, beta.conj())
240 …ize(0), batch1.sym_size(1), batch2.sym_size(2) }).bmm(batch2.transpose(1, 2).conj()), alpha.conj())
241 ….transpose(1, 2).conj().bmm(grad.unsqueeze(0).expand_symint({ batch1.sym_size(0), batch1.sym_size(…
246 tensor1: handle_r_to_c(tensor1.scalar_type(), grad * (value / tensor2).conj())
247 …nsor2: handle_r_to_c(tensor2.scalar_type(), -grad * (value * tensor1 / (tensor2 * tensor2)).conj())
252 tensor1: handle_r_to_c(tensor1.scalar_type(), grad * (tensor2 * value).conj())
253 tensor2: handle_r_to_c(tensor2.scalar_type(), grad * (tensor1 * value).conj())
257 self: maybe_multiply(grad, beta.conj())
268 self: maybe_multiply(grad, beta.conj())
269 mat: maybe_multiply(grad.ger(vec.conj()), alpha.conj())
270 vec: maybe_multiply(mat.t().conj().mv(grad), alpha.conj())
274 self: maybe_multiply(grad, beta.conj())
275 vec1: maybe_multiply(grad.mv(vec2.conj()), alpha.conj())
276 vec2: maybe_multiply(grad.t().mv(vec1.conj()), alpha.conj())
288 result: handle_r_to_c(result.scalar_type(), angle_backward(self_t.conj(), self_p).conj())
318 …self: "self.is_complex() ? grad * ((self + 1).rsqrt() * (self - 1).rsqrt()).conj() : grad * (self …
325 self: grad * (self.pow(2) + 1).rsqrt().conj()
332 self: grad * 1 / (1 - self.pow(2)).conj()
347 self: grad * (-self * self + 1).rsqrt().conj()
351 self: grad / (self * self + 1).conj()
359 self: maybe_multiply(grad, beta.conj())
360 batch1: maybe_multiply(grad.bmm(batch2.transpose(1, 2).conj()), alpha.conj())
361 batch2: maybe_multiply(batch1.transpose(1, 2).conj().bmm(grad), alpha.conj())
378 self: grad.bmm(mat2.transpose(1, 2).conj())
379 mat2: self.transpose(1, 2).conj().bmm(grad)
469 self: grad.conj()
470 result: self_t.conj()
494 self: grad * -self.sin().conj()
498 self: grad * self.sinh().conj()
508 self: at::linalg_cross(other.conj(), grad, dim)
509 other: at::linalg_cross(grad, self.conj(), dim)
598 self: grad * tensor.conj()
599 tensor: grad * self.conj()
603 self: grad.conj() * other
644 self: grad * result.conj()
648 self: grad * result.conj() * M_LN2
652 self: grad * (result.conj() + 1)
912 …self: "weight.isComplex() ? grad * (1 - weight.conj().toComplexDouble()) : grad * (1 - weight.toDo…
913 end: grad * weight.conj()
917 self: grad * (1 - weight).conj()
918 end: grad * weight.conj()
919 weight: grad * (end - self).conj()
939 self: grad.div(self.conj())
943 self: grad / (self.conj() * 2.3025850929940456)
951 self: grad / (self.conj() * 0.6931471805599453)
955 self: grad / (1 + exp(other - self)).conj()
956 other: grad / (1 + exp(self - other)).conj()
1191 self: grad.ger(vec.conj())
1192 vec: self.conj().t().mv(grad)
1351 …ow_backward_self(self_t.conj(), self_p, exponent_p) + pow_backward_exponent(exponent_t.conj(), sel…
1359 …ptions()).expand_as(result), self_p.to(result.scalar_type()), result) * self_t.conj()).sum().conj()
1363 …), self_p.to(result.scalar_type()), result, dim, keepdim) * self_t.conj()).sum(dim, keepdim).conj()
1392 self: -grad * (result * result).conj()
1449 self: -0.5 * grad * result.pow(3).conj()
1501 self: grad * self.cos().conj()
1509 self: grad * self.cosh().conj()
1582 self: grad / (2 * result.conj())
1620 …result: (at::real(var_backward(self_t.conj(), self_p, dim, correction, true).sum(dim.value_or(IntA…
1624 …result0: (at::real(var_backward(self_t.conj(), self_p, dim, correction, true).sum(dim.value_or(Int…
1630 other: handle_r_to_c(other.scalar_type(), maybe_multiply(-grad, alpha.conj()))
1638 self: handle_r_to_c(self.scalar_type(), maybe_multiply(-grad, alpha.conj()))
1643 self: handle_r_to_c(self.scalar_type(), maybe_multiply(-grad, alpha.conj()))
1710 self: grad * (1 + result.pow(2)).conj()
1862 …result: at::real(var_backward(self_t.conj(), self_p, dim, correction, true).sum(dim.value_or(IntAr…
1866 …result0: at::real(var_backward(self_t.conj(), self_p, dim, correction, true).sum(dim.value_or(IntA…
2009 …d(self_t.conj(), self_p, target_p, at::Reduction::None).conj() + mse_loss_backward(target_t.conj()…
2032 …conj(), self_p, target_p, at::Reduction::None, beta).conj() + smooth_l1_loss_backward(target_t.con…
2037 ….conj(), self_p, target_p, at::Reduction::None, delta).conj() + huber_loss_backward(target_t.conj(…
2041 …ply_loss_reduction(soft_margin_loss_backward(self_t.conj(), self_p, target, at::Reduction::None).c…
2105 …sult: self_t.copy_(leaky_relu_backward(original_self_t.conj(), result, negative_slope, true).conj(…
2109 output: log_sigmoid_backward(self_t.conj(), self_p, buffer).conj()
2169 result: self_t.copy_(threshold_backward(self_t.conj(), original_self_p, threshold).conj())
2639 grad_output: sigmoid_backward(grad, output.conj())
2640 output: grad.conj() * grad_output * (-2 * output.conj() + 1)
2641 … sigmoid_backward(grad_output_t, output_p) + output_t.conj() * grad_output_p * (-2 * output_p.conj…
2644 grad_output: tanh_backward(grad, output.conj())
2645 output: grad.conj() * (-2 * output.conj() * grad_output)
2646 …result: tanh_backward(grad_output_t, output_p) + output_t.conj() * (-2 * output_p.conj() * grad_ou…
3128 …w_backward_self(self_t.conj(), self_p, exponent_p) + pow_backward_exponent(exponent_t.conj(), self…
3132 result: pow_backward(self_t.conj(), self_p, exponent[i]).conj()