/aosp_15_r20/external/executorch/kernels/aten/ |
H A D | functions.yaml | 43 - op: add.Scalar_out 85 - op: bitwise_and.Scalar_out 89 - op: bitwise_or.Scalar_out 95 - op: bitwise_xor.Scalar_out 138 - op: div.Scalar_out 146 - op: eq.Scalar_out 160 - op: fill.Scalar_out 172 - op: fmod.Scalar_out 180 - op: ge.Scalar_out 190 - op: gt.Scalar_out [all …]
|
/aosp_15_r20/external/pytorch/test/expect/ |
H A D | HasDecompTest.test_aten_core_operators.expect | 38 aten::add.Scalar_out 88 aten::bitwise_and.Scalar_out 107 aten::bitwise_or.Scalar_out 123 aten::bitwise_xor.Scalar_out 129 aten::bucketize.Scalar_out 159 aten::copysign.Scalar_out 189 aten::div.Scalar_out 203 aten::eq.Scalar_out 242 aten::fmod.Scalar_out 255 aten::ge.Scalar_out [all …]
|
H A D | HasDecompTest.test_has_decomposition.expect | 2 aten::__lshift__.Scalar_out 6 aten::__rshift__.Scalar_out 16 aten::_add_relu.Scalar_out 140 aten::_foreach_add.Scalar_out 150 aten::_foreach_addcdiv.Scalar_out 159 aten::_foreach_addcmul.Scalar_out 179 aten::_foreach_clamp_max.Scalar_out 188 aten::_foreach_clamp_min.Scalar_out 206 aten::_foreach_div.Scalar_out 234 aten::_foreach_lerp.Scalar_out [all …]
|
/aosp_15_r20/external/executorch/kernels/portable/ |
H A D | functions.yaml | 80 - op: add.Scalar_out 175 - op: bitwise_and.Scalar_out 190 - op: bitwise_or.Scalar_out 200 - op: bitwise_xor.Scalar_out 301 - op: div.Scalar_out 322 - op: eq.Scalar_out 352 - op: fill.Scalar_out 382 - op: fmod.Scalar_out 405 - op: ge.Scalar_out 425 - op: gt.Scalar_out [all …]
|
/aosp_15_r20/external/pytorch/torch/distributed/tensor/_ops/ |
H A D | _pointwise_ops.py | 107 aten.bitwise_and.Scalar_out, 124 aten.bitwise_or.Scalar_out, 138 aten.bitwise_xor.Scalar_out, 156 aten.copysign.Scalar_out, 182 aten.eq.Scalar_out, 202 aten.float_power.Scalar_out, 213 aten.fmod.Scalar_out, 227 aten.gt.Scalar_out, 254 aten.lt.Scalar_out, 258 aten.lerp.Scalar_out, [all …]
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/ |
H A D | UpSample.h | 474 template <typename scalar_in, typename scalar_out, 475 …typename std::enable_if_t<!is_reduced_floating_point_v<scalar_out> || !std::is_same<scalar_in, flo… 476 void inline apply_grad_input(scalar_in* buffer_ptr, scalar_out* gin, int64_t size) { in apply_grad_input() 477 TORCH_CHECK((is_reduced_floating_point_v<scalar_out>), in apply_grad_input() 484 template <typename scalar_in, typename scalar_out, 485 …typename std::enable_if_t<is_reduced_floating_point_v<scalar_out> && std::is_same<scalar_in, float… 486 void inline apply_grad_input(scalar_in* buffer_ptr, scalar_out* gin, int64_t size) { in apply_grad_input() 487 using bVec = Vectorized<scalar_out>; in apply_grad_input() 492 auto [gin_fvec0, gin_fvec1] = convert_to_float<scalar_out>(gin_bvec); in apply_grad_input() 497 convert_from_float<scalar_out>(gin_fvec0, gin_fvec1).store(gin + d); in apply_grad_input()
|
H A D | native_functions.yaml | 610 autogen: _add_relu.Scalar_out 625 autogen: add.Scalar_out 1231 - func: copysign.Scalar_out(Tensor self, Scalar other, *, Tensor(a!) out) -> Tensor(a!) 2185 autogen: div.Scalar_out 2688 autogen: fill.Scalar_out 2764 autogen: floor_divide.Scalar_out 4215 autogen: mul.Scalar_out 6698 autogen: norm.Scalar_out 6864 autogen: sub.Scalar_out 6915 autogen: rsub.Scalar_out [all …]
|
/aosp_15_r20/external/executorch/kernels/portable/cpu/pattern/ |
H A D | comparison_op.h | 39 if (op == "eq.Tensor_out" || op == "eq.Scalar_out") { in get_comparison_fn() 42 if (op == "ne.Tensor_out" || op == "ne.Scalar_out") { in get_comparison_fn() 45 if (op == "ge.Tensor_out" || op == "ge.Scalar_out") { in get_comparison_fn() 48 if (op == "le.Tensor_out" || op == "le.Scalar_out") { in get_comparison_fn() 51 if (op == "gt.Tensor_out" || op == "gt.Scalar_out") { in get_comparison_fn() 54 if (op == "lt.Tensor_out" || op == "lt.Scalar_out") { in get_comparison_fn()
|
H A D | bitwise_op.h | 36 if (op == "bitwise_and.Tensor_out" || op == "bitwise_and.Scalar_out") { in get_bitwise_fn() 39 if (op == "bitwise_or.Tensor_out" || op == "bitwise_or.Scalar_out") { in get_bitwise_fn() 42 if (op == "bitwise_xor.Tensor_out" || op == "bitwise_xor.Scalar_out") { in get_bitwise_fn()
|
/aosp_15_r20/external/executorch/kernels/optimized/cpu/ |
H A D | op_le.cpp | 106 ET_SWITCH_REAL_TYPES_AND(Bool, a_type, ctx, "le.Scalar_out", CTYPE, [&]() { in opt_le_scalar_out() 108 Bool, b_type, ctx, "le.Scalar_out", CTYPE_B, [&]() { in opt_le_scalar_out() 122 Bool, a_type, ctx, "le.Scalar_out", CTYPE_A, [&]() { in opt_le_scalar_out() 124 Bool, b_type, ctx, "le.Scalar_out", CTYPE_B, [&]() { in opt_le_scalar_out() 126 Bool, common_type, ctx, "le.Scalar_out", CTYPE_IN, [&]() { in opt_le_scalar_out() 131 "le.Scalar_out", in opt_le_scalar_out()
|
H A D | op_div.cpp | 218 ET_SWITCH_REAL_TYPES(a_type, ctx, "div.Scalar_out", CTYPE, [&]() { in opt_div_scalar_out() 220 Bool, b_type, ctx, "div.Scalar_out", CTYPE_B, [&]() { in opt_div_scalar_out() 236 Bool, a_type, ctx, "div.Scalar_out", CTYPE_A, [&]() { in opt_div_scalar_out() 238 Bool, b_type, ctx, "div.Scalar_out", CTYPE_B, [&]() { in opt_div_scalar_out() 240 common_type, ctx, "div.Scalar_out", CTYPE_IN, [&]() { in opt_div_scalar_out() 242 out_type, ctx, "div.Scalar_out", CTYPE_OUT, [&]() { in opt_div_scalar_out()
|
H A D | op_add.cpp | 239 ET_SWITCH_REALB_TYPES(a_type, ctx, "add.Scalar_out", CTYPE, [&]() { in opt_add_scalar_out() 240 ET_SWITCH_SCALAR_OBJ_TYPES(b_type, ctx, "add.Scalar_out", CTYPE_B, [&]() { in opt_add_scalar_out() 258 ET_SWITCH_REALHBBF16_TYPES(a_type, ctx, "add.Scalar_out", CTYPE_A, [&]() { in opt_add_scalar_out() 259 ET_SWITCH_SCALAR_OBJ_TYPES(b_type, ctx, "add.Scalar_out", CTYPE_B, [&]() { in opt_add_scalar_out() 261 common_type, ctx, "add.Scalar_out", CTYPE_IN, [&]() { in opt_add_scalar_out() 263 out_type, ctx, "add.Scalar_out", CTYPE_OUT, [&]() { in opt_add_scalar_out()
|
H A D | op_sub.cpp | 273 ET_SWITCH_REAL_TYPES(a_type, ctx, "sub.Scalar_out", CTYPE, [&]() { in opt_sub_scalar_out() 275 b_type, ctx, "sub.Scalar_out", CTYPE_B, [&]() { in opt_sub_scalar_out() 293 ET_SWITCH_REALH_TYPES(a_type, ctx, "sub.Scalar_out", CTYPE_A, [&]() { in opt_sub_scalar_out() 295 b_type, ctx, "sub.Scalar_out", CTYPE_B, [&]() { in opt_sub_scalar_out() 297 common_type, ctx, "sub.Scalar_out", CTYPE_IN, [&]() { in opt_sub_scalar_out() 299 out_type, ctx, "sub.Scalar_out", CTYPE_OUT, [&]() { in opt_sub_scalar_out()
|
H A D | op_mul.cpp | 303 ET_SWITCH_REALB_TYPES(a_type, ctx, "mul.Scalar_out", CTYPE, [&]() { in opt_mul_scalar_out() 304 ET_SWITCH_SCALAR_OBJ_TYPES(b_type, ctx, "mul.Scalar_out", CTYPE_B, [&]() { in opt_mul_scalar_out() 318 ET_SWITCH_REALHBBF16_TYPES(a_type, ctx, "mul.Scalar_out", CTYPE_A, [&]() { in opt_mul_scalar_out() 319 ET_SWITCH_SCALAR_OBJ_TYPES(b_type, ctx, "mul.Scalar_out", CTYPE_B, [&]() { in opt_mul_scalar_out() 321 common_type, ctx, "mul.Scalar_out", CTYPE_IN, [&]() { in opt_mul_scalar_out() 323 out_type, ctx, "mul.Scalar_out", CTYPE_OUT, [&]() { in opt_mul_scalar_out()
|
/aosp_15_r20/external/executorch/kernels/optimized/ |
H A D | optimized-oss.yaml | 13 - op: add.Scalar_out 28 - op: div.Scalar_out 43 - op: le.Scalar_out 63 - op: mul.Scalar_out 83 - op: sub.Scalar_out
|
H A D | optimized.yaml | 15 - op: add.Scalar_out 30 - op: div.Scalar_out 50 - op: le.Scalar_out 75 - op: mul.Scalar_out 95 - op: sub.Scalar_out
|
/aosp_15_r20/external/pytorch/test/edge/ |
H A D | selected_operators.yaml | 140 aten::eq.Scalar_out: 182 aten::gt.Scalar_out: 224 aten::masked_fill.Scalar_out: 278 aten::ne.Scalar_out: 308 aten::remainder.Scalar_out: 326 aten::rsub.Scalar_out:
|
/aosp_15_r20/external/pytorch/aten/src/ATen/core/ |
H A D | NamedRegistrations.cpp | 165 m.impl("eq.Scalar_out", CppFunction::makeFallthrough()); in TORCH_LIBRARY_IMPL() 205 m.impl("ge.Scalar_out", CppFunction::makeFallthrough()); in TORCH_LIBRARY_IMPL() 210 m.impl("gt.Scalar_out", CppFunction::makeFallthrough()); in TORCH_LIBRARY_IMPL() 251 m.impl("le.Scalar_out", CppFunction::makeFallthrough()); in TORCH_LIBRARY_IMPL() 297 m.impl("lt.Scalar_out", CppFunction::makeFallthrough()); in TORCH_LIBRARY_IMPL() 353 m.impl("ne.Scalar_out", CppFunction::makeFallthrough()); in TORCH_LIBRARY_IMPL() 369 m.impl("pow.Scalar_out", CppFunction::makeFallthrough()); in TORCH_LIBRARY_IMPL()
|
/aosp_15_r20/external/executorch/kernels/portable/cpu/ |
H A D | op_masked_fill.cpp | 46 Bool, in_type, ctx, "masked_fill.Scalar_out", CTYPE, [&]() { in masked_fill_scalar_out() 48 Bool, val_type, ctx, "masked_fill.Scalar_out", CTYPE_VAL, [&]() { in masked_fill_scalar_out()
|
H A D | op_fill.cpp | 45 ET_SWITCH_REAL_TYPES_AND(Bool, a_type, ctx, "fill.Scalar_out", CTYPE_A, [&] { in fill_scalar_out() 47 ET_SWITCH_SCALAR_OBJ_TYPES(b_type, ctx, "fill.Scalar_out", CTYPE_B, [&] { in fill_scalar_out()
|
/aosp_15_r20/external/executorch/exir/operator/test/ |
H A D | test_operator.py | 25 self.assertEqual(out_schema, str(torch.ops.aten.mul.Scalar_out._schema)) 31 self.assertEqual(out_schema, str(torch.ops.aten.add.Scalar_out._schema))
|
/aosp_15_r20/external/federated-compute/fcp/tensorflow/ |
H A D | serve_slices_op.cc | 54 T* scalar_out) { in get_scalar_input() argument 57 *scalar_out = tensor->scalar<T>()(); in get_scalar_input()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/ |
H A D | library.cpp | 23 …m.def(TORCH_SELECTIVE_SCHEMA("quantized::add.Scalar_out(Tensor qa, Scalar b, Tensor(a!) out) -> Te… in TORCH_LIBRARY() 28 …m.def(TORCH_SELECTIVE_SCHEMA("quantized::add_relu.Scalar_out(Tensor qa, Scalar b, Tensor(a!) out) … in TORCH_LIBRARY() 202 …m.def(TORCH_SELECTIVE_SCHEMA("quantized::mul.Scalar_out(Tensor qa, Scalar b, Tensor(a!) out)-> Ten… in TORCH_LIBRARY() 207 …m.def(TORCH_SELECTIVE_SCHEMA("quantized::mul_relu.Scalar_out(Tensor qa, Scalar b, Tensor(a!) out)-… in TORCH_LIBRARY()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/ |
H A D | qmul.cpp | 352 …m.impl(TORCH_SELECTIVE_NAME("quantized::mul.Scalar_out"), TORCH_FN(QMulScalarOut</*ReLUFused=… in TORCH_LIBRARY_IMPL() 357 …m.impl(TORCH_SELECTIVE_NAME("quantized::mul_relu.Scalar_out"), TORCH_FN(QMulScalarOut</*ReLUFused=… in TORCH_LIBRARY_IMPL()
|
H A D | BinaryOps.cpp | 479 …m.impl(TORCH_SELECTIVE_NAME("quantized::add.Scalar_out"), TORCH_FN(qadd_scalar_out</*ReLUFuse… in TORCH_LIBRARY_IMPL() 484 …m.impl(TORCH_SELECTIVE_NAME("quantized::add_relu.Scalar_out"), TORCH_FN(qadd_scalar_out</*ReLUFuse… in TORCH_LIBRARY_IMPL()
|