/aosp_15_r20/external/pytorch/torch/distributed/_composable/fsdp/ |
H A D | _fsdp_collectives.py | 91 …"split_with_sizes_copy(Tensor all_gather_output, SymInt[] all_gather_input_split_sizes, int dim=0,… 95 @torch.library.impl(lib, "split_with_sizes_copy", "Meta") 96 @torch.library.impl(lib, "split_with_sizes_copy", "CUDA") 97 @torch.library.impl(lib, "split_with_sizes_copy", "CPU") 98 def split_with_sizes_copy( function 104 torch.split_with_sizes_copy( 275 torch.ops.fsdp.split_with_sizes_copy(
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/cuda/ |
H A D | TensorShape.cu | 23 // NOTE [CUDA fast path for split_with_sizes_copy.out] 24 // split_with_sizes_copy.out for contiguous operands has the following 377 // path for split_with_sizes_copy.out]. In addition, chunk_cat_cuda supports 617 // See [CUDA fast path for split_with_sizes_copy.out]
|
/aosp_15_r20/external/pytorch/test/ |
H A D | test_functionalization.py | 876 split_with_sizes_copy = torch.ops.aten.split_with_sizes_copy.default(arg0_1, [2, 2]) 877 getitem = split_with_sizes_copy[0] 878 getitem_1 = split_with_sizes_copy[1]; split_with_sizes_copy = getitem_1 = None 881 split_with_sizes_copy_1 = torch.ops.aten.split_with_sizes_copy.default(arg0_1, [2, 2]) 886 split_with_sizes_copy_2 = torch.ops.aten.split_with_sizes_copy.default(slice_scatter, [2, 2])
|
/aosp_15_r20/external/pytorch/torch/_inductor/ |
H A D | comms.py | 516 torch.ops.fsdp.split_with_sizes_copy.default, 545 cur_snode.node, torch.ops.fsdp.split_with_sizes_copy.default
|
/aosp_15_r20/external/pytorch/test/distributed/_composable/fsdp/ |
H A D | test_fully_shard_compile.py | 206 torch.ops.fsdp.split_with_sizes_copy.default, 286 file_check = file_check.check("torch.ops.fsdp.split_with_sizes_copy.")
|
/aosp_15_r20/external/executorch/backends/arm/_passes/ |
H A D | convert_split_to_slice.py | 22 exir_ops.edge.aten.split_with_sizes_copy.default,
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/ |
H A D | ts_native_functions.yaml | 156 #- split_with_sizes_copy
|
/aosp_15_r20/external/executorch/backends/arm/operator_support/ |
H A D | tosa_supported_operators.py | 87 exir_ops.edge.aten.split_with_sizes_copy.default,
|
/aosp_15_r20/external/executorch/backends/vulkan/partitioner/ |
H A D | supported_ops.py | 122 exir_ops.edge.aten.split_with_sizes_copy.default,
|
/aosp_15_r20/external/executorch/backends/vulkan/runtime/graph/ops/impl/ |
H A D | Split.cpp | 138 aten.split_with_sizes_copy.default, split_with_sizes_copy_default);
|
/aosp_15_r20/external/executorch/backends/cadence/aot/ |
H A D | functions_hifi.yaml | 105 - op: split_with_sizes_copy.out
|
H A D | functions.yaml | 125 - op: split_with_sizes_copy.out
|
/aosp_15_r20/external/executorch/kernels/aten/ |
H A D | functions.yaml | 366 - op: split_with_sizes_copy.out
|
/aosp_15_r20/external/executorch/backends/apple/mps/operators/ |
H A D | shape_ops.py | 231 target = "aten.split_with_sizes_copy.default"
|
/aosp_15_r20/external/pytorch/test/edge/ |
H A D | selected_operators.yaml | 368 aten::split_with_sizes_copy.out:
|
/aosp_15_r20/external/executorch/exir/tests/ |
H A D | test_delegate.py | 297 node.target == exir_ops.edge.aten.split_with_sizes_copy.default
|
/aosp_15_r20/external/executorch/backends/arm/ |
H A D | tosa_quant_utils.py | 32 exir_ops.edge.aten.split_with_sizes_copy.default,
|
/aosp_15_r20/external/pytorch/torch/_decomp/ |
H A D | __init__.py | 438 aten.split_with_sizes_copy,
|
H A D | decompositions.py | 1425 [aten.split_with_sizes_copy.default, aten.split_with_sizes_copy.out] 1427 def split_with_sizes_copy( function
|
/aosp_15_r20/external/executorch/backends/vulkan/ |
H A D | op_registry.py | 495 exir_ops.edge.aten.split_with_sizes_copy.default,
|
/aosp_15_r20/external/executorch/kernels/portable/ |
H A D | functions.yaml | 830 - op: split_with_sizes_copy.out
|
/aosp_15_r20/external/pytorch/torch/distributed/tensor/_ops/ |
H A D | _tensor_ops.py | 731 aten.split_with_sizes_copy.default,
|
/aosp_15_r20/external/pytorch/aten/src/ATen/functorch/ |
H A D | LegacyBatchingRegistrations.cpp | 739 m.impl("split_with_sizes_copy", split_with_sizes_copy_batching_rule); in TORCH_LIBRARY_IMPL()
|
/aosp_15_r20/external/executorch/exir/dialects/edge/op/ |
H A D | sample_input.py | 1088 "split_with_sizes_copy.default": { # (Tensor self, SymInt[] split_sizes, int dim=0) -> Tensor[]
|
/aosp_15_r20/external/executorch/backends/vulkan/test/op_tests/ |
H A D | cases.py | 890 @register_test_suite("aten.split_with_sizes_copy.default")
|