/aosp_15_r20/external/pytorch/aten/src/ATen/native/sparse/ |
H A D | SparseTensor.cpp | 94 …TORCH_CHECK(false, "is_coalesced expected sparse coordinate tensor layout but got ", self.layout()… in is_coalesced_default() 121 self.is_coalesced(), in indices_sparse() 132 self.is_coalesced(), in values_sparse() 195 std::optional<bool> is_coalesced) { in new_with_dims_and_tensor_sparse_symint() argument 218 if (is_coalesced.has_value()) { in new_with_dims_and_tensor_sparse_symint() 219 impl->set_coalesced(*is_coalesced); in new_with_dims_and_tensor_sparse_symint() 227 // is_coalesced state. The condition `!is_coalesced && self._nnz() < in new_with_dims_and_tensor_sparse_symint() 229 // respect to estimating the is_coalesced state. in new_with_dims_and_tensor_sparse_symint() 293 std::optional<bool> is_coalesced) { in sparse_coo_tensor() argument 366 is_coalesced); in sparse_coo_tensor() [all …]
|
H A D | SparseUnaryOps.cpp | 93 /*is_coalesced=*/ true); in coalesced_unary_ufunc() 108 TORCH_CHECK(self.is_coalesced(), "expected coalesced tensor for inplace operation"); in coalesced_unary_ufunc_out() 163 TORCH_CHECK(self.is_coalesced(), \ 225 if (self.is_coalesced()) { in threshold_backward_sparse() 249 if (self.is_coalesced()) { in threshold_backward_sparse_out() 281 TORCH_CHECK(self.is_coalesced(), "nan_to_num_ requires coalesced input"); in nan_to_num_sparse_()
|
H A D | SparseTensorMath.cpp | 128 r._coalesced_(t.is_coalesced()); in mul_out_sparse_zerodim() 182 return r._coalesced_(t.is_coalesced()); in pow_out_sparse_scalar() 196 if (tensor.is_coalesced()) { in coalesce_() 233 const bool should_coalesce = rounding_mode.has_value() && !t.is_coalesced(); in div_out_sparse_zerodim() 251 r._coalesced_(t_tmp.is_coalesced()); in div_out_sparse_zerodim() 306 if (!result.is_coalesced()) { in floor_divide_out_sparse_zerodim() 318 if (!dividend.is_coalesced()) { in floor_divide_out_sparse_zerodim() 331 result._coalesced_(dividend_tmp.is_coalesced()); in floor_divide_out_sparse_zerodim() 442 bool coalesced = t.is_coalesced() && src.is_coalesced(); in add_out_sparse_contiguous() 743 bool sparse_is_coalesced = (sparse_.is_coalesced() || sparse_nnz == 1); in add_out_dense_sparse_cpu() [all …]
|
H A D | SparseBinaryOpIntersectionCommon.h | 180 if ((x.is_coalesced() ^ y.is_coalesced())) { 181 return x.is_coalesced() 300 if (probably_coalesced.is_coalesced()) { 418 res._coalesced_(source.is_coalesced());
|
/aosp_15_r20/external/pytorch/torch/utils/benchmark/utils/ |
H A D | sparse_fuzzer.py | 59 def sparse_tensor_constructor(size, dtype, sparse_dim, nnz, is_coalesced): argument 62 …Note that when `is_coalesced` is False, the number of elements is doubled but the number of indices 67 …In the other hand when `is_coalesced` is True the number of elements is reduced in the coalescing … 84 if not is_coalesced: 89 if is_coalesced: 99 is_coalesced = params['coalesced'] 102 tensor = self.sparse_tensor_constructor(size, self._dtype, sparse_dim, nnz, is_coalesced) 113 "is_coalesced": tensor.is_coalesced(),
|
/aosp_15_r20/external/pytorch/torch/utils/benchmark/examples/sparse/ |
H A D | op_benchmark.py | 55 is_coalesced = float_tensor_params[name]["is_coalesced"] 56 is_coalesced_str = "True" if is_coalesced else "False" 69 for name, shape, sparse_dim, is_coalesced in descriptions: 73 is_coalesced_len = max(is_coalesced_len, len(is_coalesced)) 80 print(f" is_coalesced\n{'-' * 100}") 85 for t_str, (name, shape, sparse_dim, is_coalesced) in zip(time_str, descriptions): 89 print(f"{t_str} {name} {shape}| {sparse_dim} | {is_coalesced}")
|
/aosp_15_r20/external/pytorch/test/ |
H A D | test_sparse.py | 208 assert not x.is_coalesced() 296 self.assertEqual(x.coalesce()._nnz(), nnz if x.is_coalesced() else nnz // 2) 329 self.assertTrue(tc.is_coalesced()) 358 if t.is_coalesced(): 391 self.assertTrue(S2.is_coalesced()) 440 self.assertEqual(t.is_coalesced(), coalesced) 442 def func(indices, values, shape, is_coalesced): argument 443 …= torch.sparse_coo_tensor(indices, values, shape, check_invariants=True, is_coalesced=is_coalesced) 444 self.assertEqual(s.is_coalesced(), is_coalesced) 453 … "cannot set is_coalesced to true if indices correspond to uncoalesced COO tensor"): [all …]
|
/aosp_15_r20/external/perfetto/test/trace_processor/diff_tests/metrics/chrome/ |
H A D | touch_jank.py | 83 is_coalesced=0) 115 is_coalesced=0) 147 is_coalesced=0)
|
H A D | scroll_jank_mojo_simple_watcher.py | 83 is_coalesced=0) 115 is_coalesced=0) 153 is_coalesced=0)
|
/aosp_15_r20/external/pytorch/torch/_subclasses/ |
H A D | meta_utils.py | 110 assert_eq(m1.is_coalesced, m2.is_coalesced()) 354 is_coalesced=t.is_coalesced() if t.is_sparse else None, 493 is_coalesced: Optional[bool] = None # is_sparse variable in MetaTensorDesc 1107 # Note [is_coalesced is dispatched] 1108 # Strangely enough, is_coalesced() is a dispatched operator, 1112 r._coalesced_(t.is_coalesced) 1122 r._coalesced_(t.is_coalesced)
|
H A D | fake_tensor.py | 931 is_coalesced: Optional[bool] 987 t.is_coalesced() if t.is_sparse else None, 1685 # See Note [is_coalesced is dispatched] 2275 out._coalesced_(e.is_coalesced()) 2404 torch.ops.aten.is_coalesced.default,
|
/aosp_15_r20/external/cronet/base/tracing/stdlib/chrome/scroll_jank/ |
H A D | scroll_jank_v3.sql | 76 is_coalesced BOOL 85 EXTRACT_ARG(arg_set_id, 'chrome_latency_info.is_coalesced') AS is_coalesced 97 WHERE is_coalesced = false
|
H A D | scroll_offsets.sql | 40 WHERE is_coalesced = False; 52 EXTRACT_ARG(arg_set_id, 'scroll_deltas.original_delta_y') IS NOT NULL AS is_coalesced field 93 TRUE AS is_coalesced
|
/aosp_15_r20/out/soong/.intermediates/external/perfetto/perfetto_protos_perfetto_trace_track_event_lite_gen_headers/gen/external/perfetto/protos/perfetto/trace/track_event/ |
D | chrome_latency_info.pb.h | 640 // optional bool is_coalesced = 5; in Swap() 646 bool is_coalesced() const; in Swap() 898 // optional bool is_coalesced = 5; 913 inline bool ChromeLatencyInfo::is_coalesced() const { in is_coalesced() function 914 // @@protoc_insertion_point(field_get:perfetto.protos.ChromeLatencyInfo.is_coalesced) in is_coalesced() 923 // @@protoc_insertion_point(field_set:perfetto.protos.ChromeLatencyInfo.is_coalesced) in set_is_coalesced()
|
/aosp_15_r20/prebuilts/runtime/mainline/platform/sdk/include_gen/external/perfetto/perfetto_protos_perfetto_trace_track_event_lite_gen_headers/gen/external/perfetto/protos/perfetto/trace/track_event/ |
H A D | chrome_latency_info.pb.h | 640 // optional bool is_coalesced = 5; in Swap() 646 bool is_coalesced() const; in Swap() 898 // optional bool is_coalesced = 5; 913 inline bool ChromeLatencyInfo::is_coalesced() const { in is_coalesced() function 914 // @@protoc_insertion_point(field_get:perfetto.protos.ChromeLatencyInfo.is_coalesced) in is_coalesced() 923 // @@protoc_insertion_point(field_set:perfetto.protos.ChromeLatencyInfo.is_coalesced) in set_is_coalesced()
|
/aosp_15_r20/external/pytorch/torch/multiprocessing/ |
H A D | reductions.py | 451 is_coalesced, argument 455 return torch.sparse_coo_tensor(indices, values, shape, is_coalesced=is_coalesced) 490 sparse.is_coalesced(),
|
/aosp_15_r20/external/pytorch/benchmarks/dynamo/microbenchmarks/ |
H A D | operator_inp_utils.py | 76 return FuncCallWrapper("ST", list(e.shape), e.dtype, e.layout, e.is_coalesced()) 79 "ST", list(e.shape), e.dtype, e.layout, e.is_coalesced(), e._nnz() 83 def deserialize_sparse_tensor(size, dtype, layout, is_coalesced, nnz=None): argument
|
/aosp_15_r20/external/pytorch/torch/ |
H A D | _utils.py | 255 t._indices(), t._values(), t.size(), t.is_coalesced() 299 is_coalesced = None 301 indices, values, size, is_coalesced = data 303 indices, values, size, check_invariants=False, is_coalesced=is_coalesced
|
/aosp_15_r20/external/perfetto/test/ |
H A D | synth_common.py | 710 is_coalesced=None, argument 725 if is_coalesced is not None: 726 latency_info.is_coalesced = is_coalesced
|
/aosp_15_r20/external/pytorch/torch/sparse/ |
H A D | __init__.py | 620 indices=obj._indices(), is_coalesced=obj.is_coalesced() 655 is_coalesced=d["is_coalesced"],
|
/aosp_15_r20/external/perfetto/test/trace_processor/diff_tests/parser/track_event/ |
H A D | track_event_typed_args_args.out | 7 "chrome_latency_info.is_coalesced","chrome_latency_info.is_coalesced",1,"[NULL]"
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/sparse/cuda/ |
H A D | SparseCUDATensor.cu | 45 TORCH_INTERNAL_ASSERT(!self.is_coalesced()); in _coalesce_sparse_cuda() 46 // NOTE: Since `coalesce` is not an in-place operation when `is_coalesced` is false, in _coalesce_sparse_cuda()
|
/aosp_15_r20/out/soong/.intermediates/external/perfetto/perfetto_protos_perfetto_trace_track_event_cpp_gen/gen/external/perfetto/protos/perfetto/trace/track_event/ |
D | chrome_latency_info.gen.cc | 63 case 5 /* is_coalesced */: in ParseFromArray() 116 // Field 5: is_coalesced in Serialize()
|
/aosp_15_r20/external/perfetto/src/trace_processor/perfetto_sql/stdlib/chrome/ |
H A D | chrome_scrolls.sql | 243 coalesced_into IS NOT NULL AS is_coalesced, field 410 is_coalesced BOOL, field 534 is_coalesced, field
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/ |
H A D | TensorShape.cpp | 523 // to_broadcast conserves is_coalesced property iff only the last in sparse_broadcast_to() 527 …bool is_coalesced = self.dim()==0 || (self.is_coalesced() && (max_unchanged_dim < min_broadcast_di… in sparse_broadcast_to() local 551 return at::sparse_coo_tensor(new_indices, new_values, size, self.options(), is_coalesced); in sparse_broadcast_to() 1284 …urn at::sparse_coo_tensor(new_indices, new_values, new_sizes, self.options(), self.is_coalesced()); in narrow_copy_sparse() 1500 const auto is_coalesced = self.is_coalesced() && (dims.empty() || dims[0] == 0); in permute_sparse_coo() local 1501 // TODO: apply `is_coalesced ||= new_values.size(0) < 2`. in permute_sparse_coo() 1503 sparse_ndim, dense_ndim, new_sizes, new_indices, new_values, self.options(), is_coalesced); in permute_sparse_coo() 2004 || (self.is_coalesced() && dim == 0 in index_select_sparse_cpu() 2018 if (self.is_coalesced() && dim == 0) { in index_select_sparse_cpu() 2238 /*is_sorted=*/self.is_coalesced() && dim == 0 in index_select_sparse_cpu()
|