/external/pytorch/torch/utils/benchmark/utils/ |
D | sparse_fuzzer.py | 59 def sparse_tensor_constructor(size, dtype, sparse_dim, nnz, is_coalesced): argument 62 …Note that when `is_coalesced` is False, the number of elements is doubled but the number of indices 67 …In the other hand when `is_coalesced` is True the number of elements is reduced in the coalescing … 84 if not is_coalesced: 89 if is_coalesced: 99 is_coalesced = params['coalesced'] 102 tensor = self.sparse_tensor_constructor(size, self._dtype, sparse_dim, nnz, is_coalesced) 113 "is_coalesced": tensor.is_coalesced(),
|
/external/pytorch/torch/utils/benchmark/examples/sparse/ |
D | op_benchmark.py | 55 is_coalesced = float_tensor_params[name]["is_coalesced"] 56 is_coalesced_str = "True" if is_coalesced else "False" 69 for name, shape, sparse_dim, is_coalesced in descriptions: 73 is_coalesced_len = max(is_coalesced_len, len(is_coalesced)) 80 print(f" is_coalesced\n{'-' * 100}") 85 for t_str, (name, shape, sparse_dim, is_coalesced) in zip(time_str, descriptions): 89 print(f"{t_str} {name} {shape}| {sparse_dim} | {is_coalesced}")
|
D | fuzzer.py | 70 f"{('True' if x.is_coalesced() else 'False'):>9} | "
|
/external/pytorch/aten/src/ATen/native/sparse/ |
D | SparseTensor.cpp | 94 …TORCH_CHECK(false, "is_coalesced expected sparse coordinate tensor layout but got ", self.layout()… in is_coalesced_default() 121 self.is_coalesced(), in indices_sparse() 132 self.is_coalesced(), in values_sparse() 195 std::optional<bool> is_coalesced) { in new_with_dims_and_tensor_sparse_symint() argument 218 if (is_coalesced.has_value()) { in new_with_dims_and_tensor_sparse_symint() 219 impl->set_coalesced(*is_coalesced); in new_with_dims_and_tensor_sparse_symint() 227 // is_coalesced state. The condition `!is_coalesced && self._nnz() < in new_with_dims_and_tensor_sparse_symint() 229 // respect to estimating the is_coalesced state. in new_with_dims_and_tensor_sparse_symint() 293 std::optional<bool> is_coalesced) { in sparse_coo_tensor() argument 366 is_coalesced); in sparse_coo_tensor() [all …]
|
D | SparseUnaryOps.cpp | 93 /*is_coalesced=*/ true); in coalesced_unary_ufunc() 108 TORCH_CHECK(self.is_coalesced(), "expected coalesced tensor for inplace operation"); in coalesced_unary_ufunc_out() 163 TORCH_CHECK(self.is_coalesced(), \ 225 if (self.is_coalesced()) { in threshold_backward_sparse() 249 if (self.is_coalesced()) { in threshold_backward_sparse_out() 281 TORCH_CHECK(self.is_coalesced(), "nan_to_num_ requires coalesced input"); in nan_to_num_sparse_()
|
D | SparseBinaryOpIntersectionCommon.h | 180 if ((x.is_coalesced() ^ y.is_coalesced())) { 181 return x.is_coalesced() 300 if (probably_coalesced.is_coalesced()) { 418 res._coalesced_(source.is_coalesced());
|
D | SparseTensorMath.cpp | 128 r._coalesced_(t.is_coalesced()); in mul_out_sparse_zerodim() 182 return r._coalesced_(t.is_coalesced()); in pow_out_sparse_scalar() 196 if (tensor.is_coalesced()) { in coalesce_() 233 const bool should_coalesce = rounding_mode.has_value() && !t.is_coalesced(); in div_out_sparse_zerodim() 251 r._coalesced_(t_tmp.is_coalesced()); in div_out_sparse_zerodim() 306 if (!result.is_coalesced()) { in floor_divide_out_sparse_zerodim() 318 if (!dividend.is_coalesced()) { in floor_divide_out_sparse_zerodim() 331 result._coalesced_(dividend_tmp.is_coalesced()); in floor_divide_out_sparse_zerodim() 442 bool coalesced = t.is_coalesced() && src.is_coalesced(); in add_out_sparse_contiguous() 743 bool sparse_is_coalesced = (sparse_.is_coalesced() || sparse_nnz == 1); in add_out_dense_sparse_cpu() [all …]
|
/external/perfetto/test/trace_processor/diff_tests/metrics/chrome/ |
D | touch_jank.py | 83 is_coalesced=0) 115 is_coalesced=0) 147 is_coalesced=0)
|
D | scroll_jank_mojo_simple_watcher.py | 83 is_coalesced=0) 115 is_coalesced=0) 153 is_coalesced=0)
|
D | chrome_scroll_helper.py | 54 is_coalesced=False,
|
D | scroll_jank_general_validation_test.sql | 50 -- NOT EXTRACT_ARG(arg_set_id, 'chrome_latency_info.is_coalesced') AND
|
/external/perfetto/test/trace_processor/diff_tests/parser/track_event/ |
D | track_event_typed_args_args.out | 7 "chrome_latency_info.is_coalesced","chrome_latency_info.is_coalesced",1,"[NULL]"
|
D | track_event_typed_args.textproto | 71 is_coalesced: true
|
/external/pytorch/benchmarks/dynamo/microbenchmarks/ |
D | operator_inp_utils.py | 76 return FuncCallWrapper("ST", list(e.shape), e.dtype, e.layout, e.is_coalesced()) 79 "ST", list(e.shape), e.dtype, e.layout, e.is_coalesced(), e._nnz() 83 def deserialize_sparse_tensor(size, dtype, layout, is_coalesced, nnz=None): argument
|
/external/pytorch/test/ |
D | test_sparse.py | 208 assert not x.is_coalesced() 296 self.assertEqual(x.coalesce()._nnz(), nnz if x.is_coalesced() else nnz // 2) 329 self.assertTrue(tc.is_coalesced()) 358 if t.is_coalesced(): 391 self.assertTrue(S2.is_coalesced()) 440 self.assertEqual(t.is_coalesced(), coalesced) 442 def func(indices, values, shape, is_coalesced): argument 443 …= torch.sparse_coo_tensor(indices, values, shape, check_invariants=True, is_coalesced=is_coalesced) 444 self.assertEqual(s.is_coalesced(), is_coalesced) 453 … "cannot set is_coalesced to true if indices correspond to uncoalesced COO tensor"): [all …]
|
/external/pytorch/torch/multiprocessing/ |
D | reductions.py | 451 is_coalesced, argument 455 return torch.sparse_coo_tensor(indices, values, shape, is_coalesced=is_coalesced) 490 sparse.is_coalesced(),
|
/external/perfetto/test/trace_processor/diff_tests/stdlib/chrome/ |
D | chrome_scroll_helper.py | 44 is_coalesced=False,
|
/external/pytorch/aten/src/ATen/native/sparse/cuda/ |
D | SparseCUDATensor.cu | 45 TORCH_INTERNAL_ASSERT(!self.is_coalesced()); in _coalesce_sparse_cuda() 46 // NOTE: Since `coalesce` is not an in-place operation when `is_coalesced` is false, in _coalesce_sparse_cuda()
|
/external/pytorch/aten/src/ATen/native/ |
D | Scalar.cpp | 21 if (self.is_coalesced()) return at::_local_scalar_dense(self._values()); in item()
|
D | NonSymbolicBC.h | 12 …llopt, std::optional<bool> pin_memory=std::nullopt, std::optional<bool> is_coalesced=std::nullopt);
|
/external/pytorch/torch/_subclasses/ |
D | meta_utils.py | 110 assert_eq(m1.is_coalesced, m2.is_coalesced()) 354 is_coalesced=t.is_coalesced() if t.is_sparse else None, 493 is_coalesced: Optional[bool] = None # is_sparse variable in MetaTensorDesc 1107 # Note [is_coalesced is dispatched] 1108 # Strangely enough, is_coalesced() is a dispatched operator, 1112 r._coalesced_(t.is_coalesced) 1122 r._coalesced_(t.is_coalesced)
|
/external/perfetto/test/ |
D | synth_common.py | 713 is_coalesced=None, argument 728 if is_coalesced is not None: 729 latency_info.is_coalesced = is_coalesced
|
/external/perfetto/protos/perfetto/trace/track_event/ |
D | chrome_latency_info.proto | 85 optional bool is_coalesced = 5; field
|
/external/pytorch/torch/sparse/ |
D | __init__.py | 620 indices=obj._indices(), is_coalesced=obj.is_coalesced() 655 is_coalesced=d["is_coalesced"],
|
/external/pytorch/torch/ |
D | _utils.py | 255 t._indices(), t._values(), t.size(), t.is_coalesced() 299 is_coalesced = None 301 indices, values, size, is_coalesced = data 303 indices, values, size, check_invariants=False, is_coalesced=is_coalesced
|