/aosp_15_r20/external/pytorch/aten/src/ATen/native/sparse/ |
H A D | SparseTensor.cpp | 173 int64_t sparse_dim, in new_with_dims_sparse() 186 int64_t sparse_dim, in new_with_dims_and_tensor_sparse_symint() 317 int64_t sparse_dim = indices.size(0); in sparse_coo_tensor() local 388 int64_t sparse_dim = indices.size(0); in _validate_sparse_coo_tensor_args() local 507 auto sparse_dim = indices.sym_size(0).guard_int(__FILE__, __LINE__); in _sparse_coo_tensor_unsafe_symint() local 547 int64_t sparse_dim, in sparse_resize_() 556 int64_t sparse_dim, in sparse_resize_and_clear_() 640 int64_t sparse_dim = self.sparse_dim(); in _coalesce_sparse_cpu() local
|
H A D | SparseTensorMath.cpp | 443 int64_t sparse_dim = src.sparse_dim(); in add_out_sparse_contiguous() local 602 const int64_t sparse_dim = sparse.sparse_dim(); in add_dense_sparse_worker_non_hybrid_cpu() local 631 auto sparse_dim = sparse.sparse_dim(); in add_dense_sparse_worker_hybrid_cpu() local 663 auto sparse_dim = sparse.sparse_dim(); in add_dense_sparse_worker_non_coalesced_cpu() local 734 int64_t sparse_dim = sparse_.sparse_dim(); in add_out_dense_sparse_cpu() local 845 const int64_t sparse_dim = static_cast<int64_t>(res_shape.size()) - dense_dim; in intersection_binary_op_sparse_dense_out() local 866 const auto sparse_dim = s.sparse_dim(); in intersection_binary_op_sparse_dense_out() local 1105 int64_t sparse_dim = src.sparse_dim(); in mul_out_sparse_cpu() local 1653 const int64_t sparse_dim = input.sparse_dim(); in _sparse_sum() local
|
H A D | SoftMax.cpp | 34 int64_t get_nvalues(const IntArrayRef& sizes, int64_t sparse_dim) { in get_nvalues() 298 auto sparse_dim = input.sparse_dim(); in cpu_sparse_coo_softmax() local 411 auto sparse_dim = output.sparse_dim(); in cpu_sparse_coo_softmax_backward() local
|
H A D | FlattenIndicesCommon.h | 52 auto sparse_dim = indices.size(0); in _flatten_indices_impl() local
|
H A D | SparseCsrTensor.cpp | 400 constexpr int64_t sparse_dim = 2; in sparse_compressed_tensor_with_dims() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/ |
H A D | TensorConversions.cpp | 929 …sparse_check_arguments(const std::string& funcname, const Tensor& self, const int64_t sparse_dim) { in _to_sparse_check_arguments() 1191 Tensor dense_to_sparse(const Tensor& self, int64_t sparse_dim) { in dense_to_sparse() 1288 const auto sparse_dim = [&]() -> at::DimVector { in sparse_compressed_to_flipped() local 1289 auto sparse_dim = at::DimVector(self.sizes().slice(n_batches, 2)); in sparse_compressed_to_flipped() local 1860 Tensor sparse_coo_to_sparse(const Tensor& self, const int64_t sparse_dim) { in sparse_coo_to_sparse() 1867 Tensor sparse_compressed_to_sparse(const Tensor& self, const int64_t sparse_dim) { in sparse_compressed_to_sparse() 1961 Tensor to_sparse(const Tensor& self, const int64_t sparse_dim) { in to_sparse()
|
H A D | SparseTensorUtils.cpp | 38 int64_t sparse_dim = indices.size(0); in flatten_indices() local
|
H A D | TensorShape.cpp | 724 int64_t sparse_dim, in check_cat_sparse_dims() 740 int64_t sparse_dim = tensors[0].get().sparse_dim(); in cat_sparse_impl() local 1263 int64_t sparse_dim = self.sparse_dim(); in narrow_copy_sparse() local 1735 int64_t sparse_dim = self.sparse_dim(); in select_sparse() local 1914 const auto sparse_dim = self.sparse_dim(); in index_select_sparse_cpu() local 3151 int64_t sparse_dim = self.sparse_dim(); in check_t() local 3388 int64_t sparse_dim = self.sparse_dim(); in unsqueeze_sparse() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/sparse/cuda/ |
H A D | SoftMax.cu | 82 int64_t get_nvalues(const IntArrayRef& sizes, int64_t sparse_dim) { in get_nvalues() 393 auto sparse_dim = input.sparse_dim(); in cuda_sparse_coo_softmax() local 465 auto sparse_dim = output.sparse_dim(); in cuda_sparse_coo_softmax_backward() local
|
H A D | SparseCUDATensor.cu | 64 int64_t sparse_dim = self.sparse_dim(); in _coalesce_sparse_cuda() local
|
/aosp_15_r20/external/pytorch/torch/utils/benchmark/utils/ |
H A D | sparse_fuzzer.py | 59 def sparse_tensor_constructor(size, dtype, sparse_dim, nnz, is_coalesced): argument
|
/aosp_15_r20/external/pytorch/torch/utils/benchmark/examples/sparse/ |
H A D | compare.py | 46 def generate_coo_data(size, sparse_dim, nnz, dtype, device): argument
|
/aosp_15_r20/external/pytorch/aten/src/ATen/ |
H A D | SparseCsrTensorImpl.h | 73 inline int64_t sparse_dim() const noexcept { in sparse_dim() function
|
H A D | SparseCsrTensorImpl.cpp | 108 void SparseCsrTensorImpl::resize_and_clear_(int64_t sparse_dim, int64_t dense_dim, IntArrayRef size… in resize_and_clear_()
|
H A D | SparseTensorImpl.h | 60 int64_t sparse_dim() const { in sparse_dim() function
|
/aosp_15_r20/external/pytorch/torch/_subclasses/ |
H A D | meta_utils.py | 491 sparse_dim: Optional[int] = None # is_sparse, is_sparse_compressed variable in MetaTensorDesc
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/cuda/ |
H A D | Indexing.cu | 1634 const auto sparse_dim = self.sparse_dim(); in index_select_sparse_cuda() local
|
/aosp_15_r20/external/pytorch/torch/csrc/distributed/c10d/ |
H A D | ProcessGroupGloo.cpp | 1150 const auto sparse_dim = tensor.sparse_dim(); in populate_from_sparse_tensor() local
|
/aosp_15_r20/external/pytorch/test/ |
H A D | test_sparse.py | 192 def _gen_sparse(self, sparse_dim, nnz, with_size, dtype, device, coalesced): argument
|
H A D | test_autograd.py | 11181 def _test(size, sparse_dim, nnz, device): argument
|
/aosp_15_r20/external/pytorch/torch/testing/_internal/ |
H A D | common_utils.py | 3362 def genSparseTensor(self, size, sparse_dim, nnz, is_uncoalesced, device, dtype): argument
|