Home
last modified time | relevance | path

Searched defs:ntensors (Results 1 – 21 of 21) sorted by relevance

/aosp_15_r20/external/pytorch/aten/src/ATen/native/nested/cuda/
H A DNestedTensorMatmul.cu187 int64_t ntensors) { in group_gemm_dispatch()
201 int64_t ntensors) { in group_gemm_dispatch()
227 int64_t ntensors) { in group_gemm_dispatch()
292 int64_t ntensors = self_ptr->size(0), ntensors2 = mat2_ptr->size(0); in bmm_nested_cuda() local
/aosp_15_r20/external/pytorch/aten/src/ATen/native/nested/
H A DNestedTensorMath.cpp385 int64_t ntensors = nt_input->size(0); in NestedTensor_sum_dim_CPU() local
434 int64_t ntensors = self_ptr->size(0); in select_nested() local
509 int64_t ntensors = input_ptr->size(0); in softmax_nested() local
547 int64_t ntensors = input_ptr->size(0); in NestedTensor_all() local
709 int64_t ntensors = static_cast<int64_t>(sizes.size()); in NestedTensor_compute_size_stride() local
825 int64_t ntensors = self_ptr->size(0); in view_nested() local
916 int64_t ntensors = self_ptr->size(0); in reshape_nested() local
H A DNestedTensorMatmul.cpp24 int64_t ntensors = self.is_nested() ? get_nested_tensor_impl(self)->size(0) : self.size(0); in bmm_nested() local
250 int64_t ntensors = self_ptr->size(0), ntensors2 = mat2_ptr->size(0); in matmul_nested() local
H A DNestedTensorBackward.cpp84 int64_t ntensors = output_ptr->size(0); in nested_softmax_backward() local
129 int64_t ntensors = nt_self->size(0); in _nested_sum_backward_cpu() local
H A DNestedTensorUtils.h110 int64_t ntensors = self_ptr->size(0); in NestedTensor_get_sizes() local
143 int64_t ntensors = self_ptr->size(0); in NestedTensor_get_strides() local
H A DNestedTensorUnaryOps.cpp38 int64_t ntensors = condition_ptr->size(0); in NestedTensor_where() local
H A DNestedTensorFactories.cpp179 int64_t ntensors = self_ptr->size(0); in NestedTensor_unbind() local
/aosp_15_r20/external/pytorch/aten/src/ATen/
H A DTensorIteratorInternal.h28 const auto ntensors = base.size(); in get_data_ptrs() local
43 size_t ntensors, in serial_for_each()
H A DNestedTensorImpl.h225 int64_t ntensors = nt->size(0); in nested_tensor_impl_is_contiguous() local
H A DNestedTensorImpl.cpp147 int64_t ntensors = sizes.size(0), orig_dim = sizes.size(1); in construct_offsets() local
H A DTensorIterator.cpp54 auto ntensors = operands.size(); in get_strides() local
787 const auto ntensors = this->ntensors(); in serial_for_each() local
H A DTensorIterator.h280 int ntensors() const { in ntensors() function
/aosp_15_r20/external/pytorch/aten/src/ATen/native/cuda/
H A DCUDALoops.cuh254 constexpr int ntensors = traits::arity + 1; in gpu_kernel_impl_nocast() local
289 constexpr int ntensors = traits::arity + 1; in gpu_kernel_impl() local
H A DForeachReduceOp.cu146 const size_t ntensors = tensors.size(); in foreach_tensor_max_cuda() local
435 const size_t ntensors = tensors.size(); in foreach_tensor_norm_cuda() local
H A DCUDAJitLoops.cuh189 constexpr int ntensors = arity + 1; in jitted_gpu_kernel_generic() local
H A DLoops.cuh280 constexpr int ntensors = num_outputs + num_inputs; in gpu_kernel_multiple_outputs_impl() local
/aosp_15_r20/external/pytorch/aten/src/ATen/native/cpu/
H A DIndexKernel.cpp466 static constexpr int ntensors = 2; in cpu_hflip_vec() local
537 static constexpr int ntensors = 2; in cpu_vflip_memcpy() local
673 static constexpr int ntensors = 2; in cpu_hflip_channels_last_vec() local
H A DLoops.h117 constexpr int ntensors = traits::arity + 1; in basic_loop() local
/aosp_15_r20/external/pytorch/test/cpp/c10d/example/
H A Dallreduce.cpp14 const auto ntensors = 10; in main() local
/aosp_15_r20/external/pytorch/aten/src/ATen/cuda/
H A Djiterator_impl.h132 int ntensors = iter.ntensors(); in ArrayVariant() local
/aosp_15_r20/external/pytorch/test/
H A Dtest_nestedtensor.py1610 def test_sum(device, dtype, ntensors, max_sizes, dim, keepdim=True): argument