/aosp_15_r20/external/tensorflow/tensorflow/python/framework/ |
H A D | fast_tensor_util.pyx | 10 tensor_proto, np.ndarray[np.uint16_t, ndim=1] nparray): argument 22 tensor_proto, np.ndarray[np.uint16_t, ndim=1] nparray): argument 30 tensor_proto, np.ndarray[np.float32_t, ndim=1] nparray): argument 38 tensor_proto, np.ndarray[np.float64_t, ndim=1] nparray): argument 46 tensor_proto, np.ndarray[np.int32_t, ndim=1] nparray): argument 53 tensor_proto, np.ndarray[np.uint32_t, ndim=1] nparray): argument 60 tensor_proto, np.ndarray[np.int64_t, ndim=1] nparray): argument 67 tensor_proto, np.ndarray[np.uint64_t, ndim=1] nparray): argument 74 tensor_proto, np.ndarray[np.uint8_t, ndim=1] nparray): argument 82 tensor_proto, np.ndarray[np.uint16_t, ndim=1] nparray): argument [all …]
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/ |
H A D | ReduceOpsUtils.h | 114 inline DimVector make_dim_vector(OptionalIntArrayRef opt_dims, int64_t ndim) { in make_dim_vector() 170 inline Tensor review_reduce_result(const Tensor& result, int ndim, DimMask mask, bool keepdim) { in review_reduce_result() 199 int64_t ndim = self.dim(); in make_reduction() local 240 int64_t ndim = self.dim(); in make_reduction() local 405 int64_t ndim = self.dim(); in make_reduction() local 423 int64_t ndim = self.dim(); in make_reduction() local
|
H A D | Pool.h | 128 const int64_t ndim = input.ndimension(); in pool2d_shape_check() local 184 const int64_t ndim = input.ndimension(); in max_pool2d_backward_shape_check() local 214 const int64_t ndim = input.ndimension(); in avg_pool2d_backward_shape_check() local 299 const int64_t ndim = input.ndimension(); in max_pool3d_backward_shape_check() local 334 const int64_t ndim = input.ndimension(); in avg_pool3d_backward_shape_check() local
|
H A D | LinearAlgebraUtils.h | 351 const int64_t ndim = self.ndimension(); in _move_to_end() local 436 …e std::vector<int64_t> create_dim_backshift_permutation(int64_t dim0, int64_t dim1, int64_t ndim) { in create_dim_backshift_permutation() 457 int64_t ndim = permutation.size(); in create_reverse_permutation() local 584 auto ndim = input.dim(); in is_blas_compatible_column_major_order() local 605 auto ndim = input.dim(); in is_blas_compatible_row_major_order() local
|
H A D | AdaptiveMaxPooling2d.cpp | 16 int ndim = input.ndimension(); in TORCH_META_FUNC() local 58 int64_t ndim = grad_output.ndimension(); in TORCH_META_FUNC() local
|
H A D | AdaptiveAveragePooling.cpp | 30 int64_t ndim = input.dim(); in adaptive_avg_pool2d_out_cpu_template() local 66 int64_t ndim = grad_output.ndimension(); in adaptive_avg_pool2d_backward_out_cpu_template() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/mkldnn/xpu/detail/ |
H A D | Conv.cpp | 19 int64_t ndim, in conv_dst_size() 52 const int64_t ndim, in conv_src_fmt() 70 const int64_t ndim, in conv_weight_fmt() 93 const int64_t ndim, in compatible_weight_dims() 132 auto ndim = src.ndimension(); in conv_get_md() local
|
H A D | Deconv.cpp | 51 const int64_t ndim, in deconv_src_fmt() 73 const int64_t ndim, in deconv_weight_fmt() 92 int64_t ndim, in deconv_compatible_weight_dims() 129 auto ndim = src.ndimension(); in deconv_get_plain_md() local
|
/aosp_15_r20/external/pytorch/torch/_numpy/ |
H A D | _util.py | 47 def normalize_axis_index(ax, ndim, argname=None): argument 56 def normalize_axis_tuple(axis, ndim, argname=None, allow_duplicate=False): argument 120 def apply_keepdims(tensor, axis, ndim): argument
|
/aosp_15_r20/external/tensorflow/tensorflow/compiler/xla/pjrt/ |
H A D | transpose.cc | 540 int ndim = a_dims.size(); in RemoveTrivialDimensions() local 596 int ndim = a_dims.size(); in CoalesceDimensions() local 669 int ndim, absl::Span<int64_t const> tiling_spec, in ParseTilingSpecification() 693 const int ndim = a_dims_.size(); in BuildPlanNodes() local 904 int ndim = dims.size(); in Create() local 1041 int ndim = a_dims_.size(); in Initialize() local 1217 int ndim = permutation_.size(); in ChooseParallelizationStrategy() local
|
/aosp_15_r20/external/python/cpython3/Lib/test/ |
D | test_buffer.py | 264 def strides_from_shape(ndim, shape, itemsize, layout): argument 321 def getindex(ndim, ind, strides): argument 435 def verify_structure(memlen, itemsize, ndim, shape, strides, offset): argument 564 def randslice_from_shape(ndim, shape): argument 698 def genslices_ndim(ndim, shape): argument 716 def rslices_ndim(ndim, shape, iterations=5): argument 774 ndim, shape, strides, argument
|
/aosp_15_r20/external/pytorch/aten/src/ATen/ |
H A D | ExpandUtils.cpp | 21 auto ndim = dimsA > dimsB ? dimsA : dimsB; in infer_size_impl() local 66 int64_t ndim = static_cast<int64_t>(sizes.size()); in inferExpandGeometryImpl() local 153 size_t ndim = tensor_sizes.size(); in infer_dense_strides() local
|
H A D | TensorIteratorInternal.h | 29 const auto ndim = counter.size(); in get_data_ptrs() local 46 const auto ndim = shape.size(); in serial_for_each() local
|
/aosp_15_r20/external/executorch/backends/vulkan/runtime/api/containers/ |
H A D | Tensor.cpp | 25 const size_t ndim, in calculate_dim_order() 58 size_t ndim = sizes.size(); in calculate_strides() local 127 const size_t ndim = strides.size(); in unsqueeze_strides() local 144 int64_t ndim = sizes.size(); in calculate_padded_sizes() local
|
/aosp_15_r20/external/executorch/backends/vulkan/runtime/graph/ops/impl/utils/ |
H A D | TensorUtils.h | 69 T normalize(const T& nchw_dim, const int64_t ndim) { in normalize() 78 T nchw_dim_to_whcn_dim(const T& nchw_dim, const int64_t ndim) { in nchw_dim_to_whcn_dim()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/cuda/ |
H A D | ScanUtils.cuh | 220 int ndim = self.dim(); in scan_innermost_dim_with_indices() local 244 int ndim = self.dim(); in scan_dim_with_indices() local 419 int64_t ndim = self.dim(); in scan_innermost_dim() local 446 int ndim = self.dim(); in scan_dim() local
|
H A D | TensorModeKernel.cu | 153 int64_t ndim = ensure_nonempty_dim(self.dim()); in calculate_mode() local 188 int64_t ndim = ensure_nonempty_dim(self.dim()); in apply_mode() local 279 const TensorBase &self, int64_t dim, int64_t ndim) { in launch_apply_mode_kernel()
|
/aosp_15_r20/packages/modules/NeuralNetworks/common/cpu_operations/ |
D | L2Normalization.cpp | 139 int32_t ndim = getNumberOfDimensions(inputShape); in l2normFloat32() local 168 int32_t ndim = getNumberOfDimensions(inputShape); in l2normQuant8() local 184 int32_t ndim = getNumberOfDimensions(inputShape); in l2normQuant8Signed() local
|
/aosp_15_r20/external/pytorch/torch/csrc/utils/ |
H A D | tensor_numpy.cpp | 99 static std::vector<int64_t> to_aten_shape(int ndim, npy_intp* values) { in to_aten_shape() 109 int ndim = PySequence_Length(py_seq); in seq_to_aten_shape() local 234 int ndim = PyArray_NDIM(array); in tensor_from_numpy() local
|
/aosp_15_r20/external/pytorch/torch/csrc/inductor/aoti_runtime/ |
H A D | arrayref_tensor.h | 166 int64_t ndim, in is_contiguous_strides_for_shape() 280 int64_t ndim, in reinterpret_tensor_wrapper() 293 int64_t ndim, in reinterpret_tensor_wrapper()
|
/aosp_15_r20/external/python/cpython3/Objects/ |
D | memoryobject.c | 359 copy_rec(const Py_ssize_t *shape, Py_ssize_t ndim, Py_ssize_t itemsize, in copy_rec() 628 memory_alloc(int ndim) in memory_alloc() 699 int ndim) in mbuf_add_incomplete_view() 1257 copy_shape(Py_ssize_t *shape, const PyObject *seq, Py_ssize_t ndim, in copy_shape() 1296 cast_to_ND(PyMemoryViewObject *mv, const PyObject *shape, int ndim) in cast_to_ND() 1371 Py_ssize_t ndim = 1; in memoryview_cast_impl() local 2095 tolist_rec(PyMemoryViewObject *self, const char *ptr, Py_ssize_t ndim, const Py_ssize_t *shape, in tolist_rec() 2802 Py_ssize_t ndim, const Py_ssize_t *shape, in cmp_rec()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/cpu/ |
H A D | PaddingKernel.cpp | 16 int ndim; member 145 int ndim = p.ndim; in cpu_padding() local 252 int ndim = p.ndim; in cpu_padding_channels_last() local 326 int ndim = p.ndim; in cpu_padding_backward() local 414 int ndim = p.ndim; in cpu_padding_backward_channels_last() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/ |
H A D | Normalization.cpp | 74 int64_t ndim = qx.dim(); in q_batch_norm1d_impl() local 182 int64_t ndim = qx.dim(); in q_batch_norm2d_impl() local 277 int64_t ndim = qx.dim(); in q_batch_norm3d_impl() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/vulkan/api/ |
H A D | Tensor.cpp | 16 int64_t ndim = static_cast<int64_t>(sizes.size()); in calc_contiguous_strides() local 122 size_t ndim = gpu_sizes.size(); in calc_gpu_sizes() local 157 size_t ndim = gpu_sizes.size(); in create_image_extents() local
|
/aosp_15_r20/external/pytorch/torch/csrc/inductor/aoti_torch/ |
H A D | shim_common.cpp | 354 int64_t ndim, in aoti_torch__reinterpret_tensor() 370 int64_t ndim, in aoti_torch_empty_strided() 395 int64_t ndim, in aoti_torch_create_tensor_from_blob() 422 int64_t ndim, in aoti_torch_create_tensor_from_blob_v2() 1164 int64_t ndim, in aoti_torch__alloc_from_pool()
|