/aosp_15_r20/external/pytorch/torch/_inductor/kernel/ |
H A D | mm.py | 144 def bias_addmm(inp, mat1, mat2, *, out=None, alpha=1, beta=1): argument 159 def tuned_mm(mat1, mat2, *, layout=None): argument 279 def tuned_int_mm(mat1, mat2, *, layout=None): argument 323 def tuned_addmm(inp, mat1, mat2, *, alpha=1, beta=1, layout=None): argument 471 mat1, mat1_meta, mat2, *, out_dtype=None, layout=None argument 514 def fallback_mixed_mm(mat1, mat2, *, out): argument 531 def try_heuristic(m, n, k, choices, mat1, mat2, mat2_dtype, layout): argument 578 mat1, argument 596 def get_context(m, k, n, mat1, mat2, mat1_stride, mat2_stride): argument 639 def get_size_hints(mat1, mat2, m, n, k): argument [all …]
|
H A D | bmm.py | 107 def tuned_bmm(mat1, mat2, *, layout=None): argument 172 def tuned_baddbmm(inp, mat1, mat2, *, alpha=1, beta=1, layout=None): argument
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/mkldnn/ |
H A D | Matmul.cpp | 13 const Tensor &mat1, in mkldnn_matmul() 22 const Tensor& mat1, in use_mkldnn_bf16_matmul() 29 const Tensor& mat1, in use_mkldnn_fp16_matmul() 68 const Tensor& mat1, in use_mkldnn_bf32_matmul() 75 const Tensor& mat1, in use_mkldnn_matmul() 82 const Tensor &mat1, in mkldnn_matmul_i8i8i32() 228 const Tensor &mat1, in mkldnn_matmul() 326 inline bool checksize(const Tensor& mat1, const Tensor& mat2){ in checksize() 348 const Tensor& mat1, in use_mkldnn_bf16_matmul() 377 const Tensor& mat1, in use_mkldnn_fp16_matmul() [all …]
|
/aosp_15_r20/external/executorch/backends/vulkan/runtime/graph/ops/impl/ |
H A D | QuantizedLinear.cpp | 20 const ValueRef mat1, in check_q_8w_linear_args() 47 vTensorPtr mat1 = graph->get_tensor(args[1].refs[0]); in resize_q_8w_linear_node() local 69 const ValueRef mat1, in add_q_8w_linear_node() 138 const ValueRef mat1, in add_q_8w_linear_optimized_node() 235 const ValueRef mat1, in check_q_4w_linear_args() 272 vTensorPtr mat1 = graph->get_tensor(args[1].refs[0]); in resize_q_4w_linear_node() local 294 const ValueRef mat1, in add_q_4w_linear_node()
|
H A D | MatMul.cpp | 23 const ValueRef mat1, in check_matmul_args() 42 vTensorPtr mat1 = graph->get_tensor(args[1].refs[0]); in resize_matmul_node() local 61 const ValueRef mat1, in add_matmul_naive_buffer_node() 112 const ValueRef mat1, in add_matmul_naive_texture3d_node() 157 const ValueRef mat1, in add_matmul_optimized_node() 243 const ValueRef mat1, in add_matmul_node()
|
H A D | Linear.cpp | 24 const ValueRef mat1, in check_addmm_args() 58 vTensorPtr mat1 = graph->get_tensor(args[1].refs[0]); in resize_addmm_node() local 90 const ValueRef mat1, in add_addmm_naive_node() 140 const ValueRef mat1, in add_addmm_optimized_node() 232 const ValueRef mat1, in add_addmm_node()
|
/aosp_15_r20/external/pytorch/torch/_inductor/fx_passes/ |
H A D | decompose_mem_bound_mm.py | 42 def should_decompose_bmm(mat1, mat2) -> bool: argument 63 def should_decompose_mm(mat1, mat2) -> bool: argument 100 def repl(mat1, mat2): argument 123 def repl(mat1, mat2, mat3): argument 145 def repl(mat1, mat2): argument
|
H A D | post_grad.py | 316 def mm_plus_mm(match: Match, mat1, mat2, mat3, mat4): argument 396 def uint4x2_mixed_mm(match: Match, mat1, mat2, mat2_mm_shape, mat2_dtype): argument 419 def mixed_mm(match: Match, mat1, mat2, mat2_dtype): argument 998 def unfuse_bias_add_to_pointwise(match: Match, mat1, mat2, *, inp): argument 1041 def addmm(match, mat1, mat2, *, inp): argument 1042 def repl(inp, mat1, mat2): argument 1084 def fused_int_mm_mul(match: Match, mat1, mat2, mat3, out_dtype=None): argument
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/sparse/cuda/ |
H A D | SparseBlas.cpp | 40 const Tensor& mat1, in sparse_sampled_addmm_out_sparse_csr_cuda() 69 const Tensor& mat1, in sparse_sampled_addmm_sparse_csr_cuda() 81 const Tensor& mat1, in addmm_out_sparse_compressed_cuda() 163 const Tensor& mat1, in baddbmm_out_sparse_csr_cuda() 199 const Tensor& mat1, in bmm_out_sparse_csr_cuda()
|
H A D | SparseSemiStructuredOps.cu | 524 const Tensor& mat1, const Tensor& mat1_meta, const Tensor& mat2, in sparse_semi_structured_mad_op() 789 const Tensor& mat1, const Tensor& mat1_meta, const Tensor& mat2, in _sparse_semi_structured_mm() 798 const Tensor& input, const Tensor& mat1, const Tensor& mat1_meta, in _sparse_semi_structured_addmm()
|
H A D | SparseBlasImpl.cpp | 73 const at::sparse_csr::SparseCsrTensor& mat1, in addmm_out_legacy() 464 const at::sparse_csr::SparseCsrTensor& mat1, in block_sparse_mm() 580 const at::sparse_csr::SparseCsrTensor& mat1, in spmm() 836 const Tensor& mat1, in addmm_out_sparse_csr()
|
/aosp_15_r20/external/eigen/unsupported/test/ |
H A D | cxx11_tensor_contraction.cpp | 22 Tensor<float, 2, DataLayout> mat1(2, 3); in test_evals() local 103 Tensor<float, 3, DataLayout> mat1(2, 2, 2); in test_multidims() local 254 Tensor<float, 2, DataLayout> mat1(2, 3); in test_expr() local 273 Tensor<float, 3, DataLayout> mat1(2, 2, 2); in test_out_of_order_contraction() local 320 Tensor<float, 3, DataLayout> mat1(4, 3, 5); in test_consistency() local 469 Tensor<float, 2, DataLayout> mat1(2, 3); in test_tensor_product() local 501 TensorMap<Tensor<const float, 2, DataLayout> > mat1(in1.data(), 2, 3); in test_const_inputs() local
|
H A D | cxx11_tensor_expr.cpp | 74 TensorMap<Tensor<float, 2>> mat1(data1, 2, 3); in test_2d() local 114 Tensor<float, 3> mat1(2,3,7); in test_3d() local 168 Tensor<float, 3> mat1(2,3,7); in test_constants() local 226 Tensor<float, 3> mat1(2,3,7); in test_functors() local 256 Tensor<bool, 3> mat1(2,3,7); in test_type_casting() local 284 Tensor<float, 3> mat1(2,3,7); in test_select() local
|
H A D | cxx11_tensor_of_const_values.cpp | 20 TensorMap<Tensor<const float, 2>> mat1(data1, 2, 3); in test_assign() local 56 TensorMap<Tensor<const float, 2>> mat1(data1, 2, 3); in test_plus() local 82 TensorMap<Tensor<const float, 2>> mat1(data1, 2, 3); in test_plus_equal() local
|
H A D | cxx11_tensor_fixed_size.cpp | 133 TensorMap<TensorFixedSize<float, Sizes<2, 3> > > mat1(data1,2,3); in test_2d() local 182 TensorFixedSize<float, Sizes<2, 3, 7> > mat1; in test_3d() local 228 TensorFixedSize<float, Sizes<2, 3, 7> > mat1; in test_array() local
|
H A D | cxx11_tensor_map.cpp | 71 Tensor<int, 2> mat1(2,3); in test_2d() local 118 Tensor<int, 3> mat1(2,3,7); in test_3d() local 162 Tensor<int, 3> mat1(2,3,7); in test_from_tensor() local
|
H A D | cxx11_tensor_comparisons.cpp | 19 Tensor<float, 3> mat1(2,3,7); in test_orderings() local 49 Tensor<float, 3> mat1(2,3,7); in test_equality() local
|
H A D | cxx11_tensor_assign.cpp | 74 Tensor<int, 2> mat1(2,3); in test_2d() local 137 Tensor<int, 3> mat1(2,3,7); in test_3d() local
|
/aosp_15_r20/external/executorch/kernels/portable/cpu/util/ |
H A D | matmul_ops_util.cpp | 21 const Tensor& mat1, in check_addmm_args() 52 const Tensor& mat1, in get_bmm_out_target_size() 88 const Tensor& mat1, in get_mm_out_target_size() 98 const Tensor& mat1, in get_linear_out_target_size()
|
/aosp_15_r20/external/ruy/ruy/ |
H A D | prepacked_cache_test.cc | 59 PEMat mat1 = MakeDummyPEMat(Type::Create<std::uint8_t>(), 10, 20); in TEST() local 85 PEMat mat1 = MakeDummyPEMat(Type::Create<float>(), 10, 20); in TEST() local 111 PEMat mat1 = MakeDummyPEMat(Type::Create<std::uint8_t>(), 10, 20); in TEST() local 145 PEMat mat1 = MakeDummyPEMat(Type::Create<std::uint8_t>(), 10, 20); in TEST() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/sparse/ |
H A D | SparseBlas.cpp | 119 const Tensor& mat1, in sparse_sampled_addmm_out_sparse_csr_cpu() 155 const Tensor& mat1, in sparse_sampled_addmm_sparse_csr_cpu() 168 const Tensor& mat1, in sparse_sampled_addmm_check_inputs()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/mkl/ |
H A D | SparseBlasImpl.cpp | 294 const Tensor& mat1, in addmm_sparse_result() 359 const Tensor& mat1, in addmm_out_sparse_csr() 546 const Tensor& mat1, in add_out_sparse_csr()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/cuda/ |
H A D | SparseMM.cu | 14 … const Tensor& mat1, const Tensor& mat2, const Scalar& beta, const Scalar& alpha, Tensor& result) { in _sspaddmm_out_only_sparse_cuda() 18 … const Tensor& mat1, const Tensor& mat2, const Scalar& beta, const Scalar& alpha, Tensor& result) { in _sspaddmm_out_cuda()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/cpu/ |
H A D | SampledAddmmKernel.cpp | 18 const Tensor& mat1, in sampled_addmm_sparse_csr_kernel_impl() 82 const Tensor& mat1, in sampled_addmm_sparse_csr_kernel()
|
/aosp_15_r20/external/angle/src/common/ |
H A D | matrix_utils.h | 196 Matrix<T> compMult(const Matrix<T> &mat1) const in compMult() 212 Matrix<T> outerProduct(const Matrix<T> &mat1) const in outerProduct()
|