/aosp_15_r20/external/tensorflow/tensorflow/core/kernels/ |
H A D | scan_ops.h | 62 auto logsumexp = add(log1p(exp(sub(mi, ma))), ma); in operator() local 76 auto logsumexp = padd(plog1p(pexp(psub(mi, ma))), ma); in packetOp() local 85 LogSumExp<T> logsumexp; in reduce() local 92 LogSumExp<T> logsumexp; in reducePacket() local 136 auto logsumexp = add(log(sum_reducer.finalizeBoth( in finalizeBoth() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/transformers/cuda/ |
H A D | attention_backward.cu | 67 const Tensor& logsumexp, in _flash_attention_backward() 173 const Tensor& logsumexp, in _scaled_dot_product_cudnn_attention_backward_cuda() 265 const at::Tensor& logsumexp, in _efficient_attention_backward() 738 const at::Tensor& logsumexp, in _scaled_dot_product_flash_attention_backward_cuda() 791 const at::Tensor& logsumexp, in _scaled_dot_product_efficient_attention_backward_cuda()
|
H A D | attention.cu | 878 Tensor output, q_padded, k_padded, v_padded, logsumexp, output_shape, in _flash_attention_forward() local 1029 at::Tensor logsumexp; in _efficient_attention_forward() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/cpu/ |
H A D | FlashAttentionKernel.cpp | 176 const Tensor& logsumexp, in cpu_flash_attention() 429 const at::Tensor& logsumexp, in cpu_flash_attention_backward() 731 const Tensor& logsumexp, in flash_attention_kernel_impl() 785 const at::Tensor& logsumexp, in flash_attention_backward_kernel_impl()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/transformers/ |
H A D | attention.cpp | 803 at::Tensor logsumexp = at::empty({batchSize, qSize, num_head}, in _scaled_dot_product_flash_attention_cpu() local 822 const Tensor& logsumexp, in _scaled_dot_product_flash_attention_cpu_backward()
|
/aosp_15_r20/external/pytorch/test/cpp_extensions/ |
H A D | open_registration_extension.cpp | 463 auto logsumexp = at::empty({batch_size, num_heads, max_seqlen_q}, opts.dtype(at::kFloat)); in custom_scaled_dot_product_fused_attention_overrideable() local 480 const at::Tensor & logsumexp, in custom_scaled_dot_product_fused_attention_overrideable_backward()
|
/aosp_15_r20/external/pytorch/torch/special/ |
H A D | __init__.py | 310 logsumexp = _add_docstr(_special.special_logsumexp, variable
|
/aosp_15_r20/external/tensorflow/tensorflow/core/util/ctc/ |
H A D | ctc_beam_search.h | 267 T logsumexp = T(0.0); in Step() local
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/kernels/ctc/ |
H A D | ctc_beam_search.h | 262 float logsumexp = 0.0; in Step() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/nested/cuda/ |
H A D | NestedTensorTransformerFunctions.cpp | 332 const at::Tensor& logsumexp, in _scaled_dot_product_flash_attention_backward_nested()
|
/aosp_15_r20/external/pytorch/torch/masked/ |
H A D | _ops.py | 1460 def logsumexp( function
|
/aosp_15_r20/external/pytorch/torch/utils/ |
H A D | flop_counter.py | 475 logsumexp, argument
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/ |
H A D | ReduceOps.cpp | 1492 Tensor logsumexp(const Tensor& self, IntArrayRef dims, bool keepdim) { in logsumexp() function 1505 Tensor logsumexp(const Tensor& self, DimnameList dims, bool keepdim) { in logsumexp() function
|
/aosp_15_r20/external/pytorch/torch/csrc/api/include/torch/ |
H A D | special.h | 218 inline Tensor logsumexp(const Tensor& self, IntArrayRef dims, bool keepdim) { in logsumexp() function
|
/aosp_15_r20/external/tensorflow/tensorflow/python/keras/ |
H A D | backend.py | 2610 def logsumexp(x, axis=None, keepdims=False): function
|
/aosp_15_r20/external/pytorch/torch/_refs/ |
H A D | __init__.py | 800 def logsumexp( function
|
/aosp_15_r20/external/pytorch/torch/onnx/ |
H A D | symbolic_opset9.py | 5197 def logsumexp(g: jit_utils.GraphContext, input, dim, keepdim): function
|