Home
last modified time | relevance | path

Searched defs:logsumexp (Results 1 – 17 of 17) sorted by relevance

/aosp_15_r20/external/tensorflow/tensorflow/core/kernels/
H A Dscan_ops.h62 auto logsumexp = add(log1p(exp(sub(mi, ma))), ma); in operator() local
76 auto logsumexp = padd(plog1p(pexp(psub(mi, ma))), ma); in packetOp() local
85 LogSumExp<T> logsumexp; in reduce() local
92 LogSumExp<T> logsumexp; in reducePacket() local
136 auto logsumexp = add(log(sum_reducer.finalizeBoth( in finalizeBoth() local
/aosp_15_r20/external/pytorch/aten/src/ATen/native/transformers/cuda/
H A Dattention_backward.cu67 const Tensor& logsumexp, in _flash_attention_backward()
173 const Tensor& logsumexp, in _scaled_dot_product_cudnn_attention_backward_cuda()
265 const at::Tensor& logsumexp, in _efficient_attention_backward()
738 const at::Tensor& logsumexp, in _scaled_dot_product_flash_attention_backward_cuda()
791 const at::Tensor& logsumexp, in _scaled_dot_product_efficient_attention_backward_cuda()
H A Dattention.cu878 Tensor output, q_padded, k_padded, v_padded, logsumexp, output_shape, in _flash_attention_forward() local
1029 at::Tensor logsumexp; in _efficient_attention_forward() local
/aosp_15_r20/external/pytorch/aten/src/ATen/native/cpu/
H A DFlashAttentionKernel.cpp176 const Tensor& logsumexp, in cpu_flash_attention()
429 const at::Tensor& logsumexp, in cpu_flash_attention_backward()
731 const Tensor& logsumexp, in flash_attention_kernel_impl()
785 const at::Tensor& logsumexp, in flash_attention_backward_kernel_impl()
/aosp_15_r20/external/pytorch/aten/src/ATen/native/transformers/
H A Dattention.cpp803 at::Tensor logsumexp = at::empty({batchSize, qSize, num_head}, in _scaled_dot_product_flash_attention_cpu() local
822 const Tensor& logsumexp, in _scaled_dot_product_flash_attention_cpu_backward()
/aosp_15_r20/external/pytorch/test/cpp_extensions/
H A Dopen_registration_extension.cpp463 auto logsumexp = at::empty({batch_size, num_heads, max_seqlen_q}, opts.dtype(at::kFloat)); in custom_scaled_dot_product_fused_attention_overrideable() local
480 const at::Tensor & logsumexp, in custom_scaled_dot_product_fused_attention_overrideable_backward()
/aosp_15_r20/external/pytorch/torch/special/
H A D__init__.py310 logsumexp = _add_docstr(_special.special_logsumexp, variable
/aosp_15_r20/external/tensorflow/tensorflow/core/util/ctc/
H A Dctc_beam_search.h267 T logsumexp = T(0.0); in Step() local
/aosp_15_r20/external/tensorflow/tensorflow/lite/kernels/ctc/
H A Dctc_beam_search.h262 float logsumexp = 0.0; in Step() local
/aosp_15_r20/external/pytorch/aten/src/ATen/native/nested/cuda/
H A DNestedTensorTransformerFunctions.cpp332 const at::Tensor& logsumexp, in _scaled_dot_product_flash_attention_backward_nested()
/aosp_15_r20/external/pytorch/torch/masked/
H A D_ops.py1460 def logsumexp( function
/aosp_15_r20/external/pytorch/torch/utils/
H A Dflop_counter.py475 logsumexp, argument
/aosp_15_r20/external/pytorch/aten/src/ATen/native/
H A DReduceOps.cpp1492 Tensor logsumexp(const Tensor& self, IntArrayRef dims, bool keepdim) { in logsumexp() function
1505 Tensor logsumexp(const Tensor& self, DimnameList dims, bool keepdim) { in logsumexp() function
/aosp_15_r20/external/pytorch/torch/csrc/api/include/torch/
H A Dspecial.h218 inline Tensor logsumexp(const Tensor& self, IntArrayRef dims, bool keepdim) { in logsumexp() function
/aosp_15_r20/external/tensorflow/tensorflow/python/keras/
H A Dbackend.py2610 def logsumexp(x, axis=None, keepdims=False): function
/aosp_15_r20/external/pytorch/torch/_refs/
H A D__init__.py800 def logsumexp( function
/aosp_15_r20/external/pytorch/torch/onnx/
H A Dsymbolic_opset9.py5197 def logsumexp(g: jit_utils.GraphContext, input, dim, keepdim): function