Home
last modified time | relevance | path

Searched defs:num_heads (Results 1 – 25 of 28) sorted by relevance

12

/aosp_15_r20/external/executorch/backends/vulkan/test/op_tests/
H A Dsdpa_test.cpp217 const int num_heads, in test_reference_sdpa()
286 const int num_heads, in test_vulkan_sdpa()
444 const int num_heads = 6; in TEST() local
464 const int num_heads = 6; in TEST() local
483 const int num_heads = 32; in TEST() local
502 const int num_heads = 32; in TEST() local
/aosp_15_r20/external/pytorch/aten/src/ATen/native/nested/cuda/
H A DNestedTensorTransformerUtils.cpp94 const int64_t num_heads = tensor -> opt_size(2).value(); in is_safe_to_get_storage_as_tensor() local
173 const int64_t num_heads, in view_as_dense()
405 const int64_t num_heads = query.size(1); in sdpa_nested_preprocessing() local
482 const int64_t num_heads = query.size(1); in sdpa_nested_preprocessing_backward() local
/aosp_15_r20/external/pytorch/aten/src/ATen/native/transformers/cuda/flash_attn/
H A Dflash_api.cpp298 const int num_heads, const int head_size, const int max_seqlen_k, const int max_seqlen_q, in set_params_splitkv()
329 …Flash_fwd_params &params, std::optional<at::Tensor> &alibi_slopes_, int batch_size, int num_heads){ in set_params_alibi()
390 int num_heads = sizes[2]; in mha_fwd() local
604 int num_heads = sizes[1]; in mha_varlen_fwd() local
866 const int num_heads = sizes[2]; in mha_bwd() local
1088 const int num_heads = sizes[1]; in mha_varlen_bwd() local
1310 int num_heads = sizes[2]; in mha_fwd_kvcache() local
/aosp_15_r20/external/pytorch/torch/csrc/api/src/nn/options/
H A Dactivation.cpp26 int64_t num_heads) in MultiheadAttentionOptions()
42 int64_t num_heads, in MultiheadAttentionForwardFuncOptions()
/aosp_15_r20/external/pytorch/benchmarks/transformer/
H A Dsdp.py102 def __init__(self, num_heads, in_proj_weight, in_proj_bias, out_proj): argument
277 batch_sizes, num_heads, max_seq_lens, embed_dims, dtypes, pad_percentages argument
H A Dattention_bias_benchmarks.py81 def __init__(self, num_heads, embed_dim, device=None, dtype=None): argument
/aosp_15_r20/external/executorch/examples/qualcomm/oss_scripts/llama2/runner/
H A Drunner.cpp542 size_t num_heads = (method_meta_->num_inputs() - 3) / 2; in compute_total_nbytes() local
573 size_t num_heads = (method_meta_->num_inputs() - 3) / 2; in init_tensors() local
/aosp_15_r20/external/executorch/exir/tests/
H A Dtransformer.py15 def __init__(self, embed_dim, num_heads=2): argument
/aosp_15_r20/external/executorch/examples/models/llama3_2_vision/vision_encoder/
H A Dmodel.py23 num_heads: int = 16 variable in VisionEncoderConfig
/aosp_15_r20/external/pytorch/torch/csrc/api/include/torch/nn/modules/
H A Dactivation.h826 MultiheadAttentionImpl(int64_t embed_dim, int64_t num_heads) in MultiheadAttentionImpl()
/aosp_15_r20/external/pytorch/aten/src/ATen/native/transformers/
H A Dtransformer.cpp78 const int64_t num_heads, in transformer_encoder_layer_forward()
/aosp_15_r20/external/pytorch/test/
H A Dtest_native_mha.py160 def __init__(self, embed_dim, num_heads, qkv, proj): argument
H A Dtest_transformers.py3326 def rand_nt(sequence_list, num_heads, head_dim): argument
H A Dtest_jit.py14912 num_heads, # type: int argument
14992 def __init__(self, embed_dim, num_heads): argument
/aosp_15_r20/external/executorch/examples/mediatek/aot_utils/llm_utils/
H A Dutils.py244 num_heads, argument
/aosp_15_r20/external/pytorch/aten/src/ATen/native/transformers/cuda/
H A Dattention.cu752 const int64_t num_heads = query.size(1); in _scaled_dot_product_cudnn_attention_cuda() local
1024 int64_t num_heads = query.size(-2); in _efficient_attention_forward() local
H A Dattention_backward.cu195 const int64_t num_heads = query.size(1); in _scaled_dot_product_cudnn_attention_backward_cuda() local
/aosp_15_r20/external/executorch/backends/vulkan/runtime/graph/ops/impl/
H A DSDPA.cpp259 const int64_t num_heads = graph.size_at<int64_t>(2, q_projected); in sdpa_with_kv_cache_impl() local
/aosp_15_r20/external/pytorch/test/cpp/api/
H A Dsequential.cpp501 int64_t num_heads = 4; in TEST_F() local
/aosp_15_r20/external/pytorch/torch/nn/modules/
H A Dactivation.py1043 num_heads, argument
/aosp_15_r20/external/pytorch/test/cpp_extensions/
H A Dopen_registration_extension.cpp455 const int64_t num_heads = query.size(1); in custom_scaled_dot_product_fused_attention_overrideable() local
/aosp_15_r20/external/pytorch/aten/src/ATen/native/transformers/cuda/mem_eff_attention/
H A Dkernel_forward.h186 int32_t num_heads = 0; member
H A Dkernel_backward.h666 int32_t num_heads = -1; member
/aosp_15_r20/external/pytorch/test/inductor/
H A Dtest_cpu_repro.py296 num_heads, argument
3161 def __init__(self, hidden_size, num_heads): argument
/aosp_15_r20/external/bcc/libbpf-tools/arm64/
H A Dvmlinux_608.h27774 long unsigned int num_heads; member

12