/aosp_15_r20/external/XNNPACK/test/ |
H A D | qs8-vaddc-minmax.cc | 102 TEST(QS8_VADDC_MINMAX__NEON_LD64_X8, a_scale) { in TEST() argument 105 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 241 TEST(QS8_VADDC_MINMAX__NEON_LD64_X16, a_scale) { in TEST() argument 244 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 380 TEST(QS8_VADDC_MINMAX__NEON_LD64_X24, a_scale) { in TEST() argument 383 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 519 TEST(QS8_VADDC_MINMAX__NEON_LD64_X32, a_scale) { in TEST() argument 522 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 658 TEST(QS8_VADDC_MINMAX__NEON_LD128_X16, a_scale) { in TEST() argument 661 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local [all …]
|
H A D | qu8-vaddc-minmax.cc | 102 TEST(QU8_VADDC_MINMAX__NEON_LD64_X8, a_scale) { in TEST() argument 105 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 241 TEST(QU8_VADDC_MINMAX__NEON_LD64_X16, a_scale) { in TEST() argument 244 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 380 TEST(QU8_VADDC_MINMAX__NEON_LD64_X32, a_scale) { in TEST() argument 383 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 519 TEST(QU8_VADDC_MINMAX__NEON_LD128_X16, a_scale) { in TEST() argument 522 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 658 TEST(QU8_VADDC_MINMAX__SSE2_MUL16_LD64_X8, a_scale) { in TEST() argument 661 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local [all …]
|
H A D | qs8-vadd-minmax.cc | 123 TEST(QS8_VADD_MINMAX__NEON_LD64_X8, a_scale) { in TEST() argument 126 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 283 TEST(QS8_VADD_MINMAX__NEON_LD64_X16, a_scale) { in TEST() argument 286 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 443 TEST(QS8_VADD_MINMAX__NEON_LD64_X24, a_scale) { in TEST() argument 446 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 603 TEST(QS8_VADD_MINMAX__NEON_LD64_X32, a_scale) { in TEST() argument 606 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 763 TEST(QS8_VADD_MINMAX__NEON_LD128_X16, a_scale) { in TEST() argument 766 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local [all …]
|
H A D | qs8-vmulc-minmax-fp32.cc | 102 TEST(QS8_VMULC_MINMAX_FP32__NEON_LD64_X8, a_scale) { in TEST() argument 105 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 241 TEST(QS8_VMULC_MINMAX_FP32__NEON_LD64_X16, a_scale) { in TEST() argument 244 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 380 TEST(QS8_VMULC_MINMAX_FP32__NEON_LD128_X16, a_scale) { in TEST() argument 383 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 519 TEST(QS8_VMULC_MINMAX_FP32__NEONV8_LD64_X8, a_scale) { in TEST() argument 522 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 658 TEST(QS8_VMULC_MINMAX_FP32__NEONV8_LD64_X16, a_scale) { in TEST() argument 661 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local [all …]
|
H A D | qu8-vmulc-minmax-fp32.cc | 102 TEST(QU8_VMULC_MINMAX_FP32__NEON_LD64_X8, a_scale) { in TEST() argument 105 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 241 TEST(QU8_VMULC_MINMAX_FP32__NEON_LD64_X16, a_scale) { in TEST() argument 244 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 380 TEST(QU8_VMULC_MINMAX_FP32__NEON_LD128_X16, a_scale) { in TEST() argument 383 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 519 TEST(QU8_VMULC_MINMAX_FP32__NEONV8_LD64_X8, a_scale) { in TEST() argument 522 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 658 TEST(QU8_VMULC_MINMAX_FP32__NEONV8_LD64_X16, a_scale) { in TEST() argument 661 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local [all …]
|
H A D | qu8-vadd-minmax.cc | 123 TEST(QU8_VADD_MINMAX__NEON_LD64_X8, a_scale) { in TEST() argument 126 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 283 TEST(QU8_VADD_MINMAX__NEON_LD64_X16, a_scale) { in TEST() argument 286 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 443 TEST(QU8_VADD_MINMAX__NEON_LD64_X32, a_scale) { in TEST() argument 446 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 603 TEST(QU8_VADD_MINMAX__NEON_LD128_X16, a_scale) { in TEST() argument 606 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 763 TEST(QU8_VADD_MINMAX__SSE2_MUL16_LD64_X8, a_scale) { in TEST() argument 766 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local [all …]
|
H A D | qu8-vmul-minmax-fp32.cc | 123 TEST(QU8_VMUL_MINMAX_FP32__NEON_LD64_X8, a_scale) { in TEST() argument 126 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 283 TEST(QU8_VMUL_MINMAX_FP32__NEON_LD64_X16, a_scale) { in TEST() argument 286 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 443 TEST(QU8_VMUL_MINMAX_FP32__NEON_LD128_X16, a_scale) { in TEST() argument 446 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 603 TEST(QU8_VMUL_MINMAX_FP32__NEONV8_LD64_X8, a_scale) { in TEST() argument 606 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 763 TEST(QU8_VMUL_MINMAX_FP32__NEONV8_LD64_X16, a_scale) { in TEST() argument 766 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local [all …]
|
H A D | qs8-vmul-minmax-fp32.cc | 123 TEST(QS8_VMUL_MINMAX_FP32__NEON_LD64_X8, a_scale) { in TEST() argument 126 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 283 TEST(QS8_VMUL_MINMAX_FP32__NEON_LD64_X16, a_scale) { in TEST() argument 286 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 443 TEST(QS8_VMUL_MINMAX_FP32__NEON_LD128_X16, a_scale) { in TEST() argument 446 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 603 TEST(QS8_VMUL_MINMAX_FP32__NEONV8_LD64_X8, a_scale) { in TEST() argument 606 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 763 TEST(QS8_VMUL_MINMAX_FP32__NEONV8_LD64_X16, a_scale) { in TEST() argument 766 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local [all …]
|
H A D | qs8-vmulc-minmax-rndnu.cc | 102 TEST(QS8_VMULC_MINMAX_RNDNU__NEON_LD64_X8, a_scale) { in TEST() argument 105 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 241 TEST(QS8_VMULC_MINMAX_RNDNU__NEON_LD64_X16, a_scale) { in TEST() argument 244 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 380 TEST(QS8_VMULC_MINMAX_RNDNU__NEON_LD128_X16, a_scale) { in TEST() argument 383 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local
|
H A D | qu8-vmulc-minmax-rndnu.cc | 102 TEST(QU8_VMULC_MINMAX_RNDNU__NEON_LD64_X8, a_scale) { in TEST() argument 105 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 241 TEST(QU8_VMULC_MINMAX_RNDNU__NEON_LD64_X16, a_scale) { in TEST() argument 244 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 380 TEST(QU8_VMULC_MINMAX_RNDNU__NEON_LD128_X16, a_scale) { in TEST() argument 383 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local
|
H A D | qs8-vmul-minmax-rndnu.cc | 123 TEST(QS8_VMUL_MINMAX_RNDNU__NEON_LD64_X8, a_scale) { in TEST() argument 126 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 283 TEST(QS8_VMUL_MINMAX_RNDNU__NEON_LD64_X16, a_scale) { in TEST() argument 286 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 443 TEST(QS8_VMUL_MINMAX_RNDNU__NEON_LD128_X16, a_scale) { in TEST() argument 446 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local
|
H A D | qu8-vmul-minmax-rndnu.cc | 123 TEST(QU8_VMUL_MINMAX_RNDNU__NEON_LD64_X8, a_scale) { in TEST() argument 126 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 283 TEST(QU8_VMUL_MINMAX_RNDNU__NEON_LD64_X16, a_scale) { in TEST() argument 286 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local 443 TEST(QU8_VMUL_MINMAX_RNDNU__NEON_LD128_X16, a_scale) { in TEST() argument 446 for (float a_scale = 0.1f; a_scale <= 10.0f; a_scale *= 3.14f) { in TEST() local
|
H A D | vaddc-microkernel-tester.h | 49 inline VAddCMicrokernelTester& a_scale(float a_scale) { in a_scale() function 56 inline float a_scale() const { in a_scale() function
|
H A D | vadd-microkernel-tester.h | 58 inline VAddMicrokernelTester& a_scale(float a_scale) { in a_scale() argument 65 inline float a_scale() const { in a_scale() function
|
H A D | vmulc-microkernel-tester.h | 46 inline VMulCMicrokernelTester& a_scale(float a_scale) { in a_scale() argument 53 inline float a_scale() const { in a_scale() function
|
H A D | vmul-microkernel-tester.h | 55 inline VMulMicrokernelTester& a_scale(float a_scale) { in a_scale() function 62 inline float a_scale() const { in a_scale() function
|
/aosp_15_r20/external/executorch/kernels/quantized/cpu/ |
H A D | op_add.cpp | 51 float a_scale, in add_tensors() 127 float a_scale = static_cast<float>(a_scale_d); in quantized_add_out() local 168 double a_scale, in quantized_add_out()
|
/aosp_15_r20/external/executorch/kernels/quantized/test/ |
H A D | op_add_test.cpp | 105 double a_scale = 0.5; in TEST() local 177 double a_scale = 0.5; in TEST() local 357 double a_scale = 1; in TEST() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/qnnpack/test/ |
H A D | q8vadd.cc | 72 TEST(Q8VADD__SSE2, a_scale) { in TEST() argument 213 TEST(Q8VADD__NEON, a_scale) { in TEST() argument
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/ |
H A D | BinaryOps.cpp | 158 const auto a_scale = qa_contig.q_scale(); in qnnpack_add() local 306 const auto a_scale = qa_contig.q_scale(); in xnnp_add() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/transformers/cuda/mem_eff_attention/gemm/ |
H A D | mma_from_smem.h | 469 typename Base::TensorRefA a_scale, // Operand A_scale in shared memory in MmaPipelinedFromSharedMemory() 845 typename Base::TensorRefA a_scale, in MmaMultistageFromSharedMemory()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/qnnpack/src/ |
H A D | add.c | 24 float a_scale, in pytorch_qnnp_create_add_nc_q8()
|
/aosp_15_r20/external/pytorch/test/ |
H A D | test_fx_passes.py | 777 def pattern(a, a_dtype, a_scale, b, b_dtype, b_scale, out_scale): argument
|
/aosp_15_r20/external/ComputeLibrary/src/cpu/kernels/add/generic/neon/ |
H A D | impl.cpp | 230 const auto a_scale = is_broadcast_input_1 ? in0_scale : in1_scale; in add_sub_q8_neon_fixedpoint() local
|
/aosp_15_r20/frameworks/av/media/libstagefright/xmlparser/ |
H A D | MediaCodecsXmlParser.cpp | 1232 const char* a_scale = nullptr; in addLimit() local
|