/aosp_15_r20/external/XNNPACK/test/ |
H A D | qs8-vaddc-minmax.cc | 114 TEST(QS8_VADDC_MINMAX__NEON_LD64_X8, b_scale) { in TEST() argument 117 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 253 TEST(QS8_VADDC_MINMAX__NEON_LD64_X16, b_scale) { in TEST() argument 256 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 392 TEST(QS8_VADDC_MINMAX__NEON_LD64_X24, b_scale) { in TEST() argument 395 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 531 TEST(QS8_VADDC_MINMAX__NEON_LD64_X32, b_scale) { in TEST() argument 534 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 670 TEST(QS8_VADDC_MINMAX__NEON_LD128_X16, b_scale) { in TEST() argument 673 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local [all …]
|
H A D | qu8-vaddc-minmax.cc | 114 TEST(QU8_VADDC_MINMAX__NEON_LD64_X8, b_scale) { in TEST() argument 117 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 253 TEST(QU8_VADDC_MINMAX__NEON_LD64_X16, b_scale) { in TEST() argument 256 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 392 TEST(QU8_VADDC_MINMAX__NEON_LD64_X32, b_scale) { in TEST() argument 395 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 531 TEST(QU8_VADDC_MINMAX__NEON_LD128_X16, b_scale) { in TEST() argument 534 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 670 TEST(QU8_VADDC_MINMAX__SSE2_MUL16_LD64_X8, b_scale) { in TEST() argument 673 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local [all …]
|
H A D | qs8-vadd-minmax.cc | 135 TEST(QS8_VADD_MINMAX__NEON_LD64_X8, b_scale) { in TEST() argument 138 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 295 TEST(QS8_VADD_MINMAX__NEON_LD64_X16, b_scale) { in TEST() argument 298 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 455 TEST(QS8_VADD_MINMAX__NEON_LD64_X24, b_scale) { in TEST() argument 458 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 615 TEST(QS8_VADD_MINMAX__NEON_LD64_X32, b_scale) { in TEST() argument 618 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 775 TEST(QS8_VADD_MINMAX__NEON_LD128_X16, b_scale) { in TEST() argument 778 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local [all …]
|
H A D | qs8-vmulc-minmax-fp32.cc | 114 TEST(QS8_VMULC_MINMAX_FP32__NEON_LD64_X8, b_scale) { in TEST() argument 117 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 253 TEST(QS8_VMULC_MINMAX_FP32__NEON_LD64_X16, b_scale) { in TEST() argument 256 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 392 TEST(QS8_VMULC_MINMAX_FP32__NEON_LD128_X16, b_scale) { in TEST() argument 395 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 531 TEST(QS8_VMULC_MINMAX_FP32__NEONV8_LD64_X8, b_scale) { in TEST() argument 534 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 670 TEST(QS8_VMULC_MINMAX_FP32__NEONV8_LD64_X16, b_scale) { in TEST() argument 673 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local [all …]
|
H A D | qu8-vmulc-minmax-fp32.cc | 114 TEST(QU8_VMULC_MINMAX_FP32__NEON_LD64_X8, b_scale) { in TEST() argument 117 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 253 TEST(QU8_VMULC_MINMAX_FP32__NEON_LD64_X16, b_scale) { in TEST() argument 256 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 392 TEST(QU8_VMULC_MINMAX_FP32__NEON_LD128_X16, b_scale) { in TEST() argument 395 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 531 TEST(QU8_VMULC_MINMAX_FP32__NEONV8_LD64_X8, b_scale) { in TEST() argument 534 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 670 TEST(QU8_VMULC_MINMAX_FP32__NEONV8_LD64_X16, b_scale) { in TEST() argument 673 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local [all …]
|
H A D | qu8-vadd-minmax.cc | 135 TEST(QU8_VADD_MINMAX__NEON_LD64_X8, b_scale) { in TEST() argument 138 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 295 TEST(QU8_VADD_MINMAX__NEON_LD64_X16, b_scale) { in TEST() argument 298 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 455 TEST(QU8_VADD_MINMAX__NEON_LD64_X32, b_scale) { in TEST() argument 458 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 615 TEST(QU8_VADD_MINMAX__NEON_LD128_X16, b_scale) { in TEST() argument 618 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 775 TEST(QU8_VADD_MINMAX__SSE2_MUL16_LD64_X8, b_scale) { in TEST() argument 778 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local [all …]
|
H A D | qu8-vmul-minmax-fp32.cc | 135 TEST(QU8_VMUL_MINMAX_FP32__NEON_LD64_X8, b_scale) { in TEST() argument 138 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 295 TEST(QU8_VMUL_MINMAX_FP32__NEON_LD64_X16, b_scale) { in TEST() argument 298 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 455 TEST(QU8_VMUL_MINMAX_FP32__NEON_LD128_X16, b_scale) { in TEST() argument 458 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 615 TEST(QU8_VMUL_MINMAX_FP32__NEONV8_LD64_X8, b_scale) { in TEST() argument 618 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 775 TEST(QU8_VMUL_MINMAX_FP32__NEONV8_LD64_X16, b_scale) { in TEST() argument 778 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local [all …]
|
H A D | qs8-vmul-minmax-fp32.cc | 135 TEST(QS8_VMUL_MINMAX_FP32__NEON_LD64_X8, b_scale) { in TEST() argument 138 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 295 TEST(QS8_VMUL_MINMAX_FP32__NEON_LD64_X16, b_scale) { in TEST() argument 298 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 455 TEST(QS8_VMUL_MINMAX_FP32__NEON_LD128_X16, b_scale) { in TEST() argument 458 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 615 TEST(QS8_VMUL_MINMAX_FP32__NEONV8_LD64_X8, b_scale) { in TEST() argument 618 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 775 TEST(QS8_VMUL_MINMAX_FP32__NEONV8_LD64_X16, b_scale) { in TEST() argument 778 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local [all …]
|
H A D | qs8-vmulc-minmax-rndnu.cc | 114 TEST(QS8_VMULC_MINMAX_RNDNU__NEON_LD64_X8, b_scale) { in TEST() argument 117 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 253 TEST(QS8_VMULC_MINMAX_RNDNU__NEON_LD64_X16, b_scale) { in TEST() argument 256 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 392 TEST(QS8_VMULC_MINMAX_RNDNU__NEON_LD128_X16, b_scale) { in TEST() argument 395 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local
|
H A D | qu8-vmulc-minmax-rndnu.cc | 114 TEST(QU8_VMULC_MINMAX_RNDNU__NEON_LD64_X8, b_scale) { in TEST() argument 117 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 253 TEST(QU8_VMULC_MINMAX_RNDNU__NEON_LD64_X16, b_scale) { in TEST() argument 256 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 392 TEST(QU8_VMULC_MINMAX_RNDNU__NEON_LD128_X16, b_scale) { in TEST() argument 395 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local
|
H A D | qs8-vmul-minmax-rndnu.cc | 135 TEST(QS8_VMUL_MINMAX_RNDNU__NEON_LD64_X8, b_scale) { in TEST() argument 138 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 295 TEST(QS8_VMUL_MINMAX_RNDNU__NEON_LD64_X16, b_scale) { in TEST() argument 298 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 455 TEST(QS8_VMUL_MINMAX_RNDNU__NEON_LD128_X16, b_scale) { in TEST() argument 458 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local
|
H A D | qu8-vmul-minmax-rndnu.cc | 135 TEST(QU8_VMUL_MINMAX_RNDNU__NEON_LD64_X8, b_scale) { in TEST() argument 138 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 295 TEST(QU8_VMUL_MINMAX_RNDNU__NEON_LD64_X16, b_scale) { in TEST() argument 298 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local 455 TEST(QU8_VMUL_MINMAX_RNDNU__NEON_LD128_X16, b_scale) { in TEST() argument 458 for (float b_scale = 0.1f; b_scale <= 10.0f; b_scale *= 3.14f) { in TEST() local
|
H A D | vaddc-microkernel-tester.h | 69 inline VAddCMicrokernelTester& b_scale(float b_scale) { in b_scale() function 76 inline float b_scale() const { in b_scale() function
|
H A D | vadd-microkernel-tester.h | 78 inline VAddMicrokernelTester& b_scale(float b_scale) { in b_scale() function 85 inline float b_scale() const { in b_scale() function
|
H A D | vmulc-microkernel-tester.h | 66 inline VMulCMicrokernelTester& b_scale(float b_scale) { in b_scale() function 73 inline float b_scale() const { in b_scale() function
|
H A D | vmul-microkernel-tester.h | 75 inline VMulMicrokernelTester& b_scale(float b_scale) { in b_scale() function 82 inline float b_scale() const { in b_scale() function
|
/aosp_15_r20/external/executorch/kernels/quantized/cpu/ |
H A D | op_add.cpp | 54 float b_scale, in add_tensors() 128 float b_scale = static_cast<float>(b_scale_d); in quantized_add_out() local 173 double b_scale, in quantized_add_out()
|
/aosp_15_r20/external/executorch/kernels/quantized/test/ |
H A D | op_add_test.cpp | 108 double b_scale = 0.25; in TEST() local 180 double b_scale = 0.25; in TEST() local 360 double b_scale = 1; in TEST() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/qnnpack/test/ |
H A D | q8vadd.cc | 82 TEST(Q8VADD__SSE2, b_scale) { in TEST() argument 223 TEST(Q8VADD__NEON, b_scale) { in TEST() argument
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/ |
H A D | BinaryOps.cpp | 159 const auto b_scale = qb_contig.q_scale(); in qnnpack_add() local 307 const auto b_scale = qb_contig.q_scale(); in xnnp_add() local
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/quantized/cpu/qnnpack/src/ |
H A D | add.c | 26 float b_scale, in pytorch_qnnp_create_add_nc_q8()
|
/aosp_15_r20/external/pytorch/aten/src/ATen/test/ |
H A D | vulkan_quantized_api_test.cpp | 1511 const double b_scale = 0.1; in test_conv2d() local 1619 const double b_scale = 0.1; in TEST_F() local 1722 const double b_scale = 0.1; in TEST_F() local 2492 double b_scale = 0.19, in test_quantized_conv2d()
|
/aosp_15_r20/external/pytorch/test/ |
H A D | test_fx_passes.py | 777 def pattern(a, a_dtype, a_scale, b, b_dtype, b_scale, out_scale): argument
|
/aosp_15_r20/external/ComputeLibrary/src/cpu/kernels/add/generic/neon/ |
H A D | impl.cpp | 226 const auto b_scale = is_broadcast_input_1 ? in1_scale : in0_scale; in add_sub_q8_neon_fixedpoint() local
|
/aosp_15_r20/external/libvpx/third_party/libyuv/source/ |
H A D | row_common.cc | 855 const uint32_t b_scale = REPEAT8(value & 0xff); in ARGBShadeRow_C() local 890 const uint32_t b_scale = src_argb1[0]; in ARGBMultiplyRow_C() local
|