/aosp_15_r20/external/XNNPACK/src/qu8-vlrelu/gen/ |
H A D | vlrelu-neon-x16.c | 39 const uint16x8_t vmask0 = vcltq_s16(vacc0, vmovq_n_s16(0)); in xnn_qu8_vlrelu_ukernel__neon_x16() local
|
H A D | vlrelu-neon-x32.c | 42 const uint16x8_t vmask0 = vcltq_s16(vacc0, vmovq_n_s16(0)); in xnn_qu8_vlrelu_ukernel__neon_x32() local
|
/aosp_15_r20/external/XNNPACK/src/qs8-vlrelu/gen/ |
H A D | vlrelu-neon-x16.c | 39 const uint16x8_t vmask0 = vcltq_s16(vacc0, vmovq_n_s16(0)); in xnn_qs8_vlrelu_ukernel__neon_x16() local
|
H A D | vlrelu-neon-x32.c | 42 const uint16x8_t vmask0 = vcltq_s16(vacc0, vmovq_n_s16(0)); in xnn_qs8_vlrelu_ukernel__neon_x32() local
|
/aosp_15_r20/external/XNNPACK/src/f32-gemm/gen/ |
H A D | 4x2c4-wasmrelaxedsimd-fma.c | 109 const v128_t vmask0 = wasm_f32x4_eq(vb0, vzero); in xnn_f32_gemm_ukernel_4x2c4__wasmrelaxedsimd_fma() local
|
H A D | 4x2c4-wasmsimd.c | 109 const v128_t vmask0 = wasm_f32x4_eq(vb0, vzero); in xnn_f32_gemm_ukernel_4x2c4__wasmsimd() local
|
H A D | 4x2c4-relu-wasmrelaxedsimd-fma.c | 109 const v128_t vmask0 = wasm_f32x4_eq(vb0, vzero); in xnn_f32_gemm_relu_ukernel_4x2c4__wasmrelaxedsimd_fma() local
|
H A D | 4x2c4-minmax-wasmsimd-arm.c | 111 const v128_t vmask0 = wasm_f32x4_eq(vb0, vzero); in xnn_f32_gemm_minmax_ukernel_4x2c4__wasmsimd_arm() local
|
H A D | 4x2c4-relu-wasmsimd.c | 109 const v128_t vmask0 = wasm_f32x4_eq(vb0, vzero); in xnn_f32_gemm_relu_ukernel_4x2c4__wasmsimd() local
|
H A D | 4x2c4-minmax-wasmrelaxedsimd-fma.c | 111 const v128_t vmask0 = wasm_f32x4_eq(vb0, vzero); in xnn_f32_gemm_minmax_ukernel_4x2c4__wasmrelaxedsimd_fma() local
|
H A D | 4x2c4-minmax-wasmrelaxedsimd.c | 111 const v128_t vmask0 = wasm_f32x4_eq(vb0, vzero); in xnn_f32_gemm_minmax_ukernel_4x2c4__wasmrelaxedsimd() local
|
H A D | 4x2c4-minmax-wasmsimd-x86.c | 111 const v128_t vmask0 = wasm_f32x4_eq(vb0, vzero); in xnn_f32_gemm_minmax_ukernel_4x2c4__wasmsimd_x86() local
|
H A D | 4x2c4-minmax-sse.c | 108 const __m128 vmask0 = _mm_cmpeq_ps(_mm_setzero_ps(), vb0); in xnn_f32_gemm_minmax_ukernel_4x2c4__sse() local
|
/aosp_15_r20/external/XNNPACK/src/f32-igemm/gen/ |
H A D | 4x2c4-wasmrelaxedsimd-fma.c | 127 const v128_t vmask0 = wasm_f32x4_eq(vb0, vzero); in xnn_f32_igemm_ukernel_4x2c4__wasmrelaxedsimd_fma() local
|
H A D | 4x2c4-relu-wasmsimd.c | 127 const v128_t vmask0 = wasm_f32x4_eq(vb0, vzero); in xnn_f32_igemm_relu_ukernel_4x2c4__wasmsimd() local
|
H A D | 4x2c4-wasmsimd.c | 127 const v128_t vmask0 = wasm_f32x4_eq(vb0, vzero); in xnn_f32_igemm_ukernel_4x2c4__wasmsimd() local
|
H A D | 4x2c4-relu-wasmrelaxedsimd-fma.c | 127 const v128_t vmask0 = wasm_f32x4_eq(vb0, vzero); in xnn_f32_igemm_relu_ukernel_4x2c4__wasmrelaxedsimd_fma() local
|
H A D | 4x2c4-minmax-sse.c | 126 const __m128 vmask0 = _mm_cmpeq_ps(_mm_setzero_ps(), vb0); in xnn_f32_igemm_minmax_ukernel_4x2c4__sse() local
|
H A D | 4x2c4-minmax-wasmsimd-x86.c | 129 const v128_t vmask0 = wasm_f32x4_eq(vb0, vzero); in xnn_f32_igemm_minmax_ukernel_4x2c4__wasmsimd_x86() local
|
H A D | 4x2c4-minmax-wasmrelaxedsimd-fma.c | 129 const v128_t vmask0 = wasm_f32x4_eq(vb0, vzero); in xnn_f32_igemm_minmax_ukernel_4x2c4__wasmrelaxedsimd_fma() local
|
H A D | 4x2c4-minmax-wasmrelaxedsimd.c | 129 const v128_t vmask0 = wasm_f32x4_eq(vb0, vzero); in xnn_f32_igemm_minmax_ukernel_4x2c4__wasmrelaxedsimd() local
|
H A D | 4x2c4-minmax-wasmsimd-arm.c | 129 const v128_t vmask0 = wasm_f32x4_eq(vb0, vzero); in xnn_f32_igemm_minmax_ukernel_4x2c4__wasmsimd_arm() local
|
/aosp_15_r20/external/XNNPACK/src/f16-f32-vcvt/gen/ |
H A D | vcvt-avx-int16-x16.c | 63 const __m128i vmask0 = _mm_cmpgt_epi16(vnonsign0, vdenorm_cutoff); in xnn_f16_f32_vcvt_ukernel__avx_int16_x16() local
|
H A D | vcvt-sse41-int16-x16.c | 63 const __m128i vmask0 = _mm_cmpgt_epi16(vnonsign0, vdenorm_cutoff); in xnn_f16_f32_vcvt_ukernel__sse41_int16_x16() local
|
H A D | vcvt-neon-int16-x16.c | 59 const uint16x8_t vmask0 = vcgtq_u16(vnonsign0, vdenorm_cutoff); in xnn_f16_f32_vcvt_ukernel__neon_int16_x16() local
|