/aosp_15_r20/external/XNNPACK/src/f32-vscaleexpminusmax/gen/ |
H A D | avx2-p5-x8.c | 58 const __m256 vs0 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn0), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x8() local
|
H A D | avx2-p5-x16.c | 61 const __m256 vs0 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn0), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x16() local
|
H A D | avx2-p5-x24.c | 64 const __m256 vs0 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn0), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x24() local
|
/aosp_15_r20/external/XNNPACK/src/f32-raddstoreexpminusmax/gen/ |
H A D | scalar-rr2-p5-x2.c | 60 const float vs0 = uint32_as_float(float_as_uint32(vn0) << 23); in xnn_f32_raddstoreexpminusmax_ukernel__scalar_rr2_p5_x2() local
|
H A D | scalar-rr2-p5-x2-acc2.c | 61 const float vs0 = uint32_as_float(float_as_uint32(vn0) << 23); in xnn_f32_raddstoreexpminusmax_ukernel__scalar_rr2_p5_x2_acc2() local
|
H A D | scalar-rr2-lut64-p2-x2.c | 77 const float vs0 = uint32_as_float(xnn_table_exp2_k_over_64[vidx0] + ve0); in xnn_f32_raddstoreexpminusmax_ukernel__scalar_rr2_lut64_p2_x2() local
|
H A D | scalar-rr2-lut64-p2-x2-acc2.c | 78 const float vs0 = uint32_as_float(xnn_table_exp2_k_over_64[vidx0] + ve0); in xnn_f32_raddstoreexpminusmax_ukernel__scalar_rr2_lut64_p2_x2_acc2() local
|
H A D | scalar-rr2-p5-x4.c | 66 const float vs0 = uint32_as_float(float_as_uint32(vn0) << 23); in xnn_f32_raddstoreexpminusmax_ukernel__scalar_rr2_p5_x4() local
|
/aosp_15_r20/external/XNNPACK/src/f32-velu/gen/ |
H A D | velu-scalar-rr2-p6-x2.c | 52 float vs0 = uint32_as_float(float_as_uint32(vn0) << 23); in xnn_f32_velu_ukernel__scalar_rr2_p6_x2() local
|
H A D | velu-wasm-rr2-p6-x2.c | 52 float vs0 = uint32_as_float(float_as_uint32(vn0) << 23); in xnn_f32_velu_ukernel__wasm_rr2_p6_x2() local
|
H A D | velu-wasm-rr2-lut16-p3-x2.c | 60 float vs0 = uint32_as_float(xnn_table_exp2minus_k_over_16[vidx0] + ven0); in xnn_f32_velu_ukernel__wasm_rr2_lut16_p3_x2() local
|
H A D | velu-scalar-rr2-lut16-p3-x2.c | 60 float vs0 = uint32_as_float(xnn_table_exp2minus_k_over_16[vidx0] + ven0); in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x2() local
|
H A D | velu-scalar-rr2-p6-x3.c | 55 float vs0 = uint32_as_float(float_as_uint32(vn0) << 23); in xnn_f32_velu_ukernel__scalar_rr2_p6_x3() local
|
H A D | velu-avx2-rr1-p6-x16.c | 50 __m256 vs0 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn0), 23)); in xnn_f32_velu_ukernel__avx2_rr1_p6_x16() local
|
H A D | velu-avx512f-rr1-p6-x32.c | 52 __m512 vs0 = _mm512_castsi512_ps(_mm512_slli_epi32(_mm512_castps_si512(vn0), 23)); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x32() local
|
H A D | velu-wasm-rr2-p6-x3.c | 55 float vs0 = uint32_as_float(float_as_uint32(vn0) << 23); in xnn_f32_velu_ukernel__wasm_rr2_p6_x3() local
|
/aosp_15_r20/external/XNNPACK/src/f32-vsigmoid/gen/ |
H A D | vsigmoid-scalar-rr2-p5-div-x2.c | 49 const float vs0 = uint32_as_float(float_as_uint32(vn0) << 23); in xnn_f32_vsigmoid_ukernel__scalar_rr2_p5_div_x2() local
|
H A D | vsigmoid-scalar-rr2-lut64-p2-div-x2.c | 53 const float vs0 = uint32_as_float(xnn_table_exp2minus_k_over_64[vidx0] + ve0); in xnn_f32_vsigmoid_ukernel__scalar_rr2_lut64_p2_div_x2() local
|
H A D | vsigmoid-scalar-rr2-lut2048-p1-div-x2.c | 53 const float vs0 = uint32_as_float(xnn_table_exp2minus_k_over_2048[vidx0] + ve0); in xnn_f32_vsigmoid_ukernel__scalar_rr2_lut2048_p1_div_x2() local
|
H A D | vsigmoid-avx2-rr1-p5-div-x16.c | 49 const __m256 vs0 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn0), 23)); in xnn_f32_vsigmoid_ukernel__avx2_rr1_p5_div_x16() local
|
/aosp_15_r20/external/XNNPACK/src/f32-vscaleextexp/gen/ |
H A D | avx2-p5-x8.c | 92 …const __m256 vs0 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(_mm256_add_ps(ve0, vm… in xnn_f32_vscaleextexp_ukernel__avx2_p5_x8() local
|
H A D | avx2-p5-x16.c | 104 …const __m256 vs0 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(_mm256_add_ps(ve0, vm… in xnn_f32_vscaleextexp_ukernel__avx2_p5_x16() local
|
/aosp_15_r20/external/XNNPACK/src/f16-velu/gen/ |
H A D | velu-avx2-rr1-p3-x16.c | 51 __m256 vs0 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn0), 23)); in xnn_f16_velu_ukernel__avx2_rr1_p3_x16() local
|
H A D | velu-neonfp16arith-rr1-p3-x16.c | 52 float16x8_t vs0 = vreinterpretq_f16_s16(vshlq_n_s16(vreinterpretq_s16_f16(vn0), 10)); in xnn_f16_velu_ukernel__neonfp16arith_rr1_p3_x16() local
|
/aosp_15_r20/external/XNNPACK/src/f16-vsigmoid/gen/ |
H A D | vsigmoid-avx2-rr1-p2-div-x16.c | 49 const __m256 vs0 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn0), 23)); in xnn_f16_vsigmoid_ukernel__avx2_rr1_p2_div_x16() local
|