/aosp_15_r20/external/XNNPACK/src/f32-argmaxpool/ |
H A D | 9p8x-wasmsimd-c4.c | 94 const v128_t vm5 = wasm_f32x4_gt(vi5, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__wasmsimd_c4() local 187 const v128_t vm5 = wasm_f32x4_gt(vi5, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__wasmsimd_c4() local 302 const v128_t vm5 = wasm_f32x4_gt(vi5, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__wasmsimd_c4() local 359 const v128_t vm5 = wasm_f32x4_gt(vi5, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__wasmsimd_c4() local
|
H A D | 9p8x-neon-c4.c | 85 const uint32x4_t vm5 = vcgtq_f32(vi5, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() local 168 const uint32x4_t vm5 = vcgtq_f32(vi5, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() local 271 const uint32x4_t vm5 = vcgtq_f32(vi5, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() local 326 const uint32x4_t vm5 = vcgtq_f32(vi5, vmax); in xnn_f32_argmaxpool_ukernel_9p8x__neon_c4() local
|
H A D | 9p8x-sse2-c4.c | 94 const __m128i vm5 = _mm_castps_si128(_mm_cmpgt_ps(vi5, vmax)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() local 187 const __m128i vm5 = _mm_castps_si128(_mm_cmpgt_ps(vi5, vmax)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() local 302 const __m128i vm5 = _mm_castps_si128(_mm_cmpgt_ps(vi5, vmax)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() local 359 const __m128i vm5 = _mm_castps_si128(_mm_cmpgt_ps(vi5, vmax)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() local
|
H A D | 9x-sse2-c4.c | 113 const __m128i vm5 = _mm_castps_si128(_mm_cmpgt_ps(vi5, vmax)); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() local 164 const __m128i vm5 = _mm_castps_si128(_mm_cmpgt_ps(vi5, vmax)); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() local
|
H A D | 9x-wasmsimd-c4.c | 114 const v128_t vm5 = wasm_f32x4_gt(vi5, vmax); in xnn_f32_argmaxpool_ukernel_9x__wasmsimd_c4() local 165 const v128_t vm5 = wasm_f32x4_gt(vi5, vmax); in xnn_f32_argmaxpool_ukernel_9x__wasmsimd_c4() local
|
H A D | 9x-neon-c4.c | 104 const uint32x4_t vm5 = vcgtq_f32(vi5, vmax); in xnn_f32_argmaxpool_ukernel_9x__neon_c4() local 153 const uint32x4_t vm5 = vcgtq_f32(vi5, vmax); in xnn_f32_argmaxpool_ukernel_9x__neon_c4() local
|
/aosp_15_r20/external/XNNPACK/src/f16-raddstoreexpminusmax/gen/ |
H A D | neonfp16arith-rr2-p2-x48-acc2.c | 116 const uint16x8_t vm5 = vcltq_f16(vx5, vdenorm_cutoff); in xnn_f16_raddstoreexpminusmax_ukernel__neonfp16arith_rr2_p2_x48_acc2() local
|
H A D | neonfp16arith-rr2-p2-x48-acc3.c | 117 const uint16x8_t vm5 = vcltq_f16(vx5, vdenorm_cutoff); in xnn_f16_raddstoreexpminusmax_ukernel__neonfp16arith_rr2_p2_x48_acc3() local
|
H A D | neonfp16arith-rr2-p2-x48.c | 115 const uint16x8_t vm5 = vcltq_f16(vx5, vdenorm_cutoff); in xnn_f16_raddstoreexpminusmax_ukernel__neonfp16arith_rr2_p2_x48() local
|
H A D | neonfp16arith-rr2-p2-x64-acc2.c | 134 const uint16x8_t vm5 = vcltq_f16(vx5, vdenorm_cutoff); in xnn_f16_raddstoreexpminusmax_ukernel__neonfp16arith_rr2_p2_x64_acc2() local
|
H A D | neonfp16arith-rr2-p2-x64.c | 133 const uint16x8_t vm5 = vcltq_f16(vx5, vdenorm_cutoff); in xnn_f16_raddstoreexpminusmax_ukernel__neonfp16arith_rr2_p2_x64() local
|
H A D | neonfp16arith-rr2-p2-x72.c | 142 const uint16x8_t vm5 = vcltq_f16(vx5, vdenorm_cutoff); in xnn_f16_raddstoreexpminusmax_ukernel__neonfp16arith_rr2_p2_x72() local
|
H A D | neonfp16arith-rr2-p2-x64-acc4.c | 136 const uint16x8_t vm5 = vcltq_f16(vx5, vdenorm_cutoff); in xnn_f16_raddstoreexpminusmax_ukernel__neonfp16arith_rr2_p2_x64_acc4() local
|
H A D | neonfp16arith-rr2-p2-x72-acc3.c | 144 const uint16x8_t vm5 = vcltq_f16(vx5, vdenorm_cutoff); in xnn_f16_raddstoreexpminusmax_ukernel__neonfp16arith_rr2_p2_x72_acc3() local
|
H A D | neonfp16arith-rr2-p2-x80.c | 151 const uint16x8_t vm5 = vcltq_f16(vx5, vdenorm_cutoff); in xnn_f16_raddstoreexpminusmax_ukernel__neonfp16arith_rr2_p2_x80() local
|
H A D | neonfp16arith-rr2-p2-x80-acc2.c | 152 const uint16x8_t vm5 = vcltq_f16(vx5, vdenorm_cutoff); in xnn_f16_raddstoreexpminusmax_ukernel__neonfp16arith_rr2_p2_x80_acc2() local
|
H A D | neonfp16arith-rr2-p2-x80-acc5.c | 155 const uint16x8_t vm5 = vcltq_f16(vx5, vdenorm_cutoff); in xnn_f16_raddstoreexpminusmax_ukernel__neonfp16arith_rr2_p2_x80_acc5() local
|
H A D | neonfp16arith-rr2-p2-x96-acc2.c | 170 const uint16x8_t vm5 = vcltq_f16(vx5, vdenorm_cutoff); in xnn_f16_raddstoreexpminusmax_ukernel__neonfp16arith_rr2_p2_x96_acc2() local
|
/aosp_15_r20/external/XNNPACK/src/f16-vsigmoid/gen/ |
H A D | vsigmoid-neonfp16arith-rr2-p2-div-x48.c | 134 const uint16x8_t vm5 = vcltq_f16(vx5, vmovq_n_f16(0.0f)); in xnn_f16_vsigmoid_ukernel__neonfp16arith_rr2_p2_div_x48() local
|
H A D | vsigmoid-neonfp16arith-rr2-p2-div-x56.c | 147 const uint16x8_t vm5 = vcltq_f16(vx5, vmovq_n_f16(0.0f)); in xnn_f16_vsigmoid_ukernel__neonfp16arith_rr2_p2_div_x56() local
|
H A D | vsigmoid-neonfp16arith-rr2-p2-div-x64.c | 160 const uint16x8_t vm5 = vcltq_f16(vx5, vmovq_n_f16(0.0f)); in xnn_f16_vsigmoid_ukernel__neonfp16arith_rr2_p2_div_x64() local
|
H A D | vsigmoid-neonfp16arith-rr2-p2-nr1recps-x48.c | 155 const uint16x8_t vm5 = vcltq_f16(vx5, vmovq_n_f16(0.0f)); in xnn_f16_vsigmoid_ukernel__neonfp16arith_rr2_p2_nr1recps_x48() local
|
H A D | vsigmoid-neonfp16arith-rr2-p2-nr1fma-x48.c | 155 const uint16x8_t vm5 = vcltq_f16(vx5, vmovq_n_f16(0.0f)); in xnn_f16_vsigmoid_ukernel__neonfp16arith_rr2_p2_nr1fma_x48() local
|
H A D | vsigmoid-neonfp16arith-rr2-p2-nr1recps-x56.c | 171 const uint16x8_t vm5 = vcltq_f16(vx5, vmovq_n_f16(0.0f)); in xnn_f16_vsigmoid_ukernel__neonfp16arith_rr2_p2_nr1recps_x56() local
|
H A D | vsigmoid-neonfp16arith-rr2-p2-nr1fma-x56.c | 171 const uint16x8_t vm5 = vcltq_f16(vx5, vmovq_n_f16(0.0f)); in xnn_f16_vsigmoid_ukernel__neonfp16arith_rr2_p2_nr1fma_x56() local
|