/aosp_15_r20/external/XNNPACK/src/f32-f16-vcvt/ |
H A D | scalar-bitcast.c.in | 97 uint32_t vbias = vnonsignw + vexp_bias; variable 131 uint32_t vbias = vnonsignw + vexp_bias; variable 164 uint32_t vbias = vnonsignw + vexp_bias; variable
|
H A D | scalar-fabsf.c.in | 101 uint32_t vbias = vnonsignw + vexp_bias; variable 137 uint32_t vbias = vnonsignw + vexp_bias; variable 172 uint32_t vbias = vnonsignw + vexp_bias; variable
|
H A D | neon.c.in | 98 uint32x4_t vbias = vaddq_u32(vreinterpretq_u32_f32(vabsx), vexp_bias); variable 135 uint32x4_t vbias = vaddq_u32(vreinterpretq_u32_f32(vabsx), vexp_bias); variable
|
/aosp_15_r20/external/XNNPACK/src/f32-vmulcaddc/ |
H A D | wasmsimd.c.in | 109 const v128_t vbias = wasm_v128_load(w + ${CHANNEL_TILE}); local 136 const v128_t vbias = wasm_v128_load(w + ${CHANNEL_TILE}); local
|
/aosp_15_r20/external/XNNPACK/src/f16-vmulcaddc/gen/ |
H A D | c8-minmax-fma3-2x.c | 59 const __m256 vbias = _mm256_cvtph_ps(_mm_loadu_si128((const __m128i*) (w + 8))); in xnn_f16_vmulcaddc_minmax_ukernel_c8__fma3_2x() local 84 const __m256 vbias = _mm256_cvtph_ps(_mm_loadu_si128((const __m128i*) (w + 8))); in xnn_f16_vmulcaddc_minmax_ukernel_c8__fma3_2x() local
|
H A D | c16-minmax-fma3-2x.c | 96 const __m256 vbias = _mm256_cvtph_ps(_mm_loadu_si128((const __m128i*) (w + 16))); in xnn_f16_vmulcaddc_minmax_ukernel_c16__fma3_2x() local 121 const __m256 vbias = _mm256_cvtph_ps(_mm_loadu_si128((const __m128i*) (w + 16))); in xnn_f16_vmulcaddc_minmax_ukernel_c16__fma3_2x() local
|
/aosp_15_r20/external/XNNPACK/src/f32-vmulcaddc/gen/ |
H A D | c8-minmax-wasmrelaxedsimd-fma-2x.c | 96 const v128_t vbias = wasm_v128_load(w + 8); in xnn_f32_vmulcaddc_minmax_ukernel_c8__wasmrelaxedsimd_fma_2x() local 122 const v128_t vbias = wasm_v128_load(w + 8); in xnn_f32_vmulcaddc_minmax_ukernel_c8__wasmrelaxedsimd_fma_2x() local
|
H A D | c8-minmax-wasmsimd-x86-2x.c | 96 const v128_t vbias = wasm_v128_load(w + 8); in xnn_f32_vmulcaddc_minmax_ukernel_c8__wasmsimd_x86_2x() local 122 const v128_t vbias = wasm_v128_load(w + 8); in xnn_f32_vmulcaddc_minmax_ukernel_c8__wasmsimd_x86_2x() local
|
H A D | c8-minmax-wasmsimd-arm-2x.c | 96 const v128_t vbias = wasm_v128_load(w + 8); in xnn_f32_vmulcaddc_minmax_ukernel_c8__wasmsimd_arm_2x() local 122 const v128_t vbias = wasm_v128_load(w + 8); in xnn_f32_vmulcaddc_minmax_ukernel_c8__wasmsimd_arm_2x() local
|
H A D | c8-minmax-wasmrelaxedsimd-2x.c | 96 const v128_t vbias = wasm_v128_load(w + 8); in xnn_f32_vmulcaddc_minmax_ukernel_c8__wasmrelaxedsimd_2x() local 122 const v128_t vbias = wasm_v128_load(w + 8); in xnn_f32_vmulcaddc_minmax_ukernel_c8__wasmrelaxedsimd_2x() local
|
/aosp_15_r20/external/XNNPACK/src/f32-f16-vcvt/gen/ |
H A D | vcvt-neon-x8.c | 88 uint32x4_t vbias = vaddq_u32(vreinterpretq_u32_f32(vabsx), vexp_bias); in xnn_f32_f16_vcvt_ukernel__neon_x8() local 125 uint32x4_t vbias = vaddq_u32(vreinterpretq_u32_f32(vabsx), vexp_bias); in xnn_f32_f16_vcvt_ukernel__neon_x8() local
|
H A D | vcvt-neon-x16.c | 117 uint32x4_t vbias = vaddq_u32(vreinterpretq_u32_f32(vabsx), vexp_bias); in xnn_f32_f16_vcvt_ukernel__neon_x16() local 154 uint32x4_t vbias = vaddq_u32(vreinterpretq_u32_f32(vabsx), vexp_bias); in xnn_f32_f16_vcvt_ukernel__neon_x16() local
|
H A D | vcvt-neon-x24.c | 146 uint32x4_t vbias = vaddq_u32(vreinterpretq_u32_f32(vabsx), vexp_bias); in xnn_f32_f16_vcvt_ukernel__neon_x24() local 183 uint32x4_t vbias = vaddq_u32(vreinterpretq_u32_f32(vabsx), vexp_bias); in xnn_f32_f16_vcvt_ukernel__neon_x24() local
|
H A D | vcvt-neon-x32.c | 175 uint32x4_t vbias = vaddq_u32(vreinterpretq_u32_f32(vabsx), vexp_bias); in xnn_f32_f16_vcvt_ukernel__neon_x32() local 212 uint32x4_t vbias = vaddq_u32(vreinterpretq_u32_f32(vabsx), vexp_bias); in xnn_f32_f16_vcvt_ukernel__neon_x32() local
|
/aosp_15_r20/external/XNNPACK/src/qs8-vaddc/ |
H A D | avx2-mul32-ld64.c.in | 39 const __m256i vbias = _mm256_add_epi32( variable
|
H A D | wasmsimd.c.in | 37 v128_t vbias = wasm_i32x4_splat((int32_t) *input_b * params->wasmsimd.b_multiplier[0]); variable
|
H A D | avx512skx-mul32-ld128.c.in | 44 const __m512i vbias = _mm512_add_epi32( variable
|
/aosp_15_r20/external/XNNPACK/src/qs8-vadd/ |
H A D | avx2-mul32-ld64.c.in | 30 const __m256i vbias = _mm256_load_si256((const __m256i*) params->avx2.bias); variable
|
H A D | wasmsimd.c.in | 31 const v128_t vbias = wasm_v128_load64_splat(params->wasmsimd.bias); variable
|
H A D | avx512skx-mul32-ld128.c.in | 33 const __m512i vbias = _mm512_load_si512(params->avx512.bias); variable
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/cpu/ |
H A D | DepthwiseConvKernel.cpp | 128 const float32x4_t vbias = vsetq_lane_f32(*bias, vdupq_n_f32(0.0), 1); in convolution_depthwise3x3_winograd_impl() local 324 …const vfloat32m1_t vbias = __riscv_vfmerge_vfm_f32m1(__riscv_vfmv_v_f_f32m1(0.0, 4), *bias, mask, … in convolution_depthwise3x3_winograd_impl() local
|
/aosp_15_r20/external/XNNPACK/src/qs8-vcvt/gen/ |
H A D | vcvt-scalar-x1.c | 22 const int32_t vbias = params->scalar.bias; in xnn_qs8_vcvt_ukernel__scalar_x1() local
|
/aosp_15_r20/external/XNNPACK/src/qu8-vcvt/gen/ |
H A D | vcvt-scalar-x1.c | 22 const int32_t vbias = params->scalar.bias; in xnn_qu8_vcvt_ukernel__scalar_x1() local
|
/aosp_15_r20/external/XNNPACK/src/qs8-vlrelu/gen/ |
H A D | vlrelu-scalar-select-x1.c | 25 const int32_t vbias = params->scalar_select.bias; in xnn_qs8_vlrelu_ukernel__scalar_select_x1() local
|
/aosp_15_r20/external/XNNPACK/src/qu8-vlrelu/gen/ |
H A D | vlrelu-scalar-andxor-x1.c | 25 const int32_t vbias = params->scalar_andxor.bias; in xnn_qu8_vlrelu_ukernel__scalar_andxor_x1() local
|