1// Copyright 2020 Google LLC 2// 3// This source code is licensed under the BSD-style license found in the 4// LICENSE file in the root directory of this source tree. 5 6$assert DATATYPE in ["QS8", "QU8"] 7$assert SSE == 4 8$assert not XOP or AVX 9$assert BATCH_TILE % 8 == 0 10$assert BATCH_TILE >= 8 11$ABC = "0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZ" 12#include <assert.h> 13 14$if XOP: 15 #if defined(__GNUC__) || defined(__clang__) 16 #include <x86intrin.h> 17 #else 18 #include <immintrin.h> 19 #include <ammintrin.h> 20 #endif 21$else: 22 #include <immintrin.h> 23 24#include <xnnpack/intrinsics-polyfill.h> 25#include <xnnpack/unaligned.h> 26#include <xnnpack/vadd.h> 27 28 29$PARAMS_STRUCT = {"QS8": "sse4_mul32", "QU8": "sse4"}[DATATYPE] 30$XINT8_T = {"QS8": "int8_t", "QU8": "uint8_t"}[DATATYPE] 31$_MM_CVTEPX8_EPI32 = {"QS8": "_mm_cvtepi8_epi32", "QU8": "_mm_cvtepu8_epi32"}[DATATYPE] 32$_MM_PACKXS_EPI16 = {"QS8": "_mm_packs_epi16", "QU8": "_mm_packus_epi16"}[DATATYPE] 33$_MM_MIN_EPX8 = {"QS8": "_mm_min_epi8", "QU8": "_mm_min_epu8"}[DATATYPE] 34$_MM_MAX_EPX8 = {"QS8": "_mm_max_epi8", "QU8": "_mm_max_epu8"}[DATATYPE] 35$ISA = "xop" if XOP else "avx" if AVX else {4: "sse41"}[SSE] 36void xnn_${DATATYPE.lower()}_vadd_minmax_ukernel__${ISA}_mul32_ld32_x${BATCH_TILE}( 37 size_t n, 38 const ${XINT8_T}* input_a, 39 const ${XINT8_T}* input_b, 40 ${XINT8_T}* output, 41 const union xnn_${DATATYPE.lower()}_add_minmax_params params[restrict XNN_MIN_ELEMENTS(1)]) XNN_OOB_READS 42{ 43 const __m128i vbias = _mm_load_si128((const __m128i*) params->${PARAMS_STRUCT}.bias); 44 const __m128i va_multiplier = _mm_load_si128((const __m128i*) params->${PARAMS_STRUCT}.a_multiplier); 45 const __m128i vb_multiplier = _mm_load_si128((const __m128i*) params->${PARAMS_STRUCT}.b_multiplier); 46 const __m128i vshift = _mm_load_si128((const __m128i*) params->${PARAMS_STRUCT}.shift); 47 const __m128i voutput_zero_point = _mm_load_si128((const __m128i*) params->${PARAMS_STRUCT}.output_zero_point); 48 const __m128i voutput_min = _mm_load_si128((const __m128i*) params->${PARAMS_STRUCT}.output_min); 49 const __m128i voutput_max = _mm_load_si128((const __m128i*) params->${PARAMS_STRUCT}.output_max); 50 51 for (; n >= ${BATCH_TILE} * sizeof(${XINT8_T}); n -= ${BATCH_TILE} * sizeof(${XINT8_T})) { 52 const __m128i va${ABC[0:4]} = ${_MM_CVTEPX8_EPI32}(_mm_cvtsi32_si128((int) unaligned_load_s32(input_a))); 53 const __m128i vb${ABC[0:4]} = ${_MM_CVTEPX8_EPI32}(_mm_cvtsi32_si128((int) unaligned_load_s32(input_b))); 54 $for N in range(4, BATCH_TILE, 4): 55 const __m128i va${ABC[N:N+4]} = ${_MM_CVTEPX8_EPI32}(_mm_cvtsi32_si128((int) unaligned_load_s32(input_a + ${N}))); 56 const __m128i vb${ABC[N:N+4]} = ${_MM_CVTEPX8_EPI32}(_mm_cvtsi32_si128((int) unaligned_load_s32(input_b + ${N}))); 57 input_a += ${BATCH_TILE}; 58 input_b += ${BATCH_TILE}; 59 60 $if XOP: 61 $for N in range(0, BATCH_TILE, 4): 62 __m128i vacc${ABC[N:N+4]} = _mm_macc_epi32(va${ABC[N:N+4]}, va_multiplier, vbias); 63 64 $for N in range(0, BATCH_TILE, 4): 65 vacc${ABC[N:N+4]} = _mm_macc_epi32(vb${ABC[N:N+4]}, vb_multiplier, vacc${ABC[N:N+4]}); 66 $else: 67 $for N in range(0, BATCH_TILE, 4): 68 __m128i vacc${ABC[N:N+4]} = _mm_add_epi32(vbias, _mm_mullo_epi32(va${ABC[N:N+4]}, va_multiplier)); 69 70 $for N in range(0, BATCH_TILE, 4): 71 vacc${ABC[N:N+4]} = _mm_add_epi32(vacc${ABC[N:N+4]}, _mm_mullo_epi32(vb${ABC[N:N+4]}, vb_multiplier)); 72 73 $for N in range(0, BATCH_TILE, 4): 74 vacc${ABC[N:N+4]} = _mm_sra_epi32(vacc${ABC[N:N+4]}, vshift); 75 76 $for N in range(0, BATCH_TILE, 8): 77 const __m128i vout${ABC[N:N+8]} = _mm_adds_epi16(_mm_packs_epi32(vacc${ABC[N:N+4]}, vacc${ABC[N+4:N+8]}), voutput_zero_point); 78 79 $for N in range(0, BATCH_TILE, 16): 80 $if N + 8 < BATCH_TILE: 81 __m128i vout${ABC[N:N+16]} = ${_MM_PACKXS_EPI16}(vout${ABC[N:N+8]}, vout${ABC[N+8:N+16]}); 82 $else: 83 __m128i vout${ABC[N:N+8]}${ABC[N:N+8]} = ${_MM_PACKXS_EPI16}(vout${ABC[N:N+8]}, vout${ABC[N:N+8]}); 84 85 $for N in range(0, BATCH_TILE, 16): 86 $if N + 8 < BATCH_TILE: 87 vout${ABC[N:N+16]} = ${_MM_MAX_EPX8}(vout${ABC[N:N+16]}, voutput_min); 88 $else: 89 vout${ABC[N:N+8]}${ABC[N:N+8]} = ${_MM_MAX_EPX8}(vout${ABC[N:N+8]}${ABC[N:N+8]}, voutput_min); 90 91 $for N in range(0, BATCH_TILE, 16): 92 $if N + 8 < BATCH_TILE: 93 vout${ABC[N:N+16]} = ${_MM_MIN_EPX8}(vout${ABC[N:N+16]}, voutput_max); 94 $else: 95 vout${ABC[N:N+8]}${ABC[N:N+8]} = ${_MM_MIN_EPX8}(vout${ABC[N:N+8]}${ABC[N:N+8]}, voutput_max); 96 97 $if BATCH_TILE >= 16: 98 _mm_storeu_si128((__m128i*) output, vout${ABC[0:16]}); 99 $else: 100 _mm_storel_epi64((__m128i*) output, vout${ABC[0:8]}${ABC[0:8]}); 101 $for N in range(16, BATCH_TILE, 16): 102 $if N + 8 < BATCH_TILE: 103 _mm_storeu_si128((__m128i*) (output + ${N}), vout${ABC[N:N+16]}); 104 $else: 105 _mm_storel_epi64((__m128i*) (output + ${N}), vout${ABC[N:N+8]}${ABC[N:N+8]}); 106 output += ${BATCH_TILE}; 107 } 108 if XNN_UNLIKELY(n != 0) { 109 ${"do " if BATCH_TILE > 8 else ""}{ 110 const __m128i va${ABC[0:4]} = ${_MM_CVTEPX8_EPI32}(_mm_cvtsi32_si128((int) unaligned_load_s32(input_a))); 111 const __m128i vb${ABC[0:4]} = ${_MM_CVTEPX8_EPI32}(_mm_cvtsi32_si128((int) unaligned_load_s32(input_b))); 112 const __m128i va${ABC[4:8]} = ${_MM_CVTEPX8_EPI32}(_mm_cvtsi32_si128((int) unaligned_load_s32(input_a + 4))); 113 const __m128i vb${ABC[4:8]} = ${_MM_CVTEPX8_EPI32}(_mm_cvtsi32_si128((int) unaligned_load_s32(input_b + 4))); 114 $if BATCH_TILE > 8: 115 input_a += 8; 116 input_b += 8; 117 118 $if XOP: 119 __m128i vacc${ABC[0:4]} = _mm_macc_epi32(va${ABC[0:4]}, va_multiplier, vbias); 120 __m128i vacc${ABC[4:8]} = _mm_macc_epi32(va${ABC[4:8]}, va_multiplier, vbias); 121 122 vacc${ABC[0:4]} = _mm_macc_epi32(vb${ABC[0:4]}, vb_multiplier, vacc${ABC[0:4]}); 123 vacc${ABC[4:8]} = _mm_macc_epi32(vb${ABC[4:8]}, vb_multiplier, vacc${ABC[4:8]}); 124 $else: 125 __m128i vacc${ABC[0:4]} = _mm_add_epi32(vbias, _mm_mullo_epi32(va${ABC[0:4]}, va_multiplier)); 126 __m128i vacc${ABC[4:8]} = _mm_add_epi32(vbias, _mm_mullo_epi32(va${ABC[4:8]}, va_multiplier)); 127 128 vacc${ABC[0:4]} = _mm_add_epi32(vacc${ABC[0:4]}, _mm_mullo_epi32(vb${ABC[0:4]}, vb_multiplier)); 129 vacc${ABC[4:8]} = _mm_add_epi32(vacc${ABC[4:8]}, _mm_mullo_epi32(vb${ABC[4:8]}, vb_multiplier)); 130 131 vacc${ABC[0:4]} = _mm_sra_epi32(vacc${ABC[0:4]}, vshift); 132 vacc${ABC[4:8]} = _mm_sra_epi32(vacc${ABC[4:8]}, vshift); 133 134 const __m128i vout${ABC[0:8]} = _mm_adds_epi16(_mm_packs_epi32(vacc${ABC[0:4]}, vacc${ABC[4:8]}), voutput_zero_point); 135 136 __m128i vout${ABC[0:8]}${ABC[0:8]} = ${_MM_PACKXS_EPI16}(vout${ABC[0:8]}, vout${ABC[0:8]}); 137 vout${ABC[0:8]}${ABC[0:8]} = ${_MM_MAX_EPX8}(vout${ABC[0:8]}${ABC[0:8]}, voutput_min); 138 vout${ABC[0:8]}${ABC[0:8]} = ${_MM_MIN_EPX8}(vout${ABC[0:8]}${ABC[0:8]}, voutput_max); 139 140 $if BATCH_TILE > 8: 141 if XNN_LIKELY(n >= (8 * sizeof(${XINT8_T}))) { 142 _mm_storel_epi64((__m128i*) output, vout${ABC[0:8]}${ABC[0:8]}); 143 output += 8; 144 n -= 8 * sizeof(${XINT8_T}); 145 } else { 146 if (n & (4 * sizeof(${XINT8_T}))) { 147 unaligned_store_u32(output, (uint32_t) _mm_cvtsi128_si32(vout${ABC[0:8]}${ABC[0:8]})); 148 vout${ABC[0:8]}${ABC[0:8]} = _mm_srli_epi64(vout${ABC[0:8]}${ABC[0:8]}, 32); 149 output += 4; 150 } 151 if (n & (2 * sizeof(${XINT8_T}))) { 152 unaligned_store_u16(output, (uint16_t) _mm_extract_epi16(vout${ABC[0:8]}${ABC[0:8]}, 0)); 153 vout${ABC[0:8]}${ABC[0:8]} = _mm_srli_epi32(vout${ABC[0:8]}${ABC[0:8]}, 16); 154 output += 2; 155 } 156 if (n & (1 * sizeof(${XINT8_T}))) { 157 *output = (${XINT8_T}) _mm_extract_epi8(vout${ABC[0:8]}${ABC[0:8]}, 0); 158 } 159 n = 0; 160 } 161 $else: 162 if (n & (4 * sizeof(${XINT8_T}))) { 163 unaligned_store_u32(output, (uint32_t) _mm_cvtsi128_si32(vout${ABC[0:8]}${ABC[0:8]})); 164 vout${ABC[0:8]}${ABC[0:8]} = _mm_srli_epi64(vout${ABC[0:8]}${ABC[0:8]}, 32); 165 output += 4; 166 } 167 if (n & (2 * sizeof(${XINT8_T}))) { 168 unaligned_store_u16(output, (uint16_t) _mm_extract_epi16(vout${ABC[0:8]}${ABC[0:8]}, 0)); 169 vout${ABC[0:8]}${ABC[0:8]} = _mm_srli_epi32(vout${ABC[0:8]}${ABC[0:8]}, 16); 170 output += 2; 171 } 172 if (n & (1 * sizeof(${XINT8_T}))) { 173 *output = (${XINT8_T}) _mm_extract_epi8(vout${ABC[0:8]}${ABC[0:8]}, 0); 174 } 175 }${" while (n != 0);" if BATCH_TILE > 8 else ""} 176 } 177} 178