ref: 9261eb5c3786468e2041bcc6384d9292aaf4d5a4
parent: be5f58d6790117a1ac7056b4fffa7240efa1347d
author: Jean-Marc Valin <jmvalin@amazon.com>
date: Fri Jul 21 16:10:05 EDT 2023
Refactoring to make VNNI and SSE2 easier
--- a/dnn/vec_avx.h
+++ b/dnn/vec_avx.h
@@ -627,7 +627,33 @@
#endif
+#if defined(__AVXVNNI__) || defined(__AVX512VNNI__)
+#elif defined(__AVX2__)
+
+static inline __m256i mm256_dpbusds_epi32(__m256i src, __m256i a, __m256i b) {
+ __m256i ones, tmp;
+ ones = _mm256_set1_epi16(1);
+ tmp = _mm256_maddubs_epi16(a, b);
+ tmp = _mm256_madd_epi16(tmp, ones);
+ return _mm256_add_epi32(src, tmp);
+}
+#define _mm256_dpbusds_epi32(src, a, b) mm256_dpbusds_epi32(src, a, b)
+
+#elif defined(__SSSE3__)
+
+static inline mm256i_emu mm256_dpbusds_epi32(mm256i_emu src, mm256i_emu a, mm256i_emu b) {
+ mm256i_emu ones, tmp;
+ ones = _mm256_set1_epi16(1);
+ tmp = _mm256_maddubs_epi16(a, b);
+ tmp = _mm256_madd_epi16(tmp, ones);
+ return _mm256_add_epi32(src, tmp);
+}
+#define _mm256_dpbusds_epi32(src, a, b) mm256_dpbusds_epi32(src, a, b)
+
+#elif defined(__SSE2__)
+#endif
+
static inline void sgemv16x1(float *out, const float *weights, int rows, int cols, int col_stride, const float *x)
{
int i, j;
@@ -696,10 +722,8 @@
static inline void sparse_cgemv8x4(float *_out, const opus_int8 *w, const int *idx, const float *scale, int rows, int cols, const float *_x)
{
- __m256i ones;
int i, j;
unsigned char x[MAX_INPUTS];
- ones = _mm256_set1_epi16(1);
/*for (i=0;i<cols;i++) x[i] = 127+floor(.5+127*_x[i]);*/
vector_ps_to_epi8(x, _x, cols);
for (i=0;i<rows;i+=8)
@@ -713,38 +737,28 @@
#if 1 /* Unrolling by 4 gives some gain, comment out if it does not. */
for (;j<colblocks-3;j+=4)
{
- __m256i tmp;
__m256i vxj;
__m256i vw;
vxj = _mm256_set1_epi32(*(int*)&x[*idx++]);
vw = _mm256_loadu_si256((const __m256i *)w);
- tmp = _mm256_maddubs_epi16(vxj, vw);
- tmp = _mm256_madd_epi16(tmp, ones);
- vy0 = _mm256_add_epi32(vy0, tmp);
+ vy0 = _mm256_dpbusds_epi32(vy0, vxj, vw);
w += 32;
vxj = _mm256_set1_epi32(*(int*)&x[*idx++]);
vw = _mm256_loadu_si256((const __m256i *)w);
- tmp = _mm256_maddubs_epi16(vxj, vw);
- tmp = _mm256_madd_epi16(tmp, ones);
- vy0 = _mm256_add_epi32(vy0, tmp);
+ vy0 = _mm256_dpbusds_epi32(vy0, vxj, vw);
w += 32;
vxj = _mm256_set1_epi32(*(int*)&x[*idx++]);
vw = _mm256_loadu_si256((const __m256i *)w);
- tmp = _mm256_maddubs_epi16(vxj, vw);
- tmp = _mm256_madd_epi16(tmp, ones);
- vy0 = _mm256_add_epi32(vy0, tmp);
+ vy0 = _mm256_dpbusds_epi32(vy0, vxj, vw);
w += 32;
vxj = _mm256_set1_epi32(*(int*)&x[*idx++]);
vw = _mm256_loadu_si256((const __m256i *)w);
- tmp = _mm256_maddubs_epi16(vxj, vw);
- tmp = _mm256_madd_epi16(tmp, ones);
- vy0 = _mm256_add_epi32(vy0, tmp);
+ vy0 = _mm256_dpbusds_epi32(vy0, vxj, vw);
w += 32;
}
#endif
for (;j<colblocks;j++)
{
- __m256i tmp;
__m256i vxj;
__m256i vw;
int pos;
@@ -751,9 +765,7 @@
pos = (*idx++);
vxj = _mm256_set1_epi32(*(int*)&x[pos]);
vw = _mm256_loadu_si256((const __m256i *)w);
- tmp = _mm256_maddubs_epi16(vxj, vw);
- tmp = _mm256_madd_epi16(tmp, ones);
- vy0 = _mm256_add_epi32(vy0, tmp);
+ vy0 = _mm256_dpbusds_epi32(vy0, vxj, vw);
w += 32;
}
vout = _mm256_cvtepi32_ps(vy0);
@@ -763,10 +775,8 @@
}
static inline void cgemv8x4(float *_out, const opus_int8 *w, const float *scale, int rows, int cols, const float *_x)
{
- __m256i ones;
int i, j;
unsigned char x[MAX_INPUTS];
- ones = _mm256_set1_epi16(1);
/*for (i=0;i<cols;i++) x[i] = 127+floor(.5+127*_x[i]);*/
vector_ps_to_epi8(x, _x, cols);
for (i=0;i<rows;i+=8)
@@ -778,45 +788,33 @@
#if 1 /* Unrolling by 4 gives some gain, comment out if it does not. */
for (;j<cols-12;j+=16)
{
- __m256i tmp;
__m256i vxj;
__m256i vw;
vxj = _mm256_set1_epi32(*(int*)&x[j]);
vw = _mm256_loadu_si256((const __m256i *)w);
- tmp = _mm256_maddubs_epi16(vxj, vw);
- tmp = _mm256_madd_epi16(tmp, ones);
- vy0 = _mm256_add_epi32(vy0, tmp);
+ vy0 = _mm256_dpbusds_epi32(vy0, vxj, vw);
w += 32;
vxj = _mm256_set1_epi32(*(int*)&x[j+4]);
vw = _mm256_loadu_si256((const __m256i *)w);
- tmp = _mm256_maddubs_epi16(vxj, vw);
- tmp = _mm256_madd_epi16(tmp, ones);
- vy0 = _mm256_add_epi32(vy0, tmp);
+ vy0 = _mm256_dpbusds_epi32(vy0, vxj, vw);
w += 32;
vxj = _mm256_set1_epi32(*(int*)&x[j+8]);
vw = _mm256_loadu_si256((const __m256i *)w);
- tmp = _mm256_maddubs_epi16(vxj, vw);
- tmp = _mm256_madd_epi16(tmp, ones);
- vy0 = _mm256_add_epi32(vy0, tmp);
+ vy0 = _mm256_dpbusds_epi32(vy0, vxj, vw);
w += 32;
vxj = _mm256_set1_epi32(*(int*)&x[j+12]);
vw = _mm256_loadu_si256((const __m256i *)w);
- tmp = _mm256_maddubs_epi16(vxj, vw);
- tmp = _mm256_madd_epi16(tmp, ones);
- vy0 = _mm256_add_epi32(vy0, tmp);
+ vy0 = _mm256_dpbusds_epi32(vy0, vxj, vw);
w += 32;
}
#endif
for (;j<cols;j+=4)
{
- __m256i tmp;
__m256i vxj;
__m256i vw;
vxj = _mm256_set1_epi32(*(int*)&x[j]);
vw = _mm256_loadu_si256((const __m256i *)w);
- tmp = _mm256_maddubs_epi16(vxj, vw);
- tmp = _mm256_madd_epi16(tmp, ones);
- vy0 = _mm256_add_epi32(vy0, tmp);
+ vy0 = _mm256_dpbusds_epi32(vy0, vxj, vw);
w += 32;
}
vout = _mm256_cvtepi32_ps(vy0);
--
⑨