shithub: libvpx

Download patch

ref: acf24cc1b8b84ab62eeddac201ded0e45a0e0253
parent: 2030c49cf87681712ba200d2d1f06d09cc58adb1
author: Scott LaVarnway <[email protected]>
date: Thu Aug 20 09:09:27 EDT 2015

VPX: scaled convolve : fix windows build errors

Change-Id: Ic81d435ea928183197040cdf64b6afd7dbaf57e4

--- a/vpx_dsp/x86/vpx_subpixel_8t_intrin_ssse3.c
+++ b/vpx_dsp/x86/vpx_subpixel_8t_intrin_ssse3.c
@@ -616,7 +616,7 @@
   out7 = _mm_unpackhi_epi64(tr2_3, tr2_3);                              \
 }
 
-static void filter_horiz_w8_ssse3(const uint8_t *src_x, int src_pitch,
+static void filter_horiz_w8_ssse3(const uint8_t *src_x, ptrdiff_t src_pitch,
                                   uint8_t *dst, const int16_t *x_filter) {
   const __m128i k_256 = _mm_set1_epi16(1 << 8);
   const __m128i f_values = _mm_load_si128((const __m128i *)x_filter);
@@ -740,18 +740,17 @@
 
 static void filter_horiz_w4_ssse3(const uint8_t *src_ptr, ptrdiff_t src_pitch,
                                   uint8_t *dst, const int16_t *filter) {
-  const __m64 k_256 = _mm_set1_pi16(1 << 8);
-  const __m64 f_values_lo = *((const __m64 *)filter);
-  const __m64 f_values_hi = *((const __m64 *)(filter + 4));
+  const __m128i k_256 = _mm_set1_epi16(1 << 8);
+  const __m128i f_values = _mm_load_si128((const __m128i *)filter);
   // pack and duplicate the filter values
-  const __m64 f1f0 = _mm_shuffle_pi8(f_values_lo, _mm_set1_pi16(0x0200u));
-  const __m64 f3f2 = _mm_shuffle_pi8(f_values_lo, _mm_set1_pi16(0x0604u));
-  const __m64 f5f4 = _mm_shuffle_pi8(f_values_hi, _mm_set1_pi16(0x0a08u));
-  const __m64 f7f6 = _mm_shuffle_pi8(f_values_hi, _mm_set1_pi16(0x0e0cu));
-  const __m64 A = *((const __m64 *)src_ptr);
-  const __m64 B = *((const __m64 *)(src_ptr + src_pitch));
-  const __m64 C = *((const __m64 *)(src_ptr + src_pitch * 2));
-  const __m64 D = *((const __m64 *)(src_ptr + src_pitch * 3));
+  const __m128i f1f0 = _mm_shuffle_epi8(f_values, _mm_set1_epi16(0x0200u));
+  const __m128i f3f2 = _mm_shuffle_epi8(f_values, _mm_set1_epi16(0x0604u));
+  const __m128i f5f4 = _mm_shuffle_epi8(f_values, _mm_set1_epi16(0x0a08u));
+  const __m128i f7f6 = _mm_shuffle_epi8(f_values, _mm_set1_epi16(0x0e0cu));
+  const __m128i A = _mm_loadl_epi64((const __m128i *)src_ptr);
+  const __m128i B = _mm_loadl_epi64((const __m128i *)(src_ptr + src_pitch));
+  const __m128i C = _mm_loadl_epi64((const __m128i *)(src_ptr + src_pitch * 2));
+  const __m128i D = _mm_loadl_epi64((const __m128i *)(src_ptr + src_pitch * 3));
   // TRANSPOSE...
   // 00 01 02 03 04 05 06 07
   // 10 11 12 13 14 15 16 17
@@ -769,63 +768,57 @@
   // 06 16 26 36
   // 07 17 27 37
   //
-  // 00 01 10 11 02 03 12 13
-  const __m64 tr0_0 = _mm_unpacklo_pi16(A, B);
-  // 20 21 30 31 22 23 32 33
-  const __m64 tr0_1 = _mm_unpacklo_pi16(C, D);
-  // 04 05 14 15 06 07 16 17
-  const __m64 tr0_2 = _mm_unpackhi_pi16(A, B);
-  // 24 25 34 35 26 27 36 37
-  const __m64 tr0_3 = _mm_unpackhi_pi16(C, D);
-  // 00 01 10 11 20 21 30 31
-  const __m64 s1s0  = _mm_unpacklo_pi32(tr0_0, tr0_1);
+  // 00 01 10 11 02 03 12 13 04 05 14 15 06 07 16 17
+  const __m128i tr0_0 = _mm_unpacklo_epi16(A, B);
+  // 20 21 30 31 22 23 32 33 24 25 34 35 26 27 36 37
+  const __m128i tr0_1 = _mm_unpacklo_epi16(C, D);
+  // 00 01 10 11 20 21 30 31 02 03 12 13 22 23 32 33
+  const __m128i s1s0  = _mm_unpacklo_epi32(tr0_0, tr0_1);
+  // 04 05 14 15 24 25 34 35 06 07 16 17 26 27 36 37
+  const __m128i s5s4 = _mm_unpackhi_epi32(tr0_0, tr0_1);
   // 02 03 12 13 22 23 32 33
-  const __m64 s3s2 = _mm_unpackhi_pi32(tr0_0, tr0_1);
-  // 04 05 14 15 24 25 34 35
-  const __m64 s5s4 = _mm_unpacklo_pi32(tr0_2, tr0_3);
+  const __m128i s3s2 = _mm_srli_si128(s1s0, 8);
   // 06 07 16 17 26 27 36 37
-  const __m64 s7s6 = _mm_unpackhi_pi32(tr0_2, tr0_3);
+  const __m128i s7s6 = _mm_srli_si128(s5s4, 8);
   // multiply 2 adjacent elements with the filter and add the result
-  const __m64 x0 = _mm_maddubs_pi16(s1s0, f1f0);
-  const __m64 x1 = _mm_maddubs_pi16(s3s2, f3f2);
-  const __m64 x2 = _mm_maddubs_pi16(s5s4, f5f4);
-  const __m64 x3 = _mm_maddubs_pi16(s7s6, f7f6);
+  const __m128i x0 = _mm_maddubs_epi16(s1s0, f1f0);
+  const __m128i x1 = _mm_maddubs_epi16(s3s2, f3f2);
+  const __m128i x2 = _mm_maddubs_epi16(s5s4, f5f4);
+  const __m128i x3 = _mm_maddubs_epi16(s7s6, f7f6);
   // add and saturate the results together
-  const __m64 min_x2x1 = _mm_min_pi16(x2, x1);
-  const __m64 max_x2x1 = _mm_max_pi16(x2, x1);
-  __m64 temp = _mm_adds_pi16(x0, x3);
-  temp = _mm_adds_pi16(temp, min_x2x1);
-  temp = _mm_adds_pi16(temp, max_x2x1);
+  const __m128i min_x2x1 = _mm_min_epi16(x2, x1);
+  const __m128i max_x2x1 = _mm_max_epi16(x2, x1);
+  __m128i temp = _mm_adds_epi16(x0, x3);
+  temp = _mm_adds_epi16(temp, min_x2x1);
+  temp = _mm_adds_epi16(temp, max_x2x1);
   // round and shift by 7 bit each 16 bit
-  temp = _mm_mulhrs_pi16(temp, k_256);
+  temp = _mm_mulhrs_epi16(temp, k_256);
   // shrink to 8 bit each 16 bits
-  temp = _mm_packs_pu16(temp, temp);
+  temp = _mm_packus_epi16(temp, temp);
   // save only 4 bytes
-  *(int *)dst = _mm_cvtsi64_si32(temp);
+  *(int *)dst = _mm_cvtsi128_si32(temp);
 }
 
 static void transpose4x4_to_dst(const uint8_t *src, ptrdiff_t src_stride,
                                 uint8_t *dst, ptrdiff_t dst_stride) {
-  __m64 A = _mm_cvtsi32_si64(*(const int *)src);
-  __m64 B = _mm_cvtsi32_si64(*(const int *)(src + src_stride));
-  __m64 C = _mm_cvtsi32_si64(*(const int *)(src + src_stride * 2));
-  __m64 D = _mm_cvtsi32_si64(*(const int *)(src + src_stride * 3));
+  __m128i A = _mm_cvtsi32_si128(*(const int *)src);
+  __m128i B = _mm_cvtsi32_si128(*(const int *)(src + src_stride));
+  __m128i C = _mm_cvtsi32_si128(*(const int *)(src + src_stride * 2));
+  __m128i D = _mm_cvtsi32_si128(*(const int *)(src + src_stride * 3));
   // 00 10 01 11 02 12 03 13
-  const __m64 tr0_0 = _mm_unpacklo_pi8(A, B);
+  const __m128i tr0_0 = _mm_unpacklo_epi8(A, B);
   // 20 30 21 31 22 32 23 33
-  const __m64 tr0_1 = _mm_unpacklo_pi8(C, D);
-  // 00 10 20 30 01 11 21 31
-  A = _mm_unpacklo_pi16(tr0_0, tr0_1);
-  // 02 12 22 32 03 13 23 33
-  C = _mm_unpackhi_pi16(tr0_0, tr0_1);
+  const __m128i tr0_1 = _mm_unpacklo_epi8(C, D);
+  // 00 10 20 30 01 11 21 31 02 12 22 32 03 13 23 33
+  A = _mm_unpacklo_epi16(tr0_0, tr0_1);
+  B = _mm_srli_si128(A, 4);
+  C = _mm_srli_si128(A, 8);
+  D = _mm_srli_si128(A, 12);
 
-  B = _mm_unpackhi_pi32(A, A);
-  D = _mm_unpackhi_pi32(C, C);
-
-  *(int *)(dst) =  _mm_cvtsi64_si32(A);
-  *(int *)(dst + dst_stride) =  _mm_cvtsi64_si32(B);
-  *(int *)(dst + dst_stride * 2) =  _mm_cvtsi64_si32(C);
-  *(int *)(dst + dst_stride * 3) =  _mm_cvtsi64_si32(D);
+  *(int *)(dst) =  _mm_cvtsi128_si32(A);
+  *(int *)(dst + dst_stride) =  _mm_cvtsi128_si32(B);
+  *(int *)(dst + dst_stride * 2) =  _mm_cvtsi128_si32(C);
+  *(int *)(dst + dst_stride * 3) =  _mm_cvtsi128_si32(D);
 }
 
 static void scaledconvolve_horiz_w4(const uint8_t *src, ptrdiff_t src_stride,
@@ -865,46 +858,42 @@
 
 static void filter_vert_w4_ssse3(const uint8_t *src_ptr, ptrdiff_t src_pitch,
                                  uint8_t *dst, const int16_t *filter) {
-  const __m64 k_256 = _mm_set1_pi16(1 << 8);
-  const __m64 f_values_lo = *((const __m64 *)filter);
-  const __m64 f_values_hi = *((const __m64 *)(filter + 4));
+  const __m128i k_256 = _mm_set1_epi16(1 << 8);
+  const __m128i f_values = _mm_load_si128((const __m128i *)filter);
   // pack and duplicate the filter values
-  const __m64 f1f0 = _mm_shuffle_pi8(f_values_lo, _mm_set1_pi16(0x0200u));
-  const __m64 f3f2 = _mm_shuffle_pi8(f_values_lo, _mm_set1_pi16(0x0604u));
-  const __m64 f5f4 = _mm_shuffle_pi8(f_values_hi, _mm_set1_pi16(0x0a08u));
-  const __m64 f7f6 = _mm_shuffle_pi8(f_values_hi, _mm_set1_pi16(0x0e0cu));
-  const __m64 A = _mm_cvtsi32_si64(*(const int *)src_ptr);
-  const __m64 B = _mm_cvtsi32_si64(*(const int *)(src_ptr + src_pitch));
-  const __m64 C = _mm_cvtsi32_si64(*(const int *)(src_ptr + src_pitch * 2));
-  const __m64 D = _mm_cvtsi32_si64(*(const int *)(src_ptr + src_pitch * 3));
-  const __m64 E = _mm_cvtsi32_si64(*(const int *)(src_ptr + src_pitch * 4));
-  const __m64 F = _mm_cvtsi32_si64(*(const int *)(src_ptr + src_pitch * 5));
-  const __m64 G = _mm_cvtsi32_si64(*(const int *)(src_ptr + src_pitch * 6));
-  const __m64 H = _mm_cvtsi32_si64(*(const int *)(src_ptr + src_pitch * 7));
-  const __m64 s1s0 = _mm_unpacklo_pi8(A, B);
-  // 02 03 12 13 22 23 32 33
-  const __m64 s3s2 = _mm_unpacklo_pi8(C, D);
-  // 04 05 14 15 24 25 34 35
-  const __m64 s5s4 = _mm_unpacklo_pi8(E, F);
-  // 06 07 16 17 26 27 36 37
-  const __m64 s7s6 = _mm_unpacklo_pi8(G, H);
+  const __m128i f1f0 = _mm_shuffle_epi8(f_values, _mm_set1_epi16(0x0200u));
+  const __m128i f3f2 = _mm_shuffle_epi8(f_values, _mm_set1_epi16(0x0604u));
+  const __m128i f5f4 = _mm_shuffle_epi8(f_values, _mm_set1_epi16(0x0a08u));
+  const __m128i f7f6 = _mm_shuffle_epi8(f_values, _mm_set1_epi16(0x0e0cu));
+  const __m128i A = _mm_cvtsi32_si128(*(const int *)src_ptr);
+  const __m128i B = _mm_cvtsi32_si128(*(const int *)(src_ptr + src_pitch));
+  const __m128i C = _mm_cvtsi32_si128(*(const int *)(src_ptr + src_pitch * 2));
+  const __m128i D = _mm_cvtsi32_si128(*(const int *)(src_ptr + src_pitch * 3));
+  const __m128i E = _mm_cvtsi32_si128(*(const int *)(src_ptr + src_pitch * 4));
+  const __m128i F = _mm_cvtsi32_si128(*(const int *)(src_ptr + src_pitch * 5));
+  const __m128i G = _mm_cvtsi32_si128(*(const int *)(src_ptr + src_pitch * 6));
+  const __m128i H = _mm_cvtsi32_si128(*(const int *)(src_ptr + src_pitch * 7));
+  const __m128i s1s0 = _mm_unpacklo_epi8(A, B);
+  const __m128i s3s2 = _mm_unpacklo_epi8(C, D);
+  const __m128i s5s4 = _mm_unpacklo_epi8(E, F);
+  const __m128i s7s6 = _mm_unpacklo_epi8(G, H);
   // multiply 2 adjacent elements with the filter and add the result
-  const __m64 x0 = _mm_maddubs_pi16(s1s0, f1f0);
-  const __m64 x1 = _mm_maddubs_pi16(s3s2, f3f2);
-  const __m64 x2 = _mm_maddubs_pi16(s5s4, f5f4);
-  const __m64 x3 = _mm_maddubs_pi16(s7s6, f7f6);
+  const __m128i x0 = _mm_maddubs_epi16(s1s0, f1f0);
+  const __m128i x1 = _mm_maddubs_epi16(s3s2, f3f2);
+  const __m128i x2 = _mm_maddubs_epi16(s5s4, f5f4);
+  const __m128i x3 = _mm_maddubs_epi16(s7s6, f7f6);
   // add and saturate the results together
-  const __m64 min_x2x1 = _mm_min_pi16(x2, x1);
-  const __m64 max_x2x1 = _mm_max_pi16(x2, x1);
-  __m64 temp = _mm_adds_pi16(x0, x3);
-  temp = _mm_adds_pi16(temp, min_x2x1);
-  temp = _mm_adds_pi16(temp, max_x2x1);
+  const __m128i min_x2x1 = _mm_min_epi16(x2, x1);
+  const __m128i max_x2x1 = _mm_max_epi16(x2, x1);
+  __m128i temp = _mm_adds_epi16(x0, x3);
+  temp = _mm_adds_epi16(temp, min_x2x1);
+  temp = _mm_adds_epi16(temp, max_x2x1);
   // round and shift by 7 bit each 16 bit
-  temp = _mm_mulhrs_pi16(temp, k_256);
+  temp = _mm_mulhrs_epi16(temp, k_256);
   // shrink to 8 bit each 16 bits
-  temp = _mm_packs_pu16(temp, temp);
+  temp = _mm_packus_epi16(temp, temp);
   // save only 4 bytes
-  *(int *)dst = _mm_cvtsi64_si32(temp);
+  *(int *)dst = _mm_cvtsi128_si32(temp);
 }
 
 static void scaledconvolve_vert_w4(const uint8_t *src, ptrdiff_t src_stride,