Home | History | Annotate | Download | only in x86
      1 /*
      2  *  Copyright (c) 2017 The WebM project authors. All Rights Reserved.
      3  *
      4  *  Use of this source code is governed by a BSD-style license
      5  *  that can be found in the LICENSE file in the root of the source
      6  *  tree. An additional intellectual property rights grant can be found
      7  *  in the file PATENTS.  All contributing project authors may
      8  *  be found in the AUTHORS file in the root of the source tree.
      9  */
     10 
     11 #ifndef VPX_DSP_X86_CONVOLVE_SSSE3_H_
     12 #define VPX_DSP_X86_CONVOLVE_SSSE3_H_
     13 
     14 #include <assert.h>
     15 #include <tmmintrin.h>  // SSSE3
     16 
     17 #include "./vpx_config.h"
     18 
     19 static INLINE void shuffle_filter_ssse3(const int16_t *const filter,
     20                                         __m128i *const f) {
     21   const __m128i f_values = _mm_load_si128((const __m128i *)filter);
     22   // pack and duplicate the filter values
     23   f[0] = _mm_shuffle_epi8(f_values, _mm_set1_epi16(0x0200u));
     24   f[1] = _mm_shuffle_epi8(f_values, _mm_set1_epi16(0x0604u));
     25   f[2] = _mm_shuffle_epi8(f_values, _mm_set1_epi16(0x0a08u));
     26   f[3] = _mm_shuffle_epi8(f_values, _mm_set1_epi16(0x0e0cu));
     27 }
     28 
     29 static INLINE void shuffle_filter_odd_ssse3(const int16_t *const filter,
     30                                             __m128i *const f) {
     31   const __m128i f_values = _mm_load_si128((const __m128i *)filter);
     32   // pack and duplicate the filter values
     33   // It utilizes the fact that the high byte of filter[3] is always 0 to clean
     34   // half of f[0] and f[4].
     35   assert(filter[3] >= 0 && filter[3] < 256);
     36   f[0] = _mm_shuffle_epi8(f_values, _mm_set1_epi16(0x0007u));
     37   f[1] = _mm_shuffle_epi8(f_values, _mm_set1_epi16(0x0402u));
     38   f[2] = _mm_shuffle_epi8(f_values, _mm_set1_epi16(0x0806u));
     39   f[3] = _mm_shuffle_epi8(f_values, _mm_set1_epi16(0x0c0au));
     40   f[4] = _mm_shuffle_epi8(f_values, _mm_set1_epi16(0x070eu));
     41 }
     42 
     43 static INLINE __m128i convolve8_8_ssse3(const __m128i *const s,
     44                                         const __m128i *const f) {
     45   // multiply 2 adjacent elements with the filter and add the result
     46   const __m128i k_64 = _mm_set1_epi16(1 << 6);
     47   const __m128i x0 = _mm_maddubs_epi16(s[0], f[0]);
     48   const __m128i x1 = _mm_maddubs_epi16(s[1], f[1]);
     49   const __m128i x2 = _mm_maddubs_epi16(s[2], f[2]);
     50   const __m128i x3 = _mm_maddubs_epi16(s[3], f[3]);
     51   __m128i sum1, sum2;
     52 
     53   // sum the results together, saturating only on the final step
     54   // adding x0 with x2 and x1 with x3 is the only order that prevents
     55   // outranges for all filters
     56   sum1 = _mm_add_epi16(x0, x2);
     57   sum2 = _mm_add_epi16(x1, x3);
     58   // add the rounding offset early to avoid another saturated add
     59   sum1 = _mm_add_epi16(sum1, k_64);
     60   sum1 = _mm_adds_epi16(sum1, sum2);
     61   // shift by 7 bit each 16 bit
     62   sum1 = _mm_srai_epi16(sum1, 7);
     63   return sum1;
     64 }
     65 
     66 static INLINE __m128i convolve8_8_even_offset_ssse3(const __m128i *const s,
     67                                                     const __m128i *const f) {
     68   // multiply 2 adjacent elements with the filter and add the result
     69   const __m128i k_64 = _mm_set1_epi16(1 << 6);
     70   const __m128i x0 = _mm_maddubs_epi16(s[0], f[0]);
     71   const __m128i x1 = _mm_maddubs_epi16(s[1], f[1]);
     72   const __m128i x2 = _mm_maddubs_epi16(s[2], f[2]);
     73   const __m128i x3 = _mm_maddubs_epi16(s[3], f[3]);
     74   // compensate the subtracted 64 in f[1]. x4 is always non negative.
     75   const __m128i x4 = _mm_maddubs_epi16(s[1], _mm_set1_epi8(64));
     76   // add and saturate the results together
     77   __m128i temp = _mm_adds_epi16(x0, x3);
     78   temp = _mm_adds_epi16(temp, x1);
     79   temp = _mm_adds_epi16(temp, x2);
     80   temp = _mm_adds_epi16(temp, x4);
     81   // round and shift by 7 bit each 16 bit
     82   temp = _mm_adds_epi16(temp, k_64);
     83   temp = _mm_srai_epi16(temp, 7);
     84   return temp;
     85 }
     86 
     87 static INLINE __m128i convolve8_8_odd_offset_ssse3(const __m128i *const s,
     88                                                    const __m128i *const f) {
     89   // multiply 2 adjacent elements with the filter and add the result
     90   const __m128i k_64 = _mm_set1_epi16(1 << 6);
     91   const __m128i x0 = _mm_maddubs_epi16(s[0], f[0]);
     92   const __m128i x1 = _mm_maddubs_epi16(s[1], f[1]);
     93   const __m128i x2 = _mm_maddubs_epi16(s[2], f[2]);
     94   const __m128i x3 = _mm_maddubs_epi16(s[3], f[3]);
     95   const __m128i x4 = _mm_maddubs_epi16(s[4], f[4]);
     96   // compensate the subtracted 64 in f[2]. x5 is always non negative.
     97   const __m128i x5 = _mm_maddubs_epi16(s[2], _mm_set1_epi8(64));
     98   __m128i temp;
     99 
    100   // add and saturate the results together
    101   temp = _mm_adds_epi16(x0, x1);
    102   temp = _mm_adds_epi16(temp, x2);
    103   temp = _mm_adds_epi16(temp, x3);
    104   temp = _mm_adds_epi16(temp, x4);
    105   temp = _mm_adds_epi16(temp, x5);
    106   // round and shift by 7 bit each 16 bit
    107   temp = _mm_adds_epi16(temp, k_64);
    108   temp = _mm_srai_epi16(temp, 7);
    109   return temp;
    110 }
    111 
    112 #endif  // VPX_DSP_X86_CONVOLVE_SSSE3_H_
    113