Home | History | Annotate | Download | only in x86
      1 /*
      2  *  Copyright (c) 2015 The WebM project authors. All Rights Reserved.
      3  *
      4  *  Use of this source code is governed by a BSD-style license
      5  *  that can be found in the LICENSE file in the root of the source
      6  *  tree. An additional intellectual property rights grant can be found
      7  *  in the file PATENTS.  All contributing project authors may
      8  *  be found in the AUTHORS file in the root of the source tree.
      9  */
     10 
     11 #ifndef VPX_DSP_X86_HIGHBD_INV_TXFM_SSE2_H_
     12 #define VPX_DSP_X86_HIGHBD_INV_TXFM_SSE2_H_
     13 
     14 #include <emmintrin.h>  // SSE2
     15 
     16 #include "./vpx_config.h"
     17 #include "vpx/vpx_integer.h"
     18 #include "vpx_dsp/inv_txfm.h"
     19 #include "vpx_dsp/x86/transpose_sse2.h"
     20 #include "vpx_dsp/x86/txfm_common_sse2.h"
     21 
     22 static INLINE void extend_64bit(const __m128i in,
     23                                 __m128i *const out /*out[2]*/) {
     24   out[0] = _mm_unpacklo_epi32(in, in);  // 0, 0, 1, 1
     25   out[1] = _mm_unpackhi_epi32(in, in);  // 2, 2, 3, 3
     26 }
     27 
     28 static INLINE __m128i wraplow_16bit_shift4(const __m128i in0, const __m128i in1,
     29                                            const __m128i rounding) {
     30   __m128i temp[2];
     31   temp[0] = _mm_add_epi32(in0, rounding);
     32   temp[1] = _mm_add_epi32(in1, rounding);
     33   temp[0] = _mm_srai_epi32(temp[0], 4);
     34   temp[1] = _mm_srai_epi32(temp[1], 4);
     35   return _mm_packs_epi32(temp[0], temp[1]);
     36 }
     37 
     38 static INLINE __m128i wraplow_16bit_shift5(const __m128i in0, const __m128i in1,
     39                                            const __m128i rounding) {
     40   __m128i temp[2];
     41   temp[0] = _mm_add_epi32(in0, rounding);
     42   temp[1] = _mm_add_epi32(in1, rounding);
     43   temp[0] = _mm_srai_epi32(temp[0], 5);
     44   temp[1] = _mm_srai_epi32(temp[1], 5);
     45   return _mm_packs_epi32(temp[0], temp[1]);
     46 }
     47 
     48 static INLINE __m128i dct_const_round_shift_64bit(const __m128i in) {
     49   const __m128i t =
     50       _mm_add_epi64(in, pair_set_epi32(DCT_CONST_ROUNDING << 2, 0));
     51   return _mm_srli_si128(t, 2);
     52 }
     53 
     54 static INLINE __m128i pack_4(const __m128i in0, const __m128i in1) {
     55   const __m128i t0 = _mm_unpacklo_epi32(in0, in1);  // 0, 2
     56   const __m128i t1 = _mm_unpackhi_epi32(in0, in1);  // 1, 3
     57   return _mm_unpacklo_epi32(t0, t1);                // 0, 1, 2, 3
     58 }
     59 
     60 static INLINE void abs_extend_64bit_sse2(const __m128i in,
     61                                          __m128i *const out /*out[2]*/,
     62                                          __m128i *const sign /*sign[2]*/) {
     63   sign[0] = _mm_srai_epi32(in, 31);
     64   out[0] = _mm_xor_si128(in, sign[0]);
     65   out[0] = _mm_sub_epi32(out[0], sign[0]);
     66   sign[1] = _mm_unpackhi_epi32(sign[0], sign[0]);  // 64-bit sign of 2, 3
     67   sign[0] = _mm_unpacklo_epi32(sign[0], sign[0]);  // 64-bit sign of 0, 1
     68   out[1] = _mm_unpackhi_epi32(out[0], out[0]);     // 2, 3
     69   out[0] = _mm_unpacklo_epi32(out[0], out[0]);     // 0, 1
     70 }
     71 
     72 // Note: cospi must be non negative.
     73 static INLINE __m128i multiply_apply_sign_sse2(const __m128i in,
     74                                                const __m128i sign,
     75                                                const __m128i cospi) {
     76   __m128i out = _mm_mul_epu32(in, cospi);
     77   out = _mm_xor_si128(out, sign);
     78   return _mm_sub_epi64(out, sign);
     79 }
     80 
     81 // Note: c must be non negative.
     82 static INLINE __m128i multiplication_round_shift_sse2(
     83     const __m128i *const in /*in[2]*/, const __m128i *const sign /*sign[2]*/,
     84     const int c) {
     85   const __m128i pair_c = pair_set_epi32(c << 2, 0);
     86   __m128i t0, t1;
     87 
     88   assert(c >= 0);
     89   t0 = multiply_apply_sign_sse2(in[0], sign[0], pair_c);
     90   t1 = multiply_apply_sign_sse2(in[1], sign[1], pair_c);
     91   t0 = dct_const_round_shift_64bit(t0);
     92   t1 = dct_const_round_shift_64bit(t1);
     93 
     94   return pack_4(t0, t1);
     95 }
     96 
     97 // Note: c must be non negative.
     98 static INLINE __m128i multiplication_neg_round_shift_sse2(
     99     const __m128i *const in /*in[2]*/, const __m128i *const sign /*sign[2]*/,
    100     const int c) {
    101   const __m128i pair_c = pair_set_epi32(c << 2, 0);
    102   __m128i t0, t1;
    103 
    104   assert(c >= 0);
    105   t0 = multiply_apply_sign_sse2(in[0], sign[0], pair_c);
    106   t1 = multiply_apply_sign_sse2(in[1], sign[1], pair_c);
    107   t0 = _mm_sub_epi64(_mm_setzero_si128(), t0);
    108   t1 = _mm_sub_epi64(_mm_setzero_si128(), t1);
    109   t0 = dct_const_round_shift_64bit(t0);
    110   t1 = dct_const_round_shift_64bit(t1);
    111 
    112   return pack_4(t0, t1);
    113 }
    114 
    115 // Note: c0 and c1 must be non negative.
    116 static INLINE void highbd_butterfly_sse2(const __m128i in0, const __m128i in1,
    117                                          const int c0, const int c1,
    118                                          __m128i *const out0,
    119                                          __m128i *const out1) {
    120   const __m128i pair_c0 = pair_set_epi32(c0 << 2, 0);
    121   const __m128i pair_c1 = pair_set_epi32(c1 << 2, 0);
    122   __m128i temp1[4], temp2[4], sign1[2], sign2[2];
    123 
    124   assert(c0 >= 0);
    125   assert(c1 >= 0);
    126   abs_extend_64bit_sse2(in0, temp1, sign1);
    127   abs_extend_64bit_sse2(in1, temp2, sign2);
    128   temp1[2] = multiply_apply_sign_sse2(temp1[0], sign1[0], pair_c1);
    129   temp1[3] = multiply_apply_sign_sse2(temp1[1], sign1[1], pair_c1);
    130   temp1[0] = multiply_apply_sign_sse2(temp1[0], sign1[0], pair_c0);
    131   temp1[1] = multiply_apply_sign_sse2(temp1[1], sign1[1], pair_c0);
    132   temp2[2] = multiply_apply_sign_sse2(temp2[0], sign2[0], pair_c0);
    133   temp2[3] = multiply_apply_sign_sse2(temp2[1], sign2[1], pair_c0);
    134   temp2[0] = multiply_apply_sign_sse2(temp2[0], sign2[0], pair_c1);
    135   temp2[1] = multiply_apply_sign_sse2(temp2[1], sign2[1], pair_c1);
    136   temp1[0] = _mm_sub_epi64(temp1[0], temp2[0]);
    137   temp1[1] = _mm_sub_epi64(temp1[1], temp2[1]);
    138   temp2[0] = _mm_add_epi64(temp1[2], temp2[2]);
    139   temp2[1] = _mm_add_epi64(temp1[3], temp2[3]);
    140   temp1[0] = dct_const_round_shift_64bit(temp1[0]);
    141   temp1[1] = dct_const_round_shift_64bit(temp1[1]);
    142   temp2[0] = dct_const_round_shift_64bit(temp2[0]);
    143   temp2[1] = dct_const_round_shift_64bit(temp2[1]);
    144   *out0 = pack_4(temp1[0], temp1[1]);
    145   *out1 = pack_4(temp2[0], temp2[1]);
    146 }
    147 
    148 // Note: c0 and c1 must be non negative.
    149 static INLINE void highbd_partial_butterfly_sse2(const __m128i in, const int c0,
    150                                                  const int c1,
    151                                                  __m128i *const out0,
    152                                                  __m128i *const out1) {
    153   __m128i temp[2], sign[2];
    154 
    155   assert(c0 >= 0);
    156   assert(c1 >= 0);
    157   abs_extend_64bit_sse2(in, temp, sign);
    158   *out0 = multiplication_round_shift_sse2(temp, sign, c0);
    159   *out1 = multiplication_round_shift_sse2(temp, sign, c1);
    160 }
    161 
    162 // Note: c0 and c1 must be non negative.
    163 static INLINE void highbd_partial_butterfly_neg_sse2(const __m128i in,
    164                                                      const int c0, const int c1,
    165                                                      __m128i *const out0,
    166                                                      __m128i *const out1) {
    167   __m128i temp[2], sign[2];
    168 
    169   assert(c0 >= 0);
    170   assert(c1 >= 0);
    171   abs_extend_64bit_sse2(in, temp, sign);
    172   *out0 = multiplication_neg_round_shift_sse2(temp, sign, c1);
    173   *out1 = multiplication_round_shift_sse2(temp, sign, c0);
    174 }
    175 
    176 static INLINE void highbd_butterfly_cospi16_sse2(const __m128i in0,
    177                                                  const __m128i in1,
    178                                                  __m128i *const out0,
    179                                                  __m128i *const out1) {
    180   __m128i temp1[2], temp2, sign[2];
    181 
    182   temp2 = _mm_add_epi32(in0, in1);
    183   abs_extend_64bit_sse2(temp2, temp1, sign);
    184   *out0 = multiplication_round_shift_sse2(temp1, sign, cospi_16_64);
    185   temp2 = _mm_sub_epi32(in0, in1);
    186   abs_extend_64bit_sse2(temp2, temp1, sign);
    187   *out1 = multiplication_round_shift_sse2(temp1, sign, cospi_16_64);
    188 }
    189 
    190 // Only do addition and subtraction butterfly, size = 16, 32
    191 static INLINE void highbd_add_sub_butterfly(const __m128i *in, __m128i *out,
    192                                             int size) {
    193   int i = 0;
    194   const int num = size >> 1;
    195   const int bound = size - 1;
    196   while (i < num) {
    197     out[i] = _mm_add_epi32(in[i], in[bound - i]);
    198     out[bound - i] = _mm_sub_epi32(in[i], in[bound - i]);
    199     i++;
    200   }
    201 }
    202 
    203 static INLINE void highbd_idct8_stage4(const __m128i *const in,
    204                                        __m128i *const out) {
    205   out[0] = _mm_add_epi32(in[0], in[7]);
    206   out[1] = _mm_add_epi32(in[1], in[6]);
    207   out[2] = _mm_add_epi32(in[2], in[5]);
    208   out[3] = _mm_add_epi32(in[3], in[4]);
    209   out[4] = _mm_sub_epi32(in[3], in[4]);
    210   out[5] = _mm_sub_epi32(in[2], in[5]);
    211   out[6] = _mm_sub_epi32(in[1], in[6]);
    212   out[7] = _mm_sub_epi32(in[0], in[7]);
    213 }
    214 
    215 static INLINE void highbd_idct8x8_final_round(__m128i *const io) {
    216   io[0] = wraplow_16bit_shift5(io[0], io[8], _mm_set1_epi32(16));
    217   io[1] = wraplow_16bit_shift5(io[1], io[9], _mm_set1_epi32(16));
    218   io[2] = wraplow_16bit_shift5(io[2], io[10], _mm_set1_epi32(16));
    219   io[3] = wraplow_16bit_shift5(io[3], io[11], _mm_set1_epi32(16));
    220   io[4] = wraplow_16bit_shift5(io[4], io[12], _mm_set1_epi32(16));
    221   io[5] = wraplow_16bit_shift5(io[5], io[13], _mm_set1_epi32(16));
    222   io[6] = wraplow_16bit_shift5(io[6], io[14], _mm_set1_epi32(16));
    223   io[7] = wraplow_16bit_shift5(io[7], io[15], _mm_set1_epi32(16));
    224 }
    225 
    226 static INLINE void highbd_idct16_4col_stage7(const __m128i *const in,
    227                                              __m128i *const out) {
    228   out[0] = _mm_add_epi32(in[0], in[15]);
    229   out[1] = _mm_add_epi32(in[1], in[14]);
    230   out[2] = _mm_add_epi32(in[2], in[13]);
    231   out[3] = _mm_add_epi32(in[3], in[12]);
    232   out[4] = _mm_add_epi32(in[4], in[11]);
    233   out[5] = _mm_add_epi32(in[5], in[10]);
    234   out[6] = _mm_add_epi32(in[6], in[9]);
    235   out[7] = _mm_add_epi32(in[7], in[8]);
    236   out[8] = _mm_sub_epi32(in[7], in[8]);
    237   out[9] = _mm_sub_epi32(in[6], in[9]);
    238   out[10] = _mm_sub_epi32(in[5], in[10]);
    239   out[11] = _mm_sub_epi32(in[4], in[11]);
    240   out[12] = _mm_sub_epi32(in[3], in[12]);
    241   out[13] = _mm_sub_epi32(in[2], in[13]);
    242   out[14] = _mm_sub_epi32(in[1], in[14]);
    243   out[15] = _mm_sub_epi32(in[0], in[15]);
    244 }
    245 
    246 static INLINE __m128i add_clamp(const __m128i in0, const __m128i in1,
    247                                 const int bd) {
    248   const __m128i zero = _mm_set1_epi16(0);
    249   // Faster than _mm_set1_epi16((1 << bd) - 1).
    250   const __m128i one = _mm_set1_epi16(1);
    251   const __m128i max = _mm_sub_epi16(_mm_slli_epi16(one, bd), one);
    252   __m128i d;
    253 
    254   d = _mm_adds_epi16(in0, in1);
    255   d = _mm_max_epi16(d, zero);
    256   d = _mm_min_epi16(d, max);
    257 
    258   return d;
    259 }
    260 
    261 static INLINE void highbd_idct_1_add_kernel(const tran_low_t *input,
    262                                             uint16_t *dest, int stride, int bd,
    263                                             const int size) {
    264   int a1, i, j;
    265   tran_low_t out;
    266   __m128i dc, d;
    267 
    268   out = HIGHBD_WRAPLOW(
    269       dct_const_round_shift(input[0] * (tran_high_t)cospi_16_64), bd);
    270   out =
    271       HIGHBD_WRAPLOW(dct_const_round_shift(out * (tran_high_t)cospi_16_64), bd);
    272   a1 = ROUND_POWER_OF_TWO(out, (size == 8) ? 5 : 6);
    273   dc = _mm_set1_epi16(a1);
    274 
    275   for (i = 0; i < size; ++i) {
    276     for (j = 0; j < size; j += 8) {
    277       d = _mm_load_si128((const __m128i *)(&dest[j]));
    278       d = add_clamp(d, dc, bd);
    279       _mm_store_si128((__m128i *)(&dest[j]), d);
    280     }
    281     dest += stride;
    282   }
    283 }
    284 
    285 static INLINE void recon_and_store_4(const __m128i in, uint16_t *const dest,
    286                                      const int bd) {
    287   __m128i d;
    288 
    289   d = _mm_loadl_epi64((const __m128i *)dest);
    290   d = add_clamp(d, in, bd);
    291   _mm_storel_epi64((__m128i *)dest, d);
    292 }
    293 
    294 static INLINE void recon_and_store_4x2(const __m128i in, uint16_t *const dest,
    295                                        const int stride, const int bd) {
    296   __m128i d;
    297 
    298   d = _mm_loadl_epi64((const __m128i *)(dest + 0 * stride));
    299   d = _mm_castps_si128(
    300       _mm_loadh_pi(_mm_castsi128_ps(d), (const __m64 *)(dest + 1 * stride)));
    301   d = add_clamp(d, in, bd);
    302   _mm_storel_epi64((__m128i *)(dest + 0 * stride), d);
    303   _mm_storeh_pi((__m64 *)(dest + 1 * stride), _mm_castsi128_ps(d));
    304 }
    305 
    306 static INLINE void recon_and_store_4x4(const __m128i *const in, uint16_t *dest,
    307                                        const int stride, const int bd) {
    308   recon_and_store_4x2(in[0], dest, stride, bd);
    309   dest += 2 * stride;
    310   recon_and_store_4x2(in[1], dest, stride, bd);
    311 }
    312 
    313 static INLINE void recon_and_store_8(const __m128i in, uint16_t **const dest,
    314                                      const int stride, const int bd) {
    315   __m128i d;
    316 
    317   d = _mm_load_si128((const __m128i *)(*dest));
    318   d = add_clamp(d, in, bd);
    319   _mm_store_si128((__m128i *)(*dest), d);
    320   *dest += stride;
    321 }
    322 
    323 static INLINE void recon_and_store_8x8(const __m128i *const in, uint16_t *dest,
    324                                        const int stride, const int bd) {
    325   recon_and_store_8(in[0], &dest, stride, bd);
    326   recon_and_store_8(in[1], &dest, stride, bd);
    327   recon_and_store_8(in[2], &dest, stride, bd);
    328   recon_and_store_8(in[3], &dest, stride, bd);
    329   recon_and_store_8(in[4], &dest, stride, bd);
    330   recon_and_store_8(in[5], &dest, stride, bd);
    331   recon_and_store_8(in[6], &dest, stride, bd);
    332   recon_and_store_8(in[7], &dest, stride, bd);
    333 }
    334 
    335 static INLINE __m128i load_pack_8_32bit(const tran_low_t *const input) {
    336   const __m128i t0 = _mm_load_si128((const __m128i *)(input + 0));
    337   const __m128i t1 = _mm_load_si128((const __m128i *)(input + 4));
    338   return _mm_packs_epi32(t0, t1);
    339 }
    340 
    341 static INLINE void highbd_load_pack_transpose_32bit_8x8(const tran_low_t *input,
    342                                                         const int stride,
    343                                                         __m128i *const in) {
    344   in[0] = load_pack_8_32bit(input + 0 * stride);
    345   in[1] = load_pack_8_32bit(input + 1 * stride);
    346   in[2] = load_pack_8_32bit(input + 2 * stride);
    347   in[3] = load_pack_8_32bit(input + 3 * stride);
    348   in[4] = load_pack_8_32bit(input + 4 * stride);
    349   in[5] = load_pack_8_32bit(input + 5 * stride);
    350   in[6] = load_pack_8_32bit(input + 6 * stride);
    351   in[7] = load_pack_8_32bit(input + 7 * stride);
    352   transpose_16bit_8x8(in, in);
    353 }
    354 
    355 static INLINE void highbd_load_transpose_32bit_8x4(const tran_low_t *input,
    356                                                    const int stride,
    357                                                    __m128i *in) {
    358   in[0] = _mm_load_si128((const __m128i *)(input + 0 * stride + 0));
    359   in[1] = _mm_load_si128((const __m128i *)(input + 0 * stride + 4));
    360   in[2] = _mm_load_si128((const __m128i *)(input + 1 * stride + 0));
    361   in[3] = _mm_load_si128((const __m128i *)(input + 1 * stride + 4));
    362   in[4] = _mm_load_si128((const __m128i *)(input + 2 * stride + 0));
    363   in[5] = _mm_load_si128((const __m128i *)(input + 2 * stride + 4));
    364   in[6] = _mm_load_si128((const __m128i *)(input + 3 * stride + 0));
    365   in[7] = _mm_load_si128((const __m128i *)(input + 3 * stride + 4));
    366   transpose_32bit_8x4(in, in);
    367 }
    368 
    369 static INLINE void highbd_load_transpose_32bit_4x4(const tran_low_t *input,
    370                                                    const int stride,
    371                                                    __m128i *in) {
    372   in[0] = _mm_load_si128((const __m128i *)(input + 0 * stride));
    373   in[1] = _mm_load_si128((const __m128i *)(input + 1 * stride));
    374   in[2] = _mm_load_si128((const __m128i *)(input + 2 * stride));
    375   in[3] = _mm_load_si128((const __m128i *)(input + 3 * stride));
    376   transpose_32bit_4x4(in, in);
    377 }
    378 
    379 static INLINE void highbd_write_buffer_8(uint16_t *dest, const __m128i in,
    380                                          const int bd) {
    381   const __m128i final_rounding = _mm_set1_epi16(1 << 5);
    382   __m128i out;
    383 
    384   out = _mm_adds_epi16(in, final_rounding);
    385   out = _mm_srai_epi16(out, 6);
    386   recon_and_store_8(out, &dest, 0, bd);
    387 }
    388 
    389 static INLINE void highbd_write_buffer_4(uint16_t *const dest, const __m128i in,
    390                                          const int bd) {
    391   const __m128i final_rounding = _mm_set1_epi32(1 << 5);
    392   __m128i out;
    393 
    394   out = _mm_add_epi32(in, final_rounding);
    395   out = _mm_srai_epi32(out, 6);
    396   out = _mm_packs_epi32(out, out);
    397   recon_and_store_4(out, dest, bd);
    398 }
    399 
    400 #endif  // VPX_DSP_X86_HIGHBD_INV_TXFM_SSE2_H_
    401