1 /* Copyright (c) 2013 Jean-Marc Valin and John Ridges */ 2 /** 3 @file pitch_sse.h 4 @brief Pitch analysis 5 */ 6 7 /* 8 Redistribution and use in source and binary forms, with or without 9 modification, are permitted provided that the following conditions 10 are met: 11 12 - Redistributions of source code must retain the above copyright 13 notice, this list of conditions and the following disclaimer. 14 15 - Redistributions in binary form must reproduce the above copyright 16 notice, this list of conditions and the following disclaimer in the 17 documentation and/or other materials provided with the distribution. 18 19 THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 20 ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 21 LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR 22 A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER 23 OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, 24 EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, 25 PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR 26 PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF 27 LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING 28 NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS 29 SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 30 */ 31 32 #ifndef PITCH_SSE_H 33 #define PITCH_SSE_H 34 35 #include <xmmintrin.h> 36 #include "arch.h" 37 38 #define OVERRIDE_XCORR_KERNEL 39 static inline void xcorr_kernel(const opus_val16 *x, const opus_val16 *y, opus_val32 sum[4], int len) 40 { 41 int j; 42 __m128 xsum1, xsum2; 43 xsum1 = _mm_loadu_ps(sum); 44 xsum2 = _mm_setzero_ps(); 45 46 for (j = 0; j < len-3; j += 4) 47 { 48 __m128 x0 = _mm_loadu_ps(x+j); 49 __m128 yj = _mm_loadu_ps(y+j); 50 __m128 y3 = _mm_loadu_ps(y+j+3); 51 52 xsum1 = _mm_add_ps(xsum1,_mm_mul_ps(_mm_shuffle_ps(x0,x0,0x00),yj)); 53 xsum2 = _mm_add_ps(xsum2,_mm_mul_ps(_mm_shuffle_ps(x0,x0,0x55), 54 _mm_shuffle_ps(yj,y3,0x49))); 55 xsum1 = _mm_add_ps(xsum1,_mm_mul_ps(_mm_shuffle_ps(x0,x0,0xaa), 56 _mm_shuffle_ps(yj,y3,0x9e))); 57 xsum2 = _mm_add_ps(xsum2,_mm_mul_ps(_mm_shuffle_ps(x0,x0,0xff),y3)); 58 } 59 if (j < len) 60 { 61 xsum1 = _mm_add_ps(xsum1,_mm_mul_ps(_mm_load1_ps(x+j),_mm_loadu_ps(y+j))); 62 if (++j < len) 63 { 64 xsum2 = _mm_add_ps(xsum2,_mm_mul_ps(_mm_load1_ps(x+j),_mm_loadu_ps(y+j))); 65 if (++j < len) 66 { 67 xsum1 = _mm_add_ps(xsum1,_mm_mul_ps(_mm_load1_ps(x+j),_mm_loadu_ps(y+j))); 68 } 69 } 70 } 71 _mm_storeu_ps(sum,_mm_add_ps(xsum1,xsum2)); 72 } 73 74 #define OVERRIDE_DUAL_INNER_PROD 75 static inline void dual_inner_prod(const opus_val16 *x, const opus_val16 *y01, const opus_val16 *y02, 76 int N, opus_val32 *xy1, opus_val32 *xy2) 77 { 78 int i; 79 __m128 xsum1, xsum2; 80 xsum1 = _mm_setzero_ps(); 81 xsum2 = _mm_setzero_ps(); 82 for (i=0;i<N-3;i+=4) 83 { 84 __m128 xi = _mm_loadu_ps(x+i); 85 __m128 y1i = _mm_loadu_ps(y01+i); 86 __m128 y2i = _mm_loadu_ps(y02+i); 87 xsum1 = _mm_add_ps(xsum1,_mm_mul_ps(xi, y1i)); 88 xsum2 = _mm_add_ps(xsum2,_mm_mul_ps(xi, y2i)); 89 } 90 /* Horizontal sum */ 91 xsum1 = _mm_add_ps(xsum1, _mm_movehl_ps(xsum1, xsum1)); 92 xsum1 = _mm_add_ss(xsum1, _mm_shuffle_ps(xsum1, xsum1, 0x55)); 93 _mm_store_ss(xy1, xsum1); 94 xsum2 = _mm_add_ps(xsum2, _mm_movehl_ps(xsum2, xsum2)); 95 xsum2 = _mm_add_ss(xsum2, _mm_shuffle_ps(xsum2, xsum2, 0x55)); 96 _mm_store_ss(xy2, xsum2); 97 for (;i<N;i++) 98 { 99 *xy1 = MAC16_16(*xy1, x[i], y01[i]); 100 *xy2 = MAC16_16(*xy2, x[i], y02[i]); 101 } 102 } 103 104 #define OVERRIDE_COMB_FILTER_CONST 105 static inline void comb_filter_const(opus_val32 *y, opus_val32 *x, int T, int N, 106 opus_val16 g10, opus_val16 g11, opus_val16 g12) 107 { 108 int i; 109 __m128 x0v; 110 __m128 g10v, g11v, g12v; 111 g10v = _mm_load1_ps(&g10); 112 g11v = _mm_load1_ps(&g11); 113 g12v = _mm_load1_ps(&g12); 114 x0v = _mm_loadu_ps(&x[-T-2]); 115 for (i=0;i<N-3;i+=4) 116 { 117 __m128 yi, yi2, x1v, x2v, x3v, x4v; 118 const opus_val32 *xp = &x[i-T-2]; 119 yi = _mm_loadu_ps(x+i); 120 x4v = _mm_loadu_ps(xp+4); 121 #if 0 122 /* Slower version with all loads */ 123 x1v = _mm_loadu_ps(xp+1); 124 x2v = _mm_loadu_ps(xp+2); 125 x3v = _mm_loadu_ps(xp+3); 126 #else 127 x2v = _mm_shuffle_ps(x0v, x4v, 0x4e); 128 x1v = _mm_shuffle_ps(x0v, x2v, 0x99); 129 x3v = _mm_shuffle_ps(x2v, x4v, 0x99); 130 #endif 131 132 yi = _mm_add_ps(yi, _mm_mul_ps(g10v,x2v)); 133 #if 0 /* Set to 1 to make it bit-exact with the non-SSE version */ 134 yi = _mm_add_ps(yi, _mm_mul_ps(g11v,_mm_add_ps(x3v,x1v))); 135 yi = _mm_add_ps(yi, _mm_mul_ps(g12v,_mm_add_ps(x4v,x0v))); 136 #else 137 /* Use partial sums */ 138 yi2 = _mm_add_ps(_mm_mul_ps(g11v,_mm_add_ps(x3v,x1v)), 139 _mm_mul_ps(g12v,_mm_add_ps(x4v,x0v))); 140 yi = _mm_add_ps(yi, yi2); 141 #endif 142 x0v=x4v; 143 _mm_storeu_ps(y+i, yi); 144 } 145 #ifdef CUSTOM_MODES 146 for (;i<N;i++) 147 { 148 y[i] = x[i] 149 + MULT16_32_Q15(g10,x[i-T]) 150 + MULT16_32_Q15(g11,ADD32(x[i-T+1],x[i-T-1])) 151 + MULT16_32_Q15(g12,ADD32(x[i-T+2],x[i-T-2])); 152 } 153 #endif 154 } 155 156 #endif 157