1 /* 2 * Copyright (c) 2011 The WebRTC project authors. All Rights Reserved. 3 * 4 * Use of this source code is governed by a BSD-style license 5 * that can be found in the LICENSE file in the root of the source 6 * tree. An additional intellectual property rights grant can be found 7 * in the file PATENTS. All contributing project authors may 8 * be found in the AUTHORS file in the root of the source tree. 9 */ 10 11 #include "webrtc/modules/audio_processing/aec/aec_rdft.h" 12 13 #include <emmintrin.h> 14 15 static const ALIGN16_BEG float ALIGN16_END 16 k_swap_sign[4] = {-1.f, 1.f, -1.f, 1.f}; 17 18 static void cft1st_128_SSE2(float* a) { 19 const __m128 mm_swap_sign = _mm_load_ps(k_swap_sign); 20 int j, k2; 21 22 for (k2 = 0, j = 0; j < 128; j += 16, k2 += 4) { 23 __m128 a00v = _mm_loadu_ps(&a[j + 0]); 24 __m128 a04v = _mm_loadu_ps(&a[j + 4]); 25 __m128 a08v = _mm_loadu_ps(&a[j + 8]); 26 __m128 a12v = _mm_loadu_ps(&a[j + 12]); 27 __m128 a01v = _mm_shuffle_ps(a00v, a08v, _MM_SHUFFLE(1, 0, 1, 0)); 28 __m128 a23v = _mm_shuffle_ps(a00v, a08v, _MM_SHUFFLE(3, 2, 3, 2)); 29 __m128 a45v = _mm_shuffle_ps(a04v, a12v, _MM_SHUFFLE(1, 0, 1, 0)); 30 __m128 a67v = _mm_shuffle_ps(a04v, a12v, _MM_SHUFFLE(3, 2, 3, 2)); 31 32 const __m128 wk1rv = _mm_load_ps(&rdft_wk1r[k2]); 33 const __m128 wk1iv = _mm_load_ps(&rdft_wk1i[k2]); 34 const __m128 wk2rv = _mm_load_ps(&rdft_wk2r[k2]); 35 const __m128 wk2iv = _mm_load_ps(&rdft_wk2i[k2]); 36 const __m128 wk3rv = _mm_load_ps(&rdft_wk3r[k2]); 37 const __m128 wk3iv = _mm_load_ps(&rdft_wk3i[k2]); 38 __m128 x0v = _mm_add_ps(a01v, a23v); 39 const __m128 x1v = _mm_sub_ps(a01v, a23v); 40 const __m128 x2v = _mm_add_ps(a45v, a67v); 41 const __m128 x3v = _mm_sub_ps(a45v, a67v); 42 __m128 x0w; 43 a01v = _mm_add_ps(x0v, x2v); 44 x0v = _mm_sub_ps(x0v, x2v); 45 x0w = _mm_shuffle_ps(x0v, x0v, _MM_SHUFFLE(2, 3, 0, 1)); 46 { 47 const __m128 a45_0v = _mm_mul_ps(wk2rv, x0v); 48 const __m128 a45_1v = _mm_mul_ps(wk2iv, x0w); 49 a45v = _mm_add_ps(a45_0v, a45_1v); 50 } 51 { 52 __m128 a23_0v, a23_1v; 53 const __m128 x3w = _mm_shuffle_ps(x3v, x3v, _MM_SHUFFLE(2, 3, 0, 1)); 54 const __m128 x3s = _mm_mul_ps(mm_swap_sign, x3w); 55 x0v = _mm_add_ps(x1v, x3s); 56 x0w = _mm_shuffle_ps(x0v, x0v, _MM_SHUFFLE(2, 3, 0, 1)); 57 a23_0v = _mm_mul_ps(wk1rv, x0v); 58 a23_1v = _mm_mul_ps(wk1iv, x0w); 59 a23v = _mm_add_ps(a23_0v, a23_1v); 60 61 x0v = _mm_sub_ps(x1v, x3s); 62 x0w = _mm_shuffle_ps(x0v, x0v, _MM_SHUFFLE(2, 3, 0, 1)); 63 } 64 { 65 const __m128 a67_0v = _mm_mul_ps(wk3rv, x0v); 66 const __m128 a67_1v = _mm_mul_ps(wk3iv, x0w); 67 a67v = _mm_add_ps(a67_0v, a67_1v); 68 } 69 70 a00v = _mm_shuffle_ps(a01v, a23v, _MM_SHUFFLE(1, 0, 1, 0)); 71 a04v = _mm_shuffle_ps(a45v, a67v, _MM_SHUFFLE(1, 0, 1, 0)); 72 a08v = _mm_shuffle_ps(a01v, a23v, _MM_SHUFFLE(3, 2, 3, 2)); 73 a12v = _mm_shuffle_ps(a45v, a67v, _MM_SHUFFLE(3, 2, 3, 2)); 74 _mm_storeu_ps(&a[j + 0], a00v); 75 _mm_storeu_ps(&a[j + 4], a04v); 76 _mm_storeu_ps(&a[j + 8], a08v); 77 _mm_storeu_ps(&a[j + 12], a12v); 78 } 79 } 80 81 static void cftmdl_128_SSE2(float* a) { 82 const int l = 8; 83 const __m128 mm_swap_sign = _mm_load_ps(k_swap_sign); 84 int j0; 85 86 __m128 wk1rv = _mm_load_ps(cftmdl_wk1r); 87 for (j0 = 0; j0 < l; j0 += 2) { 88 const __m128i a_00 = _mm_loadl_epi64((__m128i*)&a[j0 + 0]); 89 const __m128i a_08 = _mm_loadl_epi64((__m128i*)&a[j0 + 8]); 90 const __m128i a_32 = _mm_loadl_epi64((__m128i*)&a[j0 + 32]); 91 const __m128i a_40 = _mm_loadl_epi64((__m128i*)&a[j0 + 40]); 92 const __m128 a_00_32 = _mm_shuffle_ps(_mm_castsi128_ps(a_00), 93 _mm_castsi128_ps(a_32), 94 _MM_SHUFFLE(1, 0, 1, 0)); 95 const __m128 a_08_40 = _mm_shuffle_ps(_mm_castsi128_ps(a_08), 96 _mm_castsi128_ps(a_40), 97 _MM_SHUFFLE(1, 0, 1, 0)); 98 __m128 x0r0_0i0_0r1_x0i1 = _mm_add_ps(a_00_32, a_08_40); 99 const __m128 x1r0_1i0_1r1_x1i1 = _mm_sub_ps(a_00_32, a_08_40); 100 101 const __m128i a_16 = _mm_loadl_epi64((__m128i*)&a[j0 + 16]); 102 const __m128i a_24 = _mm_loadl_epi64((__m128i*)&a[j0 + 24]); 103 const __m128i a_48 = _mm_loadl_epi64((__m128i*)&a[j0 + 48]); 104 const __m128i a_56 = _mm_loadl_epi64((__m128i*)&a[j0 + 56]); 105 const __m128 a_16_48 = _mm_shuffle_ps(_mm_castsi128_ps(a_16), 106 _mm_castsi128_ps(a_48), 107 _MM_SHUFFLE(1, 0, 1, 0)); 108 const __m128 a_24_56 = _mm_shuffle_ps(_mm_castsi128_ps(a_24), 109 _mm_castsi128_ps(a_56), 110 _MM_SHUFFLE(1, 0, 1, 0)); 111 const __m128 x2r0_2i0_2r1_x2i1 = _mm_add_ps(a_16_48, a_24_56); 112 const __m128 x3r0_3i0_3r1_x3i1 = _mm_sub_ps(a_16_48, a_24_56); 113 114 const __m128 xx0 = _mm_add_ps(x0r0_0i0_0r1_x0i1, x2r0_2i0_2r1_x2i1); 115 const __m128 xx1 = _mm_sub_ps(x0r0_0i0_0r1_x0i1, x2r0_2i0_2r1_x2i1); 116 117 const __m128 x3i0_3r0_3i1_x3r1 = _mm_castsi128_ps(_mm_shuffle_epi32( 118 _mm_castps_si128(x3r0_3i0_3r1_x3i1), _MM_SHUFFLE(2, 3, 0, 1))); 119 const __m128 x3_swapped = _mm_mul_ps(mm_swap_sign, x3i0_3r0_3i1_x3r1); 120 const __m128 x1_x3_add = _mm_add_ps(x1r0_1i0_1r1_x1i1, x3_swapped); 121 const __m128 x1_x3_sub = _mm_sub_ps(x1r0_1i0_1r1_x1i1, x3_swapped); 122 123 const __m128 yy0 = 124 _mm_shuffle_ps(x1_x3_add, x1_x3_sub, _MM_SHUFFLE(2, 2, 2, 2)); 125 const __m128 yy1 = 126 _mm_shuffle_ps(x1_x3_add, x1_x3_sub, _MM_SHUFFLE(3, 3, 3, 3)); 127 const __m128 yy2 = _mm_mul_ps(mm_swap_sign, yy1); 128 const __m128 yy3 = _mm_add_ps(yy0, yy2); 129 const __m128 yy4 = _mm_mul_ps(wk1rv, yy3); 130 131 _mm_storel_epi64((__m128i*)&a[j0 + 0], _mm_castps_si128(xx0)); 132 _mm_storel_epi64( 133 (__m128i*)&a[j0 + 32], 134 _mm_shuffle_epi32(_mm_castps_si128(xx0), _MM_SHUFFLE(3, 2, 3, 2))); 135 136 _mm_storel_epi64((__m128i*)&a[j0 + 16], _mm_castps_si128(xx1)); 137 _mm_storel_epi64( 138 (__m128i*)&a[j0 + 48], 139 _mm_shuffle_epi32(_mm_castps_si128(xx1), _MM_SHUFFLE(2, 3, 2, 3))); 140 a[j0 + 48] = -a[j0 + 48]; 141 142 _mm_storel_epi64((__m128i*)&a[j0 + 8], _mm_castps_si128(x1_x3_add)); 143 _mm_storel_epi64((__m128i*)&a[j0 + 24], _mm_castps_si128(x1_x3_sub)); 144 145 _mm_storel_epi64((__m128i*)&a[j0 + 40], _mm_castps_si128(yy4)); 146 _mm_storel_epi64( 147 (__m128i*)&a[j0 + 56], 148 _mm_shuffle_epi32(_mm_castps_si128(yy4), _MM_SHUFFLE(2, 3, 2, 3))); 149 } 150 151 { 152 int k = 64; 153 int k1 = 2; 154 int k2 = 2 * k1; 155 const __m128 wk2rv = _mm_load_ps(&rdft_wk2r[k2 + 0]); 156 const __m128 wk2iv = _mm_load_ps(&rdft_wk2i[k2 + 0]); 157 const __m128 wk1iv = _mm_load_ps(&rdft_wk1i[k2 + 0]); 158 const __m128 wk3rv = _mm_load_ps(&rdft_wk3r[k2 + 0]); 159 const __m128 wk3iv = _mm_load_ps(&rdft_wk3i[k2 + 0]); 160 wk1rv = _mm_load_ps(&rdft_wk1r[k2 + 0]); 161 for (j0 = k; j0 < l + k; j0 += 2) { 162 const __m128i a_00 = _mm_loadl_epi64((__m128i*)&a[j0 + 0]); 163 const __m128i a_08 = _mm_loadl_epi64((__m128i*)&a[j0 + 8]); 164 const __m128i a_32 = _mm_loadl_epi64((__m128i*)&a[j0 + 32]); 165 const __m128i a_40 = _mm_loadl_epi64((__m128i*)&a[j0 + 40]); 166 const __m128 a_00_32 = _mm_shuffle_ps(_mm_castsi128_ps(a_00), 167 _mm_castsi128_ps(a_32), 168 _MM_SHUFFLE(1, 0, 1, 0)); 169 const __m128 a_08_40 = _mm_shuffle_ps(_mm_castsi128_ps(a_08), 170 _mm_castsi128_ps(a_40), 171 _MM_SHUFFLE(1, 0, 1, 0)); 172 __m128 x0r0_0i0_0r1_x0i1 = _mm_add_ps(a_00_32, a_08_40); 173 const __m128 x1r0_1i0_1r1_x1i1 = _mm_sub_ps(a_00_32, a_08_40); 174 175 const __m128i a_16 = _mm_loadl_epi64((__m128i*)&a[j0 + 16]); 176 const __m128i a_24 = _mm_loadl_epi64((__m128i*)&a[j0 + 24]); 177 const __m128i a_48 = _mm_loadl_epi64((__m128i*)&a[j0 + 48]); 178 const __m128i a_56 = _mm_loadl_epi64((__m128i*)&a[j0 + 56]); 179 const __m128 a_16_48 = _mm_shuffle_ps(_mm_castsi128_ps(a_16), 180 _mm_castsi128_ps(a_48), 181 _MM_SHUFFLE(1, 0, 1, 0)); 182 const __m128 a_24_56 = _mm_shuffle_ps(_mm_castsi128_ps(a_24), 183 _mm_castsi128_ps(a_56), 184 _MM_SHUFFLE(1, 0, 1, 0)); 185 const __m128 x2r0_2i0_2r1_x2i1 = _mm_add_ps(a_16_48, a_24_56); 186 const __m128 x3r0_3i0_3r1_x3i1 = _mm_sub_ps(a_16_48, a_24_56); 187 188 const __m128 xx = _mm_add_ps(x0r0_0i0_0r1_x0i1, x2r0_2i0_2r1_x2i1); 189 const __m128 xx1 = _mm_sub_ps(x0r0_0i0_0r1_x0i1, x2r0_2i0_2r1_x2i1); 190 const __m128 xx2 = _mm_mul_ps(xx1, wk2rv); 191 const __m128 xx3 = 192 _mm_mul_ps(wk2iv, 193 _mm_castsi128_ps(_mm_shuffle_epi32( 194 _mm_castps_si128(xx1), _MM_SHUFFLE(2, 3, 0, 1)))); 195 const __m128 xx4 = _mm_add_ps(xx2, xx3); 196 197 const __m128 x3i0_3r0_3i1_x3r1 = _mm_castsi128_ps(_mm_shuffle_epi32( 198 _mm_castps_si128(x3r0_3i0_3r1_x3i1), _MM_SHUFFLE(2, 3, 0, 1))); 199 const __m128 x3_swapped = _mm_mul_ps(mm_swap_sign, x3i0_3r0_3i1_x3r1); 200 const __m128 x1_x3_add = _mm_add_ps(x1r0_1i0_1r1_x1i1, x3_swapped); 201 const __m128 x1_x3_sub = _mm_sub_ps(x1r0_1i0_1r1_x1i1, x3_swapped); 202 203 const __m128 xx10 = _mm_mul_ps(x1_x3_add, wk1rv); 204 const __m128 xx11 = _mm_mul_ps( 205 wk1iv, 206 _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(x1_x3_add), 207 _MM_SHUFFLE(2, 3, 0, 1)))); 208 const __m128 xx12 = _mm_add_ps(xx10, xx11); 209 210 const __m128 xx20 = _mm_mul_ps(x1_x3_sub, wk3rv); 211 const __m128 xx21 = _mm_mul_ps( 212 wk3iv, 213 _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(x1_x3_sub), 214 _MM_SHUFFLE(2, 3, 0, 1)))); 215 const __m128 xx22 = _mm_add_ps(xx20, xx21); 216 217 _mm_storel_epi64((__m128i*)&a[j0 + 0], _mm_castps_si128(xx)); 218 _mm_storel_epi64( 219 (__m128i*)&a[j0 + 32], 220 _mm_shuffle_epi32(_mm_castps_si128(xx), _MM_SHUFFLE(3, 2, 3, 2))); 221 222 _mm_storel_epi64((__m128i*)&a[j0 + 16], _mm_castps_si128(xx4)); 223 _mm_storel_epi64( 224 (__m128i*)&a[j0 + 48], 225 _mm_shuffle_epi32(_mm_castps_si128(xx4), _MM_SHUFFLE(3, 2, 3, 2))); 226 227 _mm_storel_epi64((__m128i*)&a[j0 + 8], _mm_castps_si128(xx12)); 228 _mm_storel_epi64( 229 (__m128i*)&a[j0 + 40], 230 _mm_shuffle_epi32(_mm_castps_si128(xx12), _MM_SHUFFLE(3, 2, 3, 2))); 231 232 _mm_storel_epi64((__m128i*)&a[j0 + 24], _mm_castps_si128(xx22)); 233 _mm_storel_epi64( 234 (__m128i*)&a[j0 + 56], 235 _mm_shuffle_epi32(_mm_castps_si128(xx22), _MM_SHUFFLE(3, 2, 3, 2))); 236 } 237 } 238 } 239 240 static void rftfsub_128_SSE2(float* a) { 241 const float* c = rdft_w + 32; 242 int j1, j2, k1, k2; 243 float wkr, wki, xr, xi, yr, yi; 244 245 static const ALIGN16_BEG float ALIGN16_END 246 k_half[4] = {0.5f, 0.5f, 0.5f, 0.5f}; 247 const __m128 mm_half = _mm_load_ps(k_half); 248 249 // Vectorized code (four at once). 250 // Note: commented number are indexes for the first iteration of the loop. 251 for (j1 = 1, j2 = 2; j2 + 7 < 64; j1 += 4, j2 += 8) { 252 // Load 'wk'. 253 const __m128 c_j1 = _mm_loadu_ps(&c[j1]); // 1, 2, 3, 4, 254 const __m128 c_k1 = _mm_loadu_ps(&c[29 - j1]); // 28, 29, 30, 31, 255 const __m128 wkrt = _mm_sub_ps(mm_half, c_k1); // 28, 29, 30, 31, 256 const __m128 wkr_ = 257 _mm_shuffle_ps(wkrt, wkrt, _MM_SHUFFLE(0, 1, 2, 3)); // 31, 30, 29, 28, 258 const __m128 wki_ = c_j1; // 1, 2, 3, 4, 259 // Load and shuffle 'a'. 260 const __m128 a_j2_0 = _mm_loadu_ps(&a[0 + j2]); // 2, 3, 4, 5, 261 const __m128 a_j2_4 = _mm_loadu_ps(&a[4 + j2]); // 6, 7, 8, 9, 262 const __m128 a_k2_0 = _mm_loadu_ps(&a[122 - j2]); // 120, 121, 122, 123, 263 const __m128 a_k2_4 = _mm_loadu_ps(&a[126 - j2]); // 124, 125, 126, 127, 264 const __m128 a_j2_p0 = _mm_shuffle_ps( 265 a_j2_0, a_j2_4, _MM_SHUFFLE(2, 0, 2, 0)); // 2, 4, 6, 8, 266 const __m128 a_j2_p1 = _mm_shuffle_ps( 267 a_j2_0, a_j2_4, _MM_SHUFFLE(3, 1, 3, 1)); // 3, 5, 7, 9, 268 const __m128 a_k2_p0 = _mm_shuffle_ps( 269 a_k2_4, a_k2_0, _MM_SHUFFLE(0, 2, 0, 2)); // 126, 124, 122, 120, 270 const __m128 a_k2_p1 = _mm_shuffle_ps( 271 a_k2_4, a_k2_0, _MM_SHUFFLE(1, 3, 1, 3)); // 127, 125, 123, 121, 272 // Calculate 'x'. 273 const __m128 xr_ = _mm_sub_ps(a_j2_p0, a_k2_p0); 274 // 2-126, 4-124, 6-122, 8-120, 275 const __m128 xi_ = _mm_add_ps(a_j2_p1, a_k2_p1); 276 // 3-127, 5-125, 7-123, 9-121, 277 // Calculate product into 'y'. 278 // yr = wkr * xr - wki * xi; 279 // yi = wkr * xi + wki * xr; 280 const __m128 a_ = _mm_mul_ps(wkr_, xr_); 281 const __m128 b_ = _mm_mul_ps(wki_, xi_); 282 const __m128 c_ = _mm_mul_ps(wkr_, xi_); 283 const __m128 d_ = _mm_mul_ps(wki_, xr_); 284 const __m128 yr_ = _mm_sub_ps(a_, b_); // 2-126, 4-124, 6-122, 8-120, 285 const __m128 yi_ = _mm_add_ps(c_, d_); // 3-127, 5-125, 7-123, 9-121, 286 // Update 'a'. 287 // a[j2 + 0] -= yr; 288 // a[j2 + 1] -= yi; 289 // a[k2 + 0] += yr; 290 // a[k2 + 1] -= yi; 291 const __m128 a_j2_p0n = _mm_sub_ps(a_j2_p0, yr_); // 2, 4, 6, 8, 292 const __m128 a_j2_p1n = _mm_sub_ps(a_j2_p1, yi_); // 3, 5, 7, 9, 293 const __m128 a_k2_p0n = _mm_add_ps(a_k2_p0, yr_); // 126, 124, 122, 120, 294 const __m128 a_k2_p1n = _mm_sub_ps(a_k2_p1, yi_); // 127, 125, 123, 121, 295 // Shuffle in right order and store. 296 const __m128 a_j2_0n = _mm_unpacklo_ps(a_j2_p0n, a_j2_p1n); 297 // 2, 3, 4, 5, 298 const __m128 a_j2_4n = _mm_unpackhi_ps(a_j2_p0n, a_j2_p1n); 299 // 6, 7, 8, 9, 300 const __m128 a_k2_0nt = _mm_unpackhi_ps(a_k2_p0n, a_k2_p1n); 301 // 122, 123, 120, 121, 302 const __m128 a_k2_4nt = _mm_unpacklo_ps(a_k2_p0n, a_k2_p1n); 303 // 126, 127, 124, 125, 304 const __m128 a_k2_0n = _mm_shuffle_ps( 305 a_k2_0nt, a_k2_0nt, _MM_SHUFFLE(1, 0, 3, 2)); // 120, 121, 122, 123, 306 const __m128 a_k2_4n = _mm_shuffle_ps( 307 a_k2_4nt, a_k2_4nt, _MM_SHUFFLE(1, 0, 3, 2)); // 124, 125, 126, 127, 308 _mm_storeu_ps(&a[0 + j2], a_j2_0n); 309 _mm_storeu_ps(&a[4 + j2], a_j2_4n); 310 _mm_storeu_ps(&a[122 - j2], a_k2_0n); 311 _mm_storeu_ps(&a[126 - j2], a_k2_4n); 312 } 313 // Scalar code for the remaining items. 314 for (; j2 < 64; j1 += 1, j2 += 2) { 315 k2 = 128 - j2; 316 k1 = 32 - j1; 317 wkr = 0.5f - c[k1]; 318 wki = c[j1]; 319 xr = a[j2 + 0] - a[k2 + 0]; 320 xi = a[j2 + 1] + a[k2 + 1]; 321 yr = wkr * xr - wki * xi; 322 yi = wkr * xi + wki * xr; 323 a[j2 + 0] -= yr; 324 a[j2 + 1] -= yi; 325 a[k2 + 0] += yr; 326 a[k2 + 1] -= yi; 327 } 328 } 329 330 static void rftbsub_128_SSE2(float* a) { 331 const float* c = rdft_w + 32; 332 int j1, j2, k1, k2; 333 float wkr, wki, xr, xi, yr, yi; 334 335 static const ALIGN16_BEG float ALIGN16_END 336 k_half[4] = {0.5f, 0.5f, 0.5f, 0.5f}; 337 const __m128 mm_half = _mm_load_ps(k_half); 338 339 a[1] = -a[1]; 340 // Vectorized code (four at once). 341 // Note: commented number are indexes for the first iteration of the loop. 342 for (j1 = 1, j2 = 2; j2 + 7 < 64; j1 += 4, j2 += 8) { 343 // Load 'wk'. 344 const __m128 c_j1 = _mm_loadu_ps(&c[j1]); // 1, 2, 3, 4, 345 const __m128 c_k1 = _mm_loadu_ps(&c[29 - j1]); // 28, 29, 30, 31, 346 const __m128 wkrt = _mm_sub_ps(mm_half, c_k1); // 28, 29, 30, 31, 347 const __m128 wkr_ = 348 _mm_shuffle_ps(wkrt, wkrt, _MM_SHUFFLE(0, 1, 2, 3)); // 31, 30, 29, 28, 349 const __m128 wki_ = c_j1; // 1, 2, 3, 4, 350 // Load and shuffle 'a'. 351 const __m128 a_j2_0 = _mm_loadu_ps(&a[0 + j2]); // 2, 3, 4, 5, 352 const __m128 a_j2_4 = _mm_loadu_ps(&a[4 + j2]); // 6, 7, 8, 9, 353 const __m128 a_k2_0 = _mm_loadu_ps(&a[122 - j2]); // 120, 121, 122, 123, 354 const __m128 a_k2_4 = _mm_loadu_ps(&a[126 - j2]); // 124, 125, 126, 127, 355 const __m128 a_j2_p0 = _mm_shuffle_ps( 356 a_j2_0, a_j2_4, _MM_SHUFFLE(2, 0, 2, 0)); // 2, 4, 6, 8, 357 const __m128 a_j2_p1 = _mm_shuffle_ps( 358 a_j2_0, a_j2_4, _MM_SHUFFLE(3, 1, 3, 1)); // 3, 5, 7, 9, 359 const __m128 a_k2_p0 = _mm_shuffle_ps( 360 a_k2_4, a_k2_0, _MM_SHUFFLE(0, 2, 0, 2)); // 126, 124, 122, 120, 361 const __m128 a_k2_p1 = _mm_shuffle_ps( 362 a_k2_4, a_k2_0, _MM_SHUFFLE(1, 3, 1, 3)); // 127, 125, 123, 121, 363 // Calculate 'x'. 364 const __m128 xr_ = _mm_sub_ps(a_j2_p0, a_k2_p0); 365 // 2-126, 4-124, 6-122, 8-120, 366 const __m128 xi_ = _mm_add_ps(a_j2_p1, a_k2_p1); 367 // 3-127, 5-125, 7-123, 9-121, 368 // Calculate product into 'y'. 369 // yr = wkr * xr + wki * xi; 370 // yi = wkr * xi - wki * xr; 371 const __m128 a_ = _mm_mul_ps(wkr_, xr_); 372 const __m128 b_ = _mm_mul_ps(wki_, xi_); 373 const __m128 c_ = _mm_mul_ps(wkr_, xi_); 374 const __m128 d_ = _mm_mul_ps(wki_, xr_); 375 const __m128 yr_ = _mm_add_ps(a_, b_); // 2-126, 4-124, 6-122, 8-120, 376 const __m128 yi_ = _mm_sub_ps(c_, d_); // 3-127, 5-125, 7-123, 9-121, 377 // Update 'a'. 378 // a[j2 + 0] = a[j2 + 0] - yr; 379 // a[j2 + 1] = yi - a[j2 + 1]; 380 // a[k2 + 0] = yr + a[k2 + 0]; 381 // a[k2 + 1] = yi - a[k2 + 1]; 382 const __m128 a_j2_p0n = _mm_sub_ps(a_j2_p0, yr_); // 2, 4, 6, 8, 383 const __m128 a_j2_p1n = _mm_sub_ps(yi_, a_j2_p1); // 3, 5, 7, 9, 384 const __m128 a_k2_p0n = _mm_add_ps(a_k2_p0, yr_); // 126, 124, 122, 120, 385 const __m128 a_k2_p1n = _mm_sub_ps(yi_, a_k2_p1); // 127, 125, 123, 121, 386 // Shuffle in right order and store. 387 const __m128 a_j2_0n = _mm_unpacklo_ps(a_j2_p0n, a_j2_p1n); 388 // 2, 3, 4, 5, 389 const __m128 a_j2_4n = _mm_unpackhi_ps(a_j2_p0n, a_j2_p1n); 390 // 6, 7, 8, 9, 391 const __m128 a_k2_0nt = _mm_unpackhi_ps(a_k2_p0n, a_k2_p1n); 392 // 122, 123, 120, 121, 393 const __m128 a_k2_4nt = _mm_unpacklo_ps(a_k2_p0n, a_k2_p1n); 394 // 126, 127, 124, 125, 395 const __m128 a_k2_0n = _mm_shuffle_ps( 396 a_k2_0nt, a_k2_0nt, _MM_SHUFFLE(1, 0, 3, 2)); // 120, 121, 122, 123, 397 const __m128 a_k2_4n = _mm_shuffle_ps( 398 a_k2_4nt, a_k2_4nt, _MM_SHUFFLE(1, 0, 3, 2)); // 124, 125, 126, 127, 399 _mm_storeu_ps(&a[0 + j2], a_j2_0n); 400 _mm_storeu_ps(&a[4 + j2], a_j2_4n); 401 _mm_storeu_ps(&a[122 - j2], a_k2_0n); 402 _mm_storeu_ps(&a[126 - j2], a_k2_4n); 403 } 404 // Scalar code for the remaining items. 405 for (; j2 < 64; j1 += 1, j2 += 2) { 406 k2 = 128 - j2; 407 k1 = 32 - j1; 408 wkr = 0.5f - c[k1]; 409 wki = c[j1]; 410 xr = a[j2 + 0] - a[k2 + 0]; 411 xi = a[j2 + 1] + a[k2 + 1]; 412 yr = wkr * xr + wki * xi; 413 yi = wkr * xi - wki * xr; 414 a[j2 + 0] = a[j2 + 0] - yr; 415 a[j2 + 1] = yi - a[j2 + 1]; 416 a[k2 + 0] = yr + a[k2 + 0]; 417 a[k2 + 1] = yi - a[k2 + 1]; 418 } 419 a[65] = -a[65]; 420 } 421 422 void aec_rdft_init_sse2(void) { 423 cft1st_128 = cft1st_128_SSE2; 424 cftmdl_128 = cftmdl_128_SSE2; 425 rftfsub_128 = rftfsub_128_SSE2; 426 rftbsub_128 = rftbsub_128_SSE2; 427 } 428