Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(557)

Unified Diff: webrtc/modules/audio_processing/aec/aec_rdft_sse2.cc

Issue 2348213002: Move the aec_rdft* files to a more proper place beneath APM and make them thread-safe. (Closed)
Patch Set: Rebase Created 4 years, 2 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: webrtc/modules/audio_processing/aec/aec_rdft_sse2.cc
diff --git a/webrtc/modules/audio_processing/aec/aec_rdft_sse2.cc b/webrtc/modules/audio_processing/aec/aec_rdft_sse2.cc
deleted file mode 100644
index b4e453ff53ea2efdeb6888c5c1b22e3a92df7d16..0000000000000000000000000000000000000000
--- a/webrtc/modules/audio_processing/aec/aec_rdft_sse2.cc
+++ /dev/null
@@ -1,427 +0,0 @@
-/*
- * Copyright (c) 2011 The WebRTC project authors. All Rights Reserved.
- *
- * Use of this source code is governed by a BSD-style license
- * that can be found in the LICENSE file in the root of the source
- * tree. An additional intellectual property rights grant can be found
- * in the file PATENTS. All contributing project authors may
- * be found in the AUTHORS file in the root of the source tree.
- */
-
-#include "webrtc/modules/audio_processing/aec/aec_rdft.h"
-
-#include <emmintrin.h>
-
-static const ALIGN16_BEG float ALIGN16_END
- k_swap_sign[4] = {-1.f, 1.f, -1.f, 1.f};
-
-static void cft1st_128_SSE2(float* a) {
- const __m128 mm_swap_sign = _mm_load_ps(k_swap_sign);
- int j, k2;
-
- for (k2 = 0, j = 0; j < 128; j += 16, k2 += 4) {
- __m128 a00v = _mm_loadu_ps(&a[j + 0]);
- __m128 a04v = _mm_loadu_ps(&a[j + 4]);
- __m128 a08v = _mm_loadu_ps(&a[j + 8]);
- __m128 a12v = _mm_loadu_ps(&a[j + 12]);
- __m128 a01v = _mm_shuffle_ps(a00v, a08v, _MM_SHUFFLE(1, 0, 1, 0));
- __m128 a23v = _mm_shuffle_ps(a00v, a08v, _MM_SHUFFLE(3, 2, 3, 2));
- __m128 a45v = _mm_shuffle_ps(a04v, a12v, _MM_SHUFFLE(1, 0, 1, 0));
- __m128 a67v = _mm_shuffle_ps(a04v, a12v, _MM_SHUFFLE(3, 2, 3, 2));
-
- const __m128 wk1rv = _mm_load_ps(&rdft_wk1r[k2]);
- const __m128 wk1iv = _mm_load_ps(&rdft_wk1i[k2]);
- const __m128 wk2rv = _mm_load_ps(&rdft_wk2r[k2]);
- const __m128 wk2iv = _mm_load_ps(&rdft_wk2i[k2]);
- const __m128 wk3rv = _mm_load_ps(&rdft_wk3r[k2]);
- const __m128 wk3iv = _mm_load_ps(&rdft_wk3i[k2]);
- __m128 x0v = _mm_add_ps(a01v, a23v);
- const __m128 x1v = _mm_sub_ps(a01v, a23v);
- const __m128 x2v = _mm_add_ps(a45v, a67v);
- const __m128 x3v = _mm_sub_ps(a45v, a67v);
- __m128 x0w;
- a01v = _mm_add_ps(x0v, x2v);
- x0v = _mm_sub_ps(x0v, x2v);
- x0w = _mm_shuffle_ps(x0v, x0v, _MM_SHUFFLE(2, 3, 0, 1));
- {
- const __m128 a45_0v = _mm_mul_ps(wk2rv, x0v);
- const __m128 a45_1v = _mm_mul_ps(wk2iv, x0w);
- a45v = _mm_add_ps(a45_0v, a45_1v);
- }
- {
- __m128 a23_0v, a23_1v;
- const __m128 x3w = _mm_shuffle_ps(x3v, x3v, _MM_SHUFFLE(2, 3, 0, 1));
- const __m128 x3s = _mm_mul_ps(mm_swap_sign, x3w);
- x0v = _mm_add_ps(x1v, x3s);
- x0w = _mm_shuffle_ps(x0v, x0v, _MM_SHUFFLE(2, 3, 0, 1));
- a23_0v = _mm_mul_ps(wk1rv, x0v);
- a23_1v = _mm_mul_ps(wk1iv, x0w);
- a23v = _mm_add_ps(a23_0v, a23_1v);
-
- x0v = _mm_sub_ps(x1v, x3s);
- x0w = _mm_shuffle_ps(x0v, x0v, _MM_SHUFFLE(2, 3, 0, 1));
- }
- {
- const __m128 a67_0v = _mm_mul_ps(wk3rv, x0v);
- const __m128 a67_1v = _mm_mul_ps(wk3iv, x0w);
- a67v = _mm_add_ps(a67_0v, a67_1v);
- }
-
- a00v = _mm_shuffle_ps(a01v, a23v, _MM_SHUFFLE(1, 0, 1, 0));
- a04v = _mm_shuffle_ps(a45v, a67v, _MM_SHUFFLE(1, 0, 1, 0));
- a08v = _mm_shuffle_ps(a01v, a23v, _MM_SHUFFLE(3, 2, 3, 2));
- a12v = _mm_shuffle_ps(a45v, a67v, _MM_SHUFFLE(3, 2, 3, 2));
- _mm_storeu_ps(&a[j + 0], a00v);
- _mm_storeu_ps(&a[j + 4], a04v);
- _mm_storeu_ps(&a[j + 8], a08v);
- _mm_storeu_ps(&a[j + 12], a12v);
- }
-}
-
-static void cftmdl_128_SSE2(float* a) {
- const int l = 8;
- const __m128 mm_swap_sign = _mm_load_ps(k_swap_sign);
- int j0;
-
- __m128 wk1rv = _mm_load_ps(cftmdl_wk1r);
- for (j0 = 0; j0 < l; j0 += 2) {
- const __m128i a_00 = _mm_loadl_epi64((__m128i*)&a[j0 + 0]);
- const __m128i a_08 = _mm_loadl_epi64((__m128i*)&a[j0 + 8]);
- const __m128i a_32 = _mm_loadl_epi64((__m128i*)&a[j0 + 32]);
- const __m128i a_40 = _mm_loadl_epi64((__m128i*)&a[j0 + 40]);
- const __m128 a_00_32 = _mm_shuffle_ps(_mm_castsi128_ps(a_00),
- _mm_castsi128_ps(a_32),
- _MM_SHUFFLE(1, 0, 1, 0));
- const __m128 a_08_40 = _mm_shuffle_ps(_mm_castsi128_ps(a_08),
- _mm_castsi128_ps(a_40),
- _MM_SHUFFLE(1, 0, 1, 0));
- __m128 x0r0_0i0_0r1_x0i1 = _mm_add_ps(a_00_32, a_08_40);
- const __m128 x1r0_1i0_1r1_x1i1 = _mm_sub_ps(a_00_32, a_08_40);
-
- const __m128i a_16 = _mm_loadl_epi64((__m128i*)&a[j0 + 16]);
- const __m128i a_24 = _mm_loadl_epi64((__m128i*)&a[j0 + 24]);
- const __m128i a_48 = _mm_loadl_epi64((__m128i*)&a[j0 + 48]);
- const __m128i a_56 = _mm_loadl_epi64((__m128i*)&a[j0 + 56]);
- const __m128 a_16_48 = _mm_shuffle_ps(_mm_castsi128_ps(a_16),
- _mm_castsi128_ps(a_48),
- _MM_SHUFFLE(1, 0, 1, 0));
- const __m128 a_24_56 = _mm_shuffle_ps(_mm_castsi128_ps(a_24),
- _mm_castsi128_ps(a_56),
- _MM_SHUFFLE(1, 0, 1, 0));
- const __m128 x2r0_2i0_2r1_x2i1 = _mm_add_ps(a_16_48, a_24_56);
- const __m128 x3r0_3i0_3r1_x3i1 = _mm_sub_ps(a_16_48, a_24_56);
-
- const __m128 xx0 = _mm_add_ps(x0r0_0i0_0r1_x0i1, x2r0_2i0_2r1_x2i1);
- const __m128 xx1 = _mm_sub_ps(x0r0_0i0_0r1_x0i1, x2r0_2i0_2r1_x2i1);
-
- const __m128 x3i0_3r0_3i1_x3r1 = _mm_castsi128_ps(_mm_shuffle_epi32(
- _mm_castps_si128(x3r0_3i0_3r1_x3i1), _MM_SHUFFLE(2, 3, 0, 1)));
- const __m128 x3_swapped = _mm_mul_ps(mm_swap_sign, x3i0_3r0_3i1_x3r1);
- const __m128 x1_x3_add = _mm_add_ps(x1r0_1i0_1r1_x1i1, x3_swapped);
- const __m128 x1_x3_sub = _mm_sub_ps(x1r0_1i0_1r1_x1i1, x3_swapped);
-
- const __m128 yy0 =
- _mm_shuffle_ps(x1_x3_add, x1_x3_sub, _MM_SHUFFLE(2, 2, 2, 2));
- const __m128 yy1 =
- _mm_shuffle_ps(x1_x3_add, x1_x3_sub, _MM_SHUFFLE(3, 3, 3, 3));
- const __m128 yy2 = _mm_mul_ps(mm_swap_sign, yy1);
- const __m128 yy3 = _mm_add_ps(yy0, yy2);
- const __m128 yy4 = _mm_mul_ps(wk1rv, yy3);
-
- _mm_storel_epi64((__m128i*)&a[j0 + 0], _mm_castps_si128(xx0));
- _mm_storel_epi64(
- (__m128i*)&a[j0 + 32],
- _mm_shuffle_epi32(_mm_castps_si128(xx0), _MM_SHUFFLE(3, 2, 3, 2)));
-
- _mm_storel_epi64((__m128i*)&a[j0 + 16], _mm_castps_si128(xx1));
- _mm_storel_epi64(
- (__m128i*)&a[j0 + 48],
- _mm_shuffle_epi32(_mm_castps_si128(xx1), _MM_SHUFFLE(2, 3, 2, 3)));
- a[j0 + 48] = -a[j0 + 48];
-
- _mm_storel_epi64((__m128i*)&a[j0 + 8], _mm_castps_si128(x1_x3_add));
- _mm_storel_epi64((__m128i*)&a[j0 + 24], _mm_castps_si128(x1_x3_sub));
-
- _mm_storel_epi64((__m128i*)&a[j0 + 40], _mm_castps_si128(yy4));
- _mm_storel_epi64(
- (__m128i*)&a[j0 + 56],
- _mm_shuffle_epi32(_mm_castps_si128(yy4), _MM_SHUFFLE(2, 3, 2, 3)));
- }
-
- {
- int k = 64;
- int k1 = 2;
- int k2 = 2 * k1;
- const __m128 wk2rv = _mm_load_ps(&rdft_wk2r[k2 + 0]);
- const __m128 wk2iv = _mm_load_ps(&rdft_wk2i[k2 + 0]);
- const __m128 wk1iv = _mm_load_ps(&rdft_wk1i[k2 + 0]);
- const __m128 wk3rv = _mm_load_ps(&rdft_wk3r[k2 + 0]);
- const __m128 wk3iv = _mm_load_ps(&rdft_wk3i[k2 + 0]);
- wk1rv = _mm_load_ps(&rdft_wk1r[k2 + 0]);
- for (j0 = k; j0 < l + k; j0 += 2) {
- const __m128i a_00 = _mm_loadl_epi64((__m128i*)&a[j0 + 0]);
- const __m128i a_08 = _mm_loadl_epi64((__m128i*)&a[j0 + 8]);
- const __m128i a_32 = _mm_loadl_epi64((__m128i*)&a[j0 + 32]);
- const __m128i a_40 = _mm_loadl_epi64((__m128i*)&a[j0 + 40]);
- const __m128 a_00_32 = _mm_shuffle_ps(_mm_castsi128_ps(a_00),
- _mm_castsi128_ps(a_32),
- _MM_SHUFFLE(1, 0, 1, 0));
- const __m128 a_08_40 = _mm_shuffle_ps(_mm_castsi128_ps(a_08),
- _mm_castsi128_ps(a_40),
- _MM_SHUFFLE(1, 0, 1, 0));
- __m128 x0r0_0i0_0r1_x0i1 = _mm_add_ps(a_00_32, a_08_40);
- const __m128 x1r0_1i0_1r1_x1i1 = _mm_sub_ps(a_00_32, a_08_40);
-
- const __m128i a_16 = _mm_loadl_epi64((__m128i*)&a[j0 + 16]);
- const __m128i a_24 = _mm_loadl_epi64((__m128i*)&a[j0 + 24]);
- const __m128i a_48 = _mm_loadl_epi64((__m128i*)&a[j0 + 48]);
- const __m128i a_56 = _mm_loadl_epi64((__m128i*)&a[j0 + 56]);
- const __m128 a_16_48 = _mm_shuffle_ps(_mm_castsi128_ps(a_16),
- _mm_castsi128_ps(a_48),
- _MM_SHUFFLE(1, 0, 1, 0));
- const __m128 a_24_56 = _mm_shuffle_ps(_mm_castsi128_ps(a_24),
- _mm_castsi128_ps(a_56),
- _MM_SHUFFLE(1, 0, 1, 0));
- const __m128 x2r0_2i0_2r1_x2i1 = _mm_add_ps(a_16_48, a_24_56);
- const __m128 x3r0_3i0_3r1_x3i1 = _mm_sub_ps(a_16_48, a_24_56);
-
- const __m128 xx = _mm_add_ps(x0r0_0i0_0r1_x0i1, x2r0_2i0_2r1_x2i1);
- const __m128 xx1 = _mm_sub_ps(x0r0_0i0_0r1_x0i1, x2r0_2i0_2r1_x2i1);
- const __m128 xx2 = _mm_mul_ps(xx1, wk2rv);
- const __m128 xx3 =
- _mm_mul_ps(wk2iv,
- _mm_castsi128_ps(_mm_shuffle_epi32(
- _mm_castps_si128(xx1), _MM_SHUFFLE(2, 3, 0, 1))));
- const __m128 xx4 = _mm_add_ps(xx2, xx3);
-
- const __m128 x3i0_3r0_3i1_x3r1 = _mm_castsi128_ps(_mm_shuffle_epi32(
- _mm_castps_si128(x3r0_3i0_3r1_x3i1), _MM_SHUFFLE(2, 3, 0, 1)));
- const __m128 x3_swapped = _mm_mul_ps(mm_swap_sign, x3i0_3r0_3i1_x3r1);
- const __m128 x1_x3_add = _mm_add_ps(x1r0_1i0_1r1_x1i1, x3_swapped);
- const __m128 x1_x3_sub = _mm_sub_ps(x1r0_1i0_1r1_x1i1, x3_swapped);
-
- const __m128 xx10 = _mm_mul_ps(x1_x3_add, wk1rv);
- const __m128 xx11 = _mm_mul_ps(
- wk1iv,
- _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(x1_x3_add),
- _MM_SHUFFLE(2, 3, 0, 1))));
- const __m128 xx12 = _mm_add_ps(xx10, xx11);
-
- const __m128 xx20 = _mm_mul_ps(x1_x3_sub, wk3rv);
- const __m128 xx21 = _mm_mul_ps(
- wk3iv,
- _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(x1_x3_sub),
- _MM_SHUFFLE(2, 3, 0, 1))));
- const __m128 xx22 = _mm_add_ps(xx20, xx21);
-
- _mm_storel_epi64((__m128i*)&a[j0 + 0], _mm_castps_si128(xx));
- _mm_storel_epi64(
- (__m128i*)&a[j0 + 32],
- _mm_shuffle_epi32(_mm_castps_si128(xx), _MM_SHUFFLE(3, 2, 3, 2)));
-
- _mm_storel_epi64((__m128i*)&a[j0 + 16], _mm_castps_si128(xx4));
- _mm_storel_epi64(
- (__m128i*)&a[j0 + 48],
- _mm_shuffle_epi32(_mm_castps_si128(xx4), _MM_SHUFFLE(3, 2, 3, 2)));
-
- _mm_storel_epi64((__m128i*)&a[j0 + 8], _mm_castps_si128(xx12));
- _mm_storel_epi64(
- (__m128i*)&a[j0 + 40],
- _mm_shuffle_epi32(_mm_castps_si128(xx12), _MM_SHUFFLE(3, 2, 3, 2)));
-
- _mm_storel_epi64((__m128i*)&a[j0 + 24], _mm_castps_si128(xx22));
- _mm_storel_epi64(
- (__m128i*)&a[j0 + 56],
- _mm_shuffle_epi32(_mm_castps_si128(xx22), _MM_SHUFFLE(3, 2, 3, 2)));
- }
- }
-}
-
-static void rftfsub_128_SSE2(float* a) {
- const float* c = rdft_w + 32;
- int j1, j2, k1, k2;
- float wkr, wki, xr, xi, yr, yi;
-
- static const ALIGN16_BEG float ALIGN16_END
- k_half[4] = {0.5f, 0.5f, 0.5f, 0.5f};
- const __m128 mm_half = _mm_load_ps(k_half);
-
- // Vectorized code (four at once).
- // Note: commented number are indexes for the first iteration of the loop.
- for (j1 = 1, j2 = 2; j2 + 7 < 64; j1 += 4, j2 += 8) {
- // Load 'wk'.
- const __m128 c_j1 = _mm_loadu_ps(&c[j1]); // 1, 2, 3, 4,
- const __m128 c_k1 = _mm_loadu_ps(&c[29 - j1]); // 28, 29, 30, 31,
- const __m128 wkrt = _mm_sub_ps(mm_half, c_k1); // 28, 29, 30, 31,
- const __m128 wkr_ =
- _mm_shuffle_ps(wkrt, wkrt, _MM_SHUFFLE(0, 1, 2, 3)); // 31, 30, 29, 28,
- const __m128 wki_ = c_j1; // 1, 2, 3, 4,
- // Load and shuffle 'a'.
- const __m128 a_j2_0 = _mm_loadu_ps(&a[0 + j2]); // 2, 3, 4, 5,
- const __m128 a_j2_4 = _mm_loadu_ps(&a[4 + j2]); // 6, 7, 8, 9,
- const __m128 a_k2_0 = _mm_loadu_ps(&a[122 - j2]); // 120, 121, 122, 123,
- const __m128 a_k2_4 = _mm_loadu_ps(&a[126 - j2]); // 124, 125, 126, 127,
- const __m128 a_j2_p0 = _mm_shuffle_ps(
- a_j2_0, a_j2_4, _MM_SHUFFLE(2, 0, 2, 0)); // 2, 4, 6, 8,
- const __m128 a_j2_p1 = _mm_shuffle_ps(
- a_j2_0, a_j2_4, _MM_SHUFFLE(3, 1, 3, 1)); // 3, 5, 7, 9,
- const __m128 a_k2_p0 = _mm_shuffle_ps(
- a_k2_4, a_k2_0, _MM_SHUFFLE(0, 2, 0, 2)); // 126, 124, 122, 120,
- const __m128 a_k2_p1 = _mm_shuffle_ps(
- a_k2_4, a_k2_0, _MM_SHUFFLE(1, 3, 1, 3)); // 127, 125, 123, 121,
- // Calculate 'x'.
- const __m128 xr_ = _mm_sub_ps(a_j2_p0, a_k2_p0);
- // 2-126, 4-124, 6-122, 8-120,
- const __m128 xi_ = _mm_add_ps(a_j2_p1, a_k2_p1);
- // 3-127, 5-125, 7-123, 9-121,
- // Calculate product into 'y'.
- // yr = wkr * xr - wki * xi;
- // yi = wkr * xi + wki * xr;
- const __m128 a_ = _mm_mul_ps(wkr_, xr_);
- const __m128 b_ = _mm_mul_ps(wki_, xi_);
- const __m128 c_ = _mm_mul_ps(wkr_, xi_);
- const __m128 d_ = _mm_mul_ps(wki_, xr_);
- const __m128 yr_ = _mm_sub_ps(a_, b_); // 2-126, 4-124, 6-122, 8-120,
- const __m128 yi_ = _mm_add_ps(c_, d_); // 3-127, 5-125, 7-123, 9-121,
- // Update 'a'.
- // a[j2 + 0] -= yr;
- // a[j2 + 1] -= yi;
- // a[k2 + 0] += yr;
- // a[k2 + 1] -= yi;
- const __m128 a_j2_p0n = _mm_sub_ps(a_j2_p0, yr_); // 2, 4, 6, 8,
- const __m128 a_j2_p1n = _mm_sub_ps(a_j2_p1, yi_); // 3, 5, 7, 9,
- const __m128 a_k2_p0n = _mm_add_ps(a_k2_p0, yr_); // 126, 124, 122, 120,
- const __m128 a_k2_p1n = _mm_sub_ps(a_k2_p1, yi_); // 127, 125, 123, 121,
- // Shuffle in right order and store.
- const __m128 a_j2_0n = _mm_unpacklo_ps(a_j2_p0n, a_j2_p1n);
- // 2, 3, 4, 5,
- const __m128 a_j2_4n = _mm_unpackhi_ps(a_j2_p0n, a_j2_p1n);
- // 6, 7, 8, 9,
- const __m128 a_k2_0nt = _mm_unpackhi_ps(a_k2_p0n, a_k2_p1n);
- // 122, 123, 120, 121,
- const __m128 a_k2_4nt = _mm_unpacklo_ps(a_k2_p0n, a_k2_p1n);
- // 126, 127, 124, 125,
- const __m128 a_k2_0n = _mm_shuffle_ps(
- a_k2_0nt, a_k2_0nt, _MM_SHUFFLE(1, 0, 3, 2)); // 120, 121, 122, 123,
- const __m128 a_k2_4n = _mm_shuffle_ps(
- a_k2_4nt, a_k2_4nt, _MM_SHUFFLE(1, 0, 3, 2)); // 124, 125, 126, 127,
- _mm_storeu_ps(&a[0 + j2], a_j2_0n);
- _mm_storeu_ps(&a[4 + j2], a_j2_4n);
- _mm_storeu_ps(&a[122 - j2], a_k2_0n);
- _mm_storeu_ps(&a[126 - j2], a_k2_4n);
- }
- // Scalar code for the remaining items.
- for (; j2 < 64; j1 += 1, j2 += 2) {
- k2 = 128 - j2;
- k1 = 32 - j1;
- wkr = 0.5f - c[k1];
- wki = c[j1];
- xr = a[j2 + 0] - a[k2 + 0];
- xi = a[j2 + 1] + a[k2 + 1];
- yr = wkr * xr - wki * xi;
- yi = wkr * xi + wki * xr;
- a[j2 + 0] -= yr;
- a[j2 + 1] -= yi;
- a[k2 + 0] += yr;
- a[k2 + 1] -= yi;
- }
-}
-
-static void rftbsub_128_SSE2(float* a) {
- const float* c = rdft_w + 32;
- int j1, j2, k1, k2;
- float wkr, wki, xr, xi, yr, yi;
-
- static const ALIGN16_BEG float ALIGN16_END
- k_half[4] = {0.5f, 0.5f, 0.5f, 0.5f};
- const __m128 mm_half = _mm_load_ps(k_half);
-
- a[1] = -a[1];
- // Vectorized code (four at once).
- // Note: commented number are indexes for the first iteration of the loop.
- for (j1 = 1, j2 = 2; j2 + 7 < 64; j1 += 4, j2 += 8) {
- // Load 'wk'.
- const __m128 c_j1 = _mm_loadu_ps(&c[j1]); // 1, 2, 3, 4,
- const __m128 c_k1 = _mm_loadu_ps(&c[29 - j1]); // 28, 29, 30, 31,
- const __m128 wkrt = _mm_sub_ps(mm_half, c_k1); // 28, 29, 30, 31,
- const __m128 wkr_ =
- _mm_shuffle_ps(wkrt, wkrt, _MM_SHUFFLE(0, 1, 2, 3)); // 31, 30, 29, 28,
- const __m128 wki_ = c_j1; // 1, 2, 3, 4,
- // Load and shuffle 'a'.
- const __m128 a_j2_0 = _mm_loadu_ps(&a[0 + j2]); // 2, 3, 4, 5,
- const __m128 a_j2_4 = _mm_loadu_ps(&a[4 + j2]); // 6, 7, 8, 9,
- const __m128 a_k2_0 = _mm_loadu_ps(&a[122 - j2]); // 120, 121, 122, 123,
- const __m128 a_k2_4 = _mm_loadu_ps(&a[126 - j2]); // 124, 125, 126, 127,
- const __m128 a_j2_p0 = _mm_shuffle_ps(
- a_j2_0, a_j2_4, _MM_SHUFFLE(2, 0, 2, 0)); // 2, 4, 6, 8,
- const __m128 a_j2_p1 = _mm_shuffle_ps(
- a_j2_0, a_j2_4, _MM_SHUFFLE(3, 1, 3, 1)); // 3, 5, 7, 9,
- const __m128 a_k2_p0 = _mm_shuffle_ps(
- a_k2_4, a_k2_0, _MM_SHUFFLE(0, 2, 0, 2)); // 126, 124, 122, 120,
- const __m128 a_k2_p1 = _mm_shuffle_ps(
- a_k2_4, a_k2_0, _MM_SHUFFLE(1, 3, 1, 3)); // 127, 125, 123, 121,
- // Calculate 'x'.
- const __m128 xr_ = _mm_sub_ps(a_j2_p0, a_k2_p0);
- // 2-126, 4-124, 6-122, 8-120,
- const __m128 xi_ = _mm_add_ps(a_j2_p1, a_k2_p1);
- // 3-127, 5-125, 7-123, 9-121,
- // Calculate product into 'y'.
- // yr = wkr * xr + wki * xi;
- // yi = wkr * xi - wki * xr;
- const __m128 a_ = _mm_mul_ps(wkr_, xr_);
- const __m128 b_ = _mm_mul_ps(wki_, xi_);
- const __m128 c_ = _mm_mul_ps(wkr_, xi_);
- const __m128 d_ = _mm_mul_ps(wki_, xr_);
- const __m128 yr_ = _mm_add_ps(a_, b_); // 2-126, 4-124, 6-122, 8-120,
- const __m128 yi_ = _mm_sub_ps(c_, d_); // 3-127, 5-125, 7-123, 9-121,
- // Update 'a'.
- // a[j2 + 0] = a[j2 + 0] - yr;
- // a[j2 + 1] = yi - a[j2 + 1];
- // a[k2 + 0] = yr + a[k2 + 0];
- // a[k2 + 1] = yi - a[k2 + 1];
- const __m128 a_j2_p0n = _mm_sub_ps(a_j2_p0, yr_); // 2, 4, 6, 8,
- const __m128 a_j2_p1n = _mm_sub_ps(yi_, a_j2_p1); // 3, 5, 7, 9,
- const __m128 a_k2_p0n = _mm_add_ps(a_k2_p0, yr_); // 126, 124, 122, 120,
- const __m128 a_k2_p1n = _mm_sub_ps(yi_, a_k2_p1); // 127, 125, 123, 121,
- // Shuffle in right order and store.
- const __m128 a_j2_0n = _mm_unpacklo_ps(a_j2_p0n, a_j2_p1n);
- // 2, 3, 4, 5,
- const __m128 a_j2_4n = _mm_unpackhi_ps(a_j2_p0n, a_j2_p1n);
- // 6, 7, 8, 9,
- const __m128 a_k2_0nt = _mm_unpackhi_ps(a_k2_p0n, a_k2_p1n);
- // 122, 123, 120, 121,
- const __m128 a_k2_4nt = _mm_unpacklo_ps(a_k2_p0n, a_k2_p1n);
- // 126, 127, 124, 125,
- const __m128 a_k2_0n = _mm_shuffle_ps(
- a_k2_0nt, a_k2_0nt, _MM_SHUFFLE(1, 0, 3, 2)); // 120, 121, 122, 123,
- const __m128 a_k2_4n = _mm_shuffle_ps(
- a_k2_4nt, a_k2_4nt, _MM_SHUFFLE(1, 0, 3, 2)); // 124, 125, 126, 127,
- _mm_storeu_ps(&a[0 + j2], a_j2_0n);
- _mm_storeu_ps(&a[4 + j2], a_j2_4n);
- _mm_storeu_ps(&a[122 - j2], a_k2_0n);
- _mm_storeu_ps(&a[126 - j2], a_k2_4n);
- }
- // Scalar code for the remaining items.
- for (; j2 < 64; j1 += 1, j2 += 2) {
- k2 = 128 - j2;
- k1 = 32 - j1;
- wkr = 0.5f - c[k1];
- wki = c[j1];
- xr = a[j2 + 0] - a[k2 + 0];
- xi = a[j2 + 1] + a[k2 + 1];
- yr = wkr * xr + wki * xi;
- yi = wkr * xi - wki * xr;
- a[j2 + 0] = a[j2 + 0] - yr;
- a[j2 + 1] = yi - a[j2 + 1];
- a[k2 + 0] = yr + a[k2 + 0];
- a[k2 + 1] = yi - a[k2 + 1];
- }
- a[65] = -a[65];
-}
-
-void aec_rdft_init_sse2(void) {
- cft1st_128 = cft1st_128_SSE2;
- cftmdl_128 = cftmdl_128_SSE2;
- rftfsub_128 = rftfsub_128_SSE2;
- rftbsub_128 = rftbsub_128_SSE2;
-}
« no previous file with comments | « webrtc/modules/audio_processing/aec/aec_rdft_neon.cc ('k') | webrtc/modules/audio_processing/audio_processing.gypi » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698