Index: webrtc/modules/audio_processing/aec/aec_core.c |
diff --git a/webrtc/modules/audio_processing/aec/aec_core.c b/webrtc/modules/audio_processing/aec/aec_core.c |
deleted file mode 100644 |
index 76a33cec165e373530641fd7085e80813481834d..0000000000000000000000000000000000000000 |
--- a/webrtc/modules/audio_processing/aec/aec_core.c |
+++ /dev/null |
@@ -1,1896 +0,0 @@ |
-/* |
- * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved. |
- * |
- * Use of this source code is governed by a BSD-style license |
- * that can be found in the LICENSE file in the root of the source |
- * tree. An additional intellectual property rights grant can be found |
- * in the file PATENTS. All contributing project authors may |
- * be found in the AUTHORS file in the root of the source tree. |
- */ |
- |
-/* |
- * The core AEC algorithm, which is presented with time-aligned signals. |
- */ |
- |
-#include "webrtc/modules/audio_processing/aec/aec_core.h" |
- |
-#ifdef WEBRTC_AEC_DEBUG_DUMP |
-#include <stdio.h> |
-#endif |
- |
-#include <assert.h> |
-#include <math.h> |
-#include <stddef.h> // size_t |
-#include <stdlib.h> |
-#include <string.h> |
- |
-#include "webrtc/common_audio/ring_buffer.h" |
-#include "webrtc/common_audio/signal_processing/include/signal_processing_library.h" |
-#include "webrtc/modules/audio_processing/aec/aec_common.h" |
-#include "webrtc/modules/audio_processing/aec/aec_core_internal.h" |
-#include "webrtc/modules/audio_processing/aec/aec_rdft.h" |
-#include "webrtc/modules/audio_processing/logging/aec_logging.h" |
-#include "webrtc/modules/audio_processing/utility/delay_estimator_wrapper.h" |
-#include "webrtc/system_wrappers/include/cpu_features_wrapper.h" |
-#include "webrtc/typedefs.h" |
- |
-// Buffer size (samples) |
-static const size_t kBufSizePartitions = 250; // 1 second of audio in 16 kHz. |
- |
-// Metrics |
-static const int subCountLen = 4; |
-static const int countLen = 50; |
-static const int kDelayMetricsAggregationWindow = 1250; // 5 seconds at 16 kHz. |
- |
-// Quantities to control H band scaling for SWB input |
-static const float cnScaleHband = |
- (float)0.4; // scale for comfort noise in H band |
-// Initial bin for averaging nlp gain in low band |
-static const int freqAvgIc = PART_LEN / 2; |
- |
-// Matlab code to produce table: |
-// win = sqrt(hanning(63)); win = [0 ; win(1:32)]; |
-// fprintf(1, '\t%.14f, %.14f, %.14f,\n', win); |
-ALIGN16_BEG const float ALIGN16_END WebRtcAec_sqrtHanning[65] = { |
- 0.00000000000000f, 0.02454122852291f, 0.04906767432742f, 0.07356456359967f, |
- 0.09801714032956f, 0.12241067519922f, 0.14673047445536f, 0.17096188876030f, |
- 0.19509032201613f, 0.21910124015687f, 0.24298017990326f, 0.26671275747490f, |
- 0.29028467725446f, 0.31368174039889f, 0.33688985339222f, 0.35989503653499f, |
- 0.38268343236509f, 0.40524131400499f, 0.42755509343028f, 0.44961132965461f, |
- 0.47139673682600f, 0.49289819222978f, 0.51410274419322f, 0.53499761988710f, |
- 0.55557023301960f, 0.57580819141785f, 0.59569930449243f, 0.61523159058063f, |
- 0.63439328416365f, 0.65317284295378f, 0.67155895484702f, 0.68954054473707f, |
- 0.70710678118655f, 0.72424708295147f, 0.74095112535496f, 0.75720884650648f, |
- 0.77301045336274f, 0.78834642762661f, 0.80320753148064f, 0.81758481315158f, |
- 0.83146961230255f, 0.84485356524971f, 0.85772861000027f, 0.87008699110871f, |
- 0.88192126434835f, 0.89322430119552f, 0.90398929312344f, 0.91420975570353f, |
- 0.92387953251129f, 0.93299279883474f, 0.94154406518302f, 0.94952818059304f, |
- 0.95694033573221f, 0.96377606579544f, 0.97003125319454f, 0.97570213003853f, |
- 0.98078528040323f, 0.98527764238894f, 0.98917650996478f, 0.99247953459871f, |
- 0.99518472667220f, 0.99729045667869f, 0.99879545620517f, 0.99969881869620f, |
- 1.00000000000000f}; |
- |
-// Matlab code to produce table: |
-// weightCurve = [0 ; 0.3 * sqrt(linspace(0,1,64))' + 0.1]; |
-// fprintf(1, '\t%.4f, %.4f, %.4f, %.4f, %.4f, %.4f,\n', weightCurve); |
-ALIGN16_BEG const float ALIGN16_END WebRtcAec_weightCurve[65] = { |
- 0.0000f, 0.1000f, 0.1378f, 0.1535f, 0.1655f, 0.1756f, 0.1845f, 0.1926f, |
- 0.2000f, 0.2069f, 0.2134f, 0.2195f, 0.2254f, 0.2309f, 0.2363f, 0.2414f, |
- 0.2464f, 0.2512f, 0.2558f, 0.2604f, 0.2648f, 0.2690f, 0.2732f, 0.2773f, |
- 0.2813f, 0.2852f, 0.2890f, 0.2927f, 0.2964f, 0.3000f, 0.3035f, 0.3070f, |
- 0.3104f, 0.3138f, 0.3171f, 0.3204f, 0.3236f, 0.3268f, 0.3299f, 0.3330f, |
- 0.3360f, 0.3390f, 0.3420f, 0.3449f, 0.3478f, 0.3507f, 0.3535f, 0.3563f, |
- 0.3591f, 0.3619f, 0.3646f, 0.3673f, 0.3699f, 0.3726f, 0.3752f, 0.3777f, |
- 0.3803f, 0.3828f, 0.3854f, 0.3878f, 0.3903f, 0.3928f, 0.3952f, 0.3976f, |
- 0.4000f}; |
- |
-// Matlab code to produce table: |
-// overDriveCurve = [sqrt(linspace(0,1,65))' + 1]; |
-// fprintf(1, '\t%.4f, %.4f, %.4f, %.4f, %.4f, %.4f,\n', overDriveCurve); |
-ALIGN16_BEG const float ALIGN16_END WebRtcAec_overDriveCurve[65] = { |
- 1.0000f, 1.1250f, 1.1768f, 1.2165f, 1.2500f, 1.2795f, 1.3062f, 1.3307f, |
- 1.3536f, 1.3750f, 1.3953f, 1.4146f, 1.4330f, 1.4507f, 1.4677f, 1.4841f, |
- 1.5000f, 1.5154f, 1.5303f, 1.5449f, 1.5590f, 1.5728f, 1.5863f, 1.5995f, |
- 1.6124f, 1.6250f, 1.6374f, 1.6495f, 1.6614f, 1.6731f, 1.6847f, 1.6960f, |
- 1.7071f, 1.7181f, 1.7289f, 1.7395f, 1.7500f, 1.7603f, 1.7706f, 1.7806f, |
- 1.7906f, 1.8004f, 1.8101f, 1.8197f, 1.8292f, 1.8385f, 1.8478f, 1.8570f, |
- 1.8660f, 1.8750f, 1.8839f, 1.8927f, 1.9014f, 1.9100f, 1.9186f, 1.9270f, |
- 1.9354f, 1.9437f, 1.9520f, 1.9601f, 1.9682f, 1.9763f, 1.9843f, 1.9922f, |
- 2.0000f}; |
- |
-// Delay Agnostic AEC parameters, still under development and may change. |
-static const float kDelayQualityThresholdMax = 0.07f; |
-static const float kDelayQualityThresholdMin = 0.01f; |
-static const int kInitialShiftOffset = 5; |
-#if !defined(WEBRTC_ANDROID) |
-static const int kDelayCorrectionStart = 1500; // 10 ms chunks |
-#endif |
- |
-// Target suppression levels for nlp modes. |
-// log{0.001, 0.00001, 0.00000001} |
-static const float kTargetSupp[3] = {-6.9f, -11.5f, -18.4f}; |
- |
-// Two sets of parameters, one for the extended filter mode. |
-static const float kExtendedMinOverDrive[3] = {3.0f, 6.0f, 15.0f}; |
-static const float kNormalMinOverDrive[3] = {1.0f, 2.0f, 5.0f}; |
-const float WebRtcAec_kExtendedSmoothingCoefficients[2][2] = {{0.9f, 0.1f}, |
- {0.92f, 0.08f}}; |
-const float WebRtcAec_kNormalSmoothingCoefficients[2][2] = {{0.9f, 0.1f}, |
- {0.93f, 0.07f}}; |
- |
-// Number of partitions forming the NLP's "preferred" bands. |
-enum { kPrefBandSize = 24 }; |
- |
-#ifdef WEBRTC_AEC_DEBUG_DUMP |
-extern int webrtc_aec_instance_count; |
-#endif |
- |
-WebRtcAecFilterFar WebRtcAec_FilterFar; |
-WebRtcAecScaleErrorSignal WebRtcAec_ScaleErrorSignal; |
-WebRtcAecFilterAdaptation WebRtcAec_FilterAdaptation; |
-WebRtcAecOverdriveAndSuppress WebRtcAec_OverdriveAndSuppress; |
-WebRtcAecComfortNoise WebRtcAec_ComfortNoise; |
-WebRtcAecSubBandCoherence WebRtcAec_SubbandCoherence; |
-WebRtcAecStoreAsComplex WebRtcAec_StoreAsComplex; |
-WebRtcAecPartitionDelay WebRtcAec_PartitionDelay; |
-WebRtcAecWindowData WebRtcAec_WindowData; |
- |
-__inline static float MulRe(float aRe, float aIm, float bRe, float bIm) { |
- return aRe * bRe - aIm * bIm; |
-} |
- |
-__inline static float MulIm(float aRe, float aIm, float bRe, float bIm) { |
- return aRe * bIm + aIm * bRe; |
-} |
- |
-static int CmpFloat(const void* a, const void* b) { |
- const float* da = (const float*)a; |
- const float* db = (const float*)b; |
- |
- return (*da > *db) - (*da < *db); |
-} |
- |
-static void FilterFar(int num_partitions, |
- int x_fft_buf_block_pos, |
- float x_fft_buf[2][kExtendedNumPartitions * PART_LEN1], |
- float h_fft_buf[2][kExtendedNumPartitions * PART_LEN1], |
- float y_fft[2][PART_LEN1]) { |
- int i; |
- for (i = 0; i < num_partitions; i++) { |
- int j; |
- int xPos = (i + x_fft_buf_block_pos) * PART_LEN1; |
- int pos = i * PART_LEN1; |
- // Check for wrap |
- if (i + x_fft_buf_block_pos >= num_partitions) { |
- xPos -= num_partitions * (PART_LEN1); |
- } |
- |
- for (j = 0; j < PART_LEN1; j++) { |
- y_fft[0][j] += MulRe(x_fft_buf[0][xPos + j], x_fft_buf[1][xPos + j], |
- h_fft_buf[0][pos + j], h_fft_buf[1][pos + j]); |
- y_fft[1][j] += MulIm(x_fft_buf[0][xPos + j], x_fft_buf[1][xPos + j], |
- h_fft_buf[0][pos + j], h_fft_buf[1][pos + j]); |
- } |
- } |
-} |
- |
-static void ScaleErrorSignal(int extended_filter_enabled, |
- float normal_mu, |
- float normal_error_threshold, |
- float x_pow[PART_LEN1], |
- float ef[2][PART_LEN1]) { |
- const float mu = extended_filter_enabled ? kExtendedMu : normal_mu; |
- const float error_threshold = extended_filter_enabled |
- ? kExtendedErrorThreshold |
- : normal_error_threshold; |
- int i; |
- float abs_ef; |
- for (i = 0; i < (PART_LEN1); i++) { |
- ef[0][i] /= (x_pow[i] + 1e-10f); |
- ef[1][i] /= (x_pow[i] + 1e-10f); |
- abs_ef = sqrtf(ef[0][i] * ef[0][i] + ef[1][i] * ef[1][i]); |
- |
- if (abs_ef > error_threshold) { |
- abs_ef = error_threshold / (abs_ef + 1e-10f); |
- ef[0][i] *= abs_ef; |
- ef[1][i] *= abs_ef; |
- } |
- |
- // Stepsize factor |
- ef[0][i] *= mu; |
- ef[1][i] *= mu; |
- } |
-} |
- |
-static void FilterAdaptation( |
- int num_partitions, |
- int x_fft_buf_block_pos, |
- float x_fft_buf[2][kExtendedNumPartitions * PART_LEN1], |
- float e_fft[2][PART_LEN1], |
- float h_fft_buf[2][kExtendedNumPartitions * PART_LEN1]) { |
- int i, j; |
- float fft[PART_LEN2]; |
- for (i = 0; i < num_partitions; i++) { |
- int xPos = (i + x_fft_buf_block_pos) * (PART_LEN1); |
- int pos; |
- // Check for wrap |
- if (i + x_fft_buf_block_pos >= num_partitions) { |
- xPos -= num_partitions * PART_LEN1; |
- } |
- |
- pos = i * PART_LEN1; |
- |
- for (j = 0; j < PART_LEN; j++) { |
- fft[2 * j] = MulRe(x_fft_buf[0][xPos + j], -x_fft_buf[1][xPos + j], |
- e_fft[0][j], e_fft[1][j]); |
- fft[2 * j + 1] = MulIm(x_fft_buf[0][xPos + j], -x_fft_buf[1][xPos + j], |
- e_fft[0][j], e_fft[1][j]); |
- } |
- fft[1] = |
- MulRe(x_fft_buf[0][xPos + PART_LEN], -x_fft_buf[1][xPos + PART_LEN], |
- e_fft[0][PART_LEN], e_fft[1][PART_LEN]); |
- |
- aec_rdft_inverse_128(fft); |
- memset(fft + PART_LEN, 0, sizeof(float) * PART_LEN); |
- |
- // fft scaling |
- { |
- float scale = 2.0f / PART_LEN2; |
- for (j = 0; j < PART_LEN; j++) { |
- fft[j] *= scale; |
- } |
- } |
- aec_rdft_forward_128(fft); |
- |
- h_fft_buf[0][pos] += fft[0]; |
- h_fft_buf[0][pos + PART_LEN] += fft[1]; |
- |
- for (j = 1; j < PART_LEN; j++) { |
- h_fft_buf[0][pos + j] += fft[2 * j]; |
- h_fft_buf[1][pos + j] += fft[2 * j + 1]; |
- } |
- } |
-} |
- |
-static void OverdriveAndSuppress(AecCore* aec, |
- float hNl[PART_LEN1], |
- const float hNlFb, |
- float efw[2][PART_LEN1]) { |
- int i; |
- for (i = 0; i < PART_LEN1; i++) { |
- // Weight subbands |
- if (hNl[i] > hNlFb) { |
- hNl[i] = WebRtcAec_weightCurve[i] * hNlFb + |
- (1 - WebRtcAec_weightCurve[i]) * hNl[i]; |
- } |
- hNl[i] = powf(hNl[i], aec->overDriveSm * WebRtcAec_overDriveCurve[i]); |
- |
- // Suppress error signal |
- efw[0][i] *= hNl[i]; |
- efw[1][i] *= hNl[i]; |
- |
- // Ooura fft returns incorrect sign on imaginary component. It matters here |
- // because we are making an additive change with comfort noise. |
- efw[1][i] *= -1; |
- } |
-} |
- |
-static int PartitionDelay(const AecCore* aec) { |
- // Measures the energy in each filter partition and returns the partition with |
- // highest energy. |
- // TODO(bjornv): Spread computational cost by computing one partition per |
- // block? |
- float wfEnMax = 0; |
- int i; |
- int delay = 0; |
- |
- for (i = 0; i < aec->num_partitions; i++) { |
- int j; |
- int pos = i * PART_LEN1; |
- float wfEn = 0; |
- for (j = 0; j < PART_LEN1; j++) { |
- wfEn += aec->wfBuf[0][pos + j] * aec->wfBuf[0][pos + j] + |
- aec->wfBuf[1][pos + j] * aec->wfBuf[1][pos + j]; |
- } |
- |
- if (wfEn > wfEnMax) { |
- wfEnMax = wfEn; |
- delay = i; |
- } |
- } |
- return delay; |
-} |
- |
-// Threshold to protect against the ill-effects of a zero far-end. |
-const float WebRtcAec_kMinFarendPSD = 15; |
- |
-// Updates the following smoothed Power Spectral Densities (PSD): |
-// - sd : near-end |
-// - se : residual echo |
-// - sx : far-end |
-// - sde : cross-PSD of near-end and residual echo |
-// - sxd : cross-PSD of near-end and far-end |
-// |
-// In addition to updating the PSDs, also the filter diverge state is |
-// determined. |
-static void SmoothedPSD(AecCore* aec, |
- float efw[2][PART_LEN1], |
- float dfw[2][PART_LEN1], |
- float xfw[2][PART_LEN1], |
- int* extreme_filter_divergence) { |
- // Power estimate smoothing coefficients. |
- const float* ptrGCoh = |
- aec->extended_filter_enabled |
- ? WebRtcAec_kExtendedSmoothingCoefficients[aec->mult - 1] |
- : WebRtcAec_kNormalSmoothingCoefficients[aec->mult - 1]; |
- int i; |
- float sdSum = 0, seSum = 0; |
- |
- for (i = 0; i < PART_LEN1; i++) { |
- aec->sd[i] = ptrGCoh[0] * aec->sd[i] + |
- ptrGCoh[1] * (dfw[0][i] * dfw[0][i] + dfw[1][i] * dfw[1][i]); |
- aec->se[i] = ptrGCoh[0] * aec->se[i] + |
- ptrGCoh[1] * (efw[0][i] * efw[0][i] + efw[1][i] * efw[1][i]); |
- // We threshold here to protect against the ill-effects of a zero farend. |
- // The threshold is not arbitrarily chosen, but balances protection and |
- // adverse interaction with the algorithm's tuning. |
- // TODO(bjornv): investigate further why this is so sensitive. |
- aec->sx[i] = ptrGCoh[0] * aec->sx[i] + |
- ptrGCoh[1] * WEBRTC_SPL_MAX( |
- xfw[0][i] * xfw[0][i] + xfw[1][i] * xfw[1][i], |
- WebRtcAec_kMinFarendPSD); |
- |
- aec->sde[i][0] = |
- ptrGCoh[0] * aec->sde[i][0] + |
- ptrGCoh[1] * (dfw[0][i] * efw[0][i] + dfw[1][i] * efw[1][i]); |
- aec->sde[i][1] = |
- ptrGCoh[0] * aec->sde[i][1] + |
- ptrGCoh[1] * (dfw[0][i] * efw[1][i] - dfw[1][i] * efw[0][i]); |
- |
- aec->sxd[i][0] = |
- ptrGCoh[0] * aec->sxd[i][0] + |
- ptrGCoh[1] * (dfw[0][i] * xfw[0][i] + dfw[1][i] * xfw[1][i]); |
- aec->sxd[i][1] = |
- ptrGCoh[0] * aec->sxd[i][1] + |
- ptrGCoh[1] * (dfw[0][i] * xfw[1][i] - dfw[1][i] * xfw[0][i]); |
- |
- sdSum += aec->sd[i]; |
- seSum += aec->se[i]; |
- } |
- |
- // Divergent filter safeguard update. |
- aec->divergeState = (aec->divergeState ? 1.05f : 1.0f) * seSum > sdSum; |
- |
- // Signal extreme filter divergence if the error is significantly larger |
- // than the nearend (13 dB). |
- *extreme_filter_divergence = (seSum > (19.95f * sdSum)); |
-} |
- |
-// Window time domain data to be used by the fft. |
-__inline static void WindowData(float* x_windowed, const float* x) { |
- int i; |
- for (i = 0; i < PART_LEN; i++) { |
- x_windowed[i] = x[i] * WebRtcAec_sqrtHanning[i]; |
- x_windowed[PART_LEN + i] = |
- x[PART_LEN + i] * WebRtcAec_sqrtHanning[PART_LEN - i]; |
- } |
-} |
- |
-// Puts fft output data into a complex valued array. |
-__inline static void StoreAsComplex(const float* data, |
- float data_complex[2][PART_LEN1]) { |
- int i; |
- data_complex[0][0] = data[0]; |
- data_complex[1][0] = 0; |
- for (i = 1; i < PART_LEN; i++) { |
- data_complex[0][i] = data[2 * i]; |
- data_complex[1][i] = data[2 * i + 1]; |
- } |
- data_complex[0][PART_LEN] = data[1]; |
- data_complex[1][PART_LEN] = 0; |
-} |
- |
-static void SubbandCoherence(AecCore* aec, |
- float efw[2][PART_LEN1], |
- float dfw[2][PART_LEN1], |
- float xfw[2][PART_LEN1], |
- float* fft, |
- float* cohde, |
- float* cohxd, |
- int* extreme_filter_divergence) { |
- int i; |
- |
- SmoothedPSD(aec, efw, dfw, xfw, extreme_filter_divergence); |
- |
- // Subband coherence |
- for (i = 0; i < PART_LEN1; i++) { |
- cohde[i] = |
- (aec->sde[i][0] * aec->sde[i][0] + aec->sde[i][1] * aec->sde[i][1]) / |
- (aec->sd[i] * aec->se[i] + 1e-10f); |
- cohxd[i] = |
- (aec->sxd[i][0] * aec->sxd[i][0] + aec->sxd[i][1] * aec->sxd[i][1]) / |
- (aec->sx[i] * aec->sd[i] + 1e-10f); |
- } |
-} |
- |
-static void GetHighbandGain(const float* lambda, float* nlpGainHband) { |
- int i; |
- |
- *nlpGainHband = (float)0.0; |
- for (i = freqAvgIc; i < PART_LEN1 - 1; i++) { |
- *nlpGainHband += lambda[i]; |
- } |
- *nlpGainHband /= (float)(PART_LEN1 - 1 - freqAvgIc); |
-} |
- |
-static void ComfortNoise(AecCore* aec, |
- float efw[2][PART_LEN1], |
- float comfortNoiseHband[2][PART_LEN1], |
- const float* noisePow, |
- const float* lambda) { |
- int i, num; |
- float rand[PART_LEN]; |
- float noise, noiseAvg, tmp, tmpAvg; |
- int16_t randW16[PART_LEN]; |
- float u[2][PART_LEN1]; |
- |
- const float pi2 = 6.28318530717959f; |
- |
- // Generate a uniform random array on [0 1] |
- WebRtcSpl_RandUArray(randW16, PART_LEN, &aec->seed); |
- for (i = 0; i < PART_LEN; i++) { |
- rand[i] = ((float)randW16[i]) / 32768; |
- } |
- |
- // Reject LF noise |
- u[0][0] = 0; |
- u[1][0] = 0; |
- for (i = 1; i < PART_LEN1; i++) { |
- tmp = pi2 * rand[i - 1]; |
- |
- noise = sqrtf(noisePow[i]); |
- u[0][i] = noise * cosf(tmp); |
- u[1][i] = -noise * sinf(tmp); |
- } |
- u[1][PART_LEN] = 0; |
- |
- for (i = 0; i < PART_LEN1; i++) { |
- // This is the proper weighting to match the background noise power |
- tmp = sqrtf(WEBRTC_SPL_MAX(1 - lambda[i] * lambda[i], 0)); |
- // tmp = 1 - lambda[i]; |
- efw[0][i] += tmp * u[0][i]; |
- efw[1][i] += tmp * u[1][i]; |
- } |
- |
- // For H band comfort noise |
- // TODO: don't compute noise and "tmp" twice. Use the previous results. |
- noiseAvg = 0.0; |
- tmpAvg = 0.0; |
- num = 0; |
- if (aec->num_bands > 1) { |
- // average noise scale |
- // average over second half of freq spectrum (i.e., 4->8khz) |
- // TODO: we shouldn't need num. We know how many elements we're summing. |
- for (i = PART_LEN1 >> 1; i < PART_LEN1; i++) { |
- num++; |
- noiseAvg += sqrtf(noisePow[i]); |
- } |
- noiseAvg /= (float)num; |
- |
- // average nlp scale |
- // average over second half of freq spectrum (i.e., 4->8khz) |
- // TODO: we shouldn't need num. We know how many elements we're summing. |
- num = 0; |
- for (i = PART_LEN1 >> 1; i < PART_LEN1; i++) { |
- num++; |
- tmpAvg += sqrtf(WEBRTC_SPL_MAX(1 - lambda[i] * lambda[i], 0)); |
- } |
- tmpAvg /= (float)num; |
- |
- // Use average noise for H band |
- // TODO: we should probably have a new random vector here. |
- // Reject LF noise |
- u[0][0] = 0; |
- u[1][0] = 0; |
- for (i = 1; i < PART_LEN1; i++) { |
- tmp = pi2 * rand[i - 1]; |
- |
- // Use average noise for H band |
- u[0][i] = noiseAvg * (float)cos(tmp); |
- u[1][i] = -noiseAvg * (float)sin(tmp); |
- } |
- u[1][PART_LEN] = 0; |
- |
- for (i = 0; i < PART_LEN1; i++) { |
- // Use average NLP weight for H band |
- comfortNoiseHband[0][i] = tmpAvg * u[0][i]; |
- comfortNoiseHband[1][i] = tmpAvg * u[1][i]; |
- } |
- } else { |
- memset(comfortNoiseHband, 0, |
- 2 * PART_LEN1 * sizeof(comfortNoiseHband[0][0])); |
- } |
-} |
- |
-static void InitLevel(PowerLevel* level) { |
- const float kBigFloat = 1E17f; |
- |
- level->averagelevel = 0; |
- level->framelevel = 0; |
- level->minlevel = kBigFloat; |
- level->frsum = 0; |
- level->sfrsum = 0; |
- level->frcounter = 0; |
- level->sfrcounter = 0; |
-} |
- |
-static void InitStats(Stats* stats) { |
- stats->instant = kOffsetLevel; |
- stats->average = kOffsetLevel; |
- stats->max = kOffsetLevel; |
- stats->min = kOffsetLevel * (-1); |
- stats->sum = 0; |
- stats->hisum = 0; |
- stats->himean = kOffsetLevel; |
- stats->counter = 0; |
- stats->hicounter = 0; |
-} |
- |
-static void InitMetrics(AecCore* self) { |
- self->stateCounter = 0; |
- InitLevel(&self->farlevel); |
- InitLevel(&self->nearlevel); |
- InitLevel(&self->linoutlevel); |
- InitLevel(&self->nlpoutlevel); |
- |
- InitStats(&self->erl); |
- InitStats(&self->erle); |
- InitStats(&self->aNlp); |
- InitStats(&self->rerl); |
-} |
- |
-static float CalculatePower(const float* in, size_t num_samples) { |
- size_t k; |
- float energy = 0.0f; |
- |
- for (k = 0; k < num_samples; ++k) { |
- energy += in[k] * in[k]; |
- } |
- return energy / num_samples; |
-} |
- |
-static void UpdateLevel(PowerLevel* level, float energy) { |
- level->sfrsum += energy; |
- level->sfrcounter++; |
- |
- if (level->sfrcounter > subCountLen) { |
- level->framelevel = level->sfrsum / (subCountLen * PART_LEN); |
- level->sfrsum = 0; |
- level->sfrcounter = 0; |
- if (level->framelevel > 0) { |
- if (level->framelevel < level->minlevel) { |
- level->minlevel = level->framelevel; // New minimum. |
- } else { |
- level->minlevel *= (1 + 0.001f); // Small increase. |
- } |
- } |
- level->frcounter++; |
- level->frsum += level->framelevel; |
- if (level->frcounter > countLen) { |
- level->averagelevel = level->frsum / countLen; |
- level->frsum = 0; |
- level->frcounter = 0; |
- } |
- } |
-} |
- |
-static void UpdateMetrics(AecCore* aec) { |
- float dtmp, dtmp2; |
- |
- const float actThresholdNoisy = 8.0f; |
- const float actThresholdClean = 40.0f; |
- const float safety = 0.99995f; |
- |
- // To make noisePower consistent with the legacy code, a factor of |
- // 2.0f / PART_LEN2 is applied to noisyPower, since the legacy code uses |
- // the energy of a frame as the audio levels, while the new code uses a |
- // a per-sample energy (i.e., power). |
- const float noisyPower = 300000.0f * 2.0f / PART_LEN2; |
- |
- float actThreshold; |
- float echo, suppressedEcho; |
- |
- if (aec->echoState) { // Check if echo is likely present |
- aec->stateCounter++; |
- } |
- |
- if (aec->farlevel.frcounter == 0) { |
- if (aec->farlevel.minlevel < noisyPower) { |
- actThreshold = actThresholdClean; |
- } else { |
- actThreshold = actThresholdNoisy; |
- } |
- |
- if ((aec->stateCounter > (0.5f * countLen * subCountLen)) && |
- (aec->farlevel.sfrcounter == 0) |
- |
- // Estimate in active far-end segments only |
- && (aec->farlevel.averagelevel > |
- (actThreshold * aec->farlevel.minlevel))) { |
- // Subtract noise power |
- echo = aec->nearlevel.averagelevel - safety * aec->nearlevel.minlevel; |
- |
- // ERL |
- dtmp = 10 * (float)log10(aec->farlevel.averagelevel / |
- aec->nearlevel.averagelevel + |
- 1e-10f); |
- dtmp2 = 10 * (float)log10(aec->farlevel.averagelevel / echo + 1e-10f); |
- |
- aec->erl.instant = dtmp; |
- if (dtmp > aec->erl.max) { |
- aec->erl.max = dtmp; |
- } |
- |
- if (dtmp < aec->erl.min) { |
- aec->erl.min = dtmp; |
- } |
- |
- aec->erl.counter++; |
- aec->erl.sum += dtmp; |
- aec->erl.average = aec->erl.sum / aec->erl.counter; |
- |
- // Upper mean |
- if (dtmp > aec->erl.average) { |
- aec->erl.hicounter++; |
- aec->erl.hisum += dtmp; |
- aec->erl.himean = aec->erl.hisum / aec->erl.hicounter; |
- } |
- |
- // A_NLP |
- dtmp = 10 * (float)log10(aec->nearlevel.averagelevel / |
- aec->linoutlevel.averagelevel + 1e-10f); |
- |
- // subtract noise power |
- suppressedEcho = aec->linoutlevel.averagelevel - |
- safety * aec->linoutlevel.minlevel; |
- |
- dtmp2 = 10 * (float)log10(echo / suppressedEcho + 1e-10f); |
- |
- aec->aNlp.instant = dtmp2; |
- if (dtmp > aec->aNlp.max) { |
- aec->aNlp.max = dtmp; |
- } |
- |
- if (dtmp < aec->aNlp.min) { |
- aec->aNlp.min = dtmp; |
- } |
- |
- aec->aNlp.counter++; |
- aec->aNlp.sum += dtmp; |
- aec->aNlp.average = aec->aNlp.sum / aec->aNlp.counter; |
- |
- // Upper mean |
- if (dtmp > aec->aNlp.average) { |
- aec->aNlp.hicounter++; |
- aec->aNlp.hisum += dtmp; |
- aec->aNlp.himean = aec->aNlp.hisum / aec->aNlp.hicounter; |
- } |
- |
- // ERLE |
- |
- // subtract noise power |
- suppressedEcho = 2 * (aec->nlpoutlevel.averagelevel - |
- safety * aec->nlpoutlevel.minlevel); |
- |
- dtmp = 10 * (float)log10(aec->nearlevel.averagelevel / |
- (2 * aec->nlpoutlevel.averagelevel) + |
- 1e-10f); |
- dtmp2 = 10 * (float)log10(echo / suppressedEcho + 1e-10f); |
- |
- dtmp = dtmp2; |
- aec->erle.instant = dtmp; |
- if (dtmp > aec->erle.max) { |
- aec->erle.max = dtmp; |
- } |
- |
- if (dtmp < aec->erle.min) { |
- aec->erle.min = dtmp; |
- } |
- |
- aec->erle.counter++; |
- aec->erle.sum += dtmp; |
- aec->erle.average = aec->erle.sum / aec->erle.counter; |
- |
- // Upper mean |
- if (dtmp > aec->erle.average) { |
- aec->erle.hicounter++; |
- aec->erle.hisum += dtmp; |
- aec->erle.himean = aec->erle.hisum / aec->erle.hicounter; |
- } |
- } |
- |
- aec->stateCounter = 0; |
- } |
-} |
- |
-static void UpdateDelayMetrics(AecCore* self) { |
- int i = 0; |
- int delay_values = 0; |
- int median = 0; |
- int lookahead = WebRtc_lookahead(self->delay_estimator); |
- const int kMsPerBlock = PART_LEN / (self->mult * 8); |
- int64_t l1_norm = 0; |
- |
- if (self->num_delay_values == 0) { |
- // We have no new delay value data. Even though -1 is a valid |median| in |
- // the sense that we allow negative values, it will practically never be |
- // used since multiples of |kMsPerBlock| will always be returned. |
- // We therefore use -1 to indicate in the logs that the delay estimator was |
- // not able to estimate the delay. |
- self->delay_median = -1; |
- self->delay_std = -1; |
- self->fraction_poor_delays = -1; |
- return; |
- } |
- |
- // Start value for median count down. |
- delay_values = self->num_delay_values >> 1; |
- // Get median of delay values since last update. |
- for (i = 0; i < kHistorySizeBlocks; i++) { |
- delay_values -= self->delay_histogram[i]; |
- if (delay_values < 0) { |
- median = i; |
- break; |
- } |
- } |
- // Account for lookahead. |
- self->delay_median = (median - lookahead) * kMsPerBlock; |
- |
- // Calculate the L1 norm, with median value as central moment. |
- for (i = 0; i < kHistorySizeBlocks; i++) { |
- l1_norm += abs(i - median) * self->delay_histogram[i]; |
- } |
- self->delay_std = |
- (int)((l1_norm + self->num_delay_values / 2) / self->num_delay_values) * |
- kMsPerBlock; |
- |
- // Determine fraction of delays that are out of bounds, that is, either |
- // negative (anti-causal system) or larger than the AEC filter length. |
- { |
- int num_delays_out_of_bounds = self->num_delay_values; |
- const int histogram_length = |
- sizeof(self->delay_histogram) / sizeof(self->delay_histogram[0]); |
- for (i = lookahead; i < lookahead + self->num_partitions; ++i) { |
- if (i < histogram_length) |
- num_delays_out_of_bounds -= self->delay_histogram[i]; |
- } |
- self->fraction_poor_delays = |
- (float)num_delays_out_of_bounds / self->num_delay_values; |
- } |
- |
- // Reset histogram. |
- memset(self->delay_histogram, 0, sizeof(self->delay_histogram)); |
- self->num_delay_values = 0; |
- |
- return; |
-} |
- |
-static void ScaledInverseFft(float freq_data[2][PART_LEN1], |
- float time_data[PART_LEN2], |
- float scale, |
- int conjugate) { |
- int i; |
- const float normalization = scale / ((float)PART_LEN2); |
- const float sign = (conjugate ? -1 : 1); |
- time_data[0] = freq_data[0][0] * normalization; |
- time_data[1] = freq_data[0][PART_LEN] * normalization; |
- for (i = 1; i < PART_LEN; i++) { |
- time_data[2 * i] = freq_data[0][i] * normalization; |
- time_data[2 * i + 1] = sign * freq_data[1][i] * normalization; |
- } |
- aec_rdft_inverse_128(time_data); |
-} |
- |
-static void Fft(float time_data[PART_LEN2], float freq_data[2][PART_LEN1]) { |
- int i; |
- aec_rdft_forward_128(time_data); |
- |
- // Reorder fft output data. |
- freq_data[1][0] = 0; |
- freq_data[1][PART_LEN] = 0; |
- freq_data[0][0] = time_data[0]; |
- freq_data[0][PART_LEN] = time_data[1]; |
- for (i = 1; i < PART_LEN; i++) { |
- freq_data[0][i] = time_data[2 * i]; |
- freq_data[1][i] = time_data[2 * i + 1]; |
- } |
-} |
- |
-static int SignalBasedDelayCorrection(AecCore* self) { |
- int delay_correction = 0; |
- int last_delay = -2; |
- assert(self != NULL); |
-#if !defined(WEBRTC_ANDROID) |
- // On desktops, turn on correction after |kDelayCorrectionStart| frames. This |
- // is to let the delay estimation get a chance to converge. Also, if the |
- // playout audio volume is low (or even muted) the delay estimation can return |
- // a very large delay, which will break the AEC if it is applied. |
- if (self->frame_count < kDelayCorrectionStart) { |
- return 0; |
- } |
-#endif |
- |
- // 1. Check for non-negative delay estimate. Note that the estimates we get |
- // from the delay estimation are not compensated for lookahead. Hence, a |
- // negative |last_delay| is an invalid one. |
- // 2. Verify that there is a delay change. In addition, only allow a change |
- // if the delay is outside a certain region taking the AEC filter length |
- // into account. |
- // TODO(bjornv): Investigate if we can remove the non-zero delay change check. |
- // 3. Only allow delay correction if the delay estimation quality exceeds |
- // |delay_quality_threshold|. |
- // 4. Finally, verify that the proposed |delay_correction| is feasible by |
- // comparing with the size of the far-end buffer. |
- last_delay = WebRtc_last_delay(self->delay_estimator); |
- if ((last_delay >= 0) && (last_delay != self->previous_delay) && |
- (WebRtc_last_delay_quality(self->delay_estimator) > |
- self->delay_quality_threshold)) { |
- int delay = last_delay - WebRtc_lookahead(self->delay_estimator); |
- // Allow for a slack in the actual delay, defined by a |lower_bound| and an |
- // |upper_bound|. The adaptive echo cancellation filter is currently |
- // |num_partitions| (of 64 samples) long. If the delay estimate is negative |
- // or at least 3/4 of the filter length we open up for correction. |
- const int lower_bound = 0; |
- const int upper_bound = self->num_partitions * 3 / 4; |
- const int do_correction = delay <= lower_bound || delay > upper_bound; |
- if (do_correction == 1) { |
- int available_read = (int)WebRtc_available_read(self->far_time_buf); |
- // With |shift_offset| we gradually rely on the delay estimates. For |
- // positive delays we reduce the correction by |shift_offset| to lower the |
- // risk of pushing the AEC into a non causal state. For negative delays |
- // we rely on the values up to a rounding error, hence compensate by 1 |
- // element to make sure to push the delay into the causal region. |
- delay_correction = -delay; |
- delay_correction += delay > self->shift_offset ? self->shift_offset : 1; |
- self->shift_offset--; |
- self->shift_offset = (self->shift_offset <= 1 ? 1 : self->shift_offset); |
- if (delay_correction > available_read - self->mult - 1) { |
- // There is not enough data in the buffer to perform this shift. Hence, |
- // we do not rely on the delay estimate and do nothing. |
- delay_correction = 0; |
- } else { |
- self->previous_delay = last_delay; |
- ++self->delay_correction_count; |
- } |
- } |
- } |
- // Update the |delay_quality_threshold| once we have our first delay |
- // correction. |
- if (self->delay_correction_count > 0) { |
- float delay_quality = WebRtc_last_delay_quality(self->delay_estimator); |
- delay_quality = |
- (delay_quality > kDelayQualityThresholdMax ? kDelayQualityThresholdMax |
- : delay_quality); |
- self->delay_quality_threshold = |
- (delay_quality > self->delay_quality_threshold |
- ? delay_quality |
- : self->delay_quality_threshold); |
- } |
- return delay_correction; |
-} |
- |
-static void EchoSubtraction(AecCore* aec, |
- int num_partitions, |
- int extended_filter_enabled, |
- float normal_mu, |
- float normal_error_threshold, |
- float* x_fft, |
- int* x_fft_buf_block_pos, |
- float x_fft_buf[2] |
- [kExtendedNumPartitions * PART_LEN1], |
- float* const y, |
- float x_pow[PART_LEN1], |
- float h_fft_buf[2] |
- [kExtendedNumPartitions * PART_LEN1], |
- float echo_subtractor_output[PART_LEN]) { |
- float s_fft[2][PART_LEN1]; |
- float e_extended[PART_LEN2]; |
- float s_extended[PART_LEN2]; |
- float* s; |
- float e[PART_LEN]; |
- float e_fft[2][PART_LEN1]; |
- int i; |
- |
- // Update the x_fft_buf block position. |
- (*x_fft_buf_block_pos)--; |
- if ((*x_fft_buf_block_pos) == -1) { |
- *x_fft_buf_block_pos = num_partitions - 1; |
- } |
- |
- // Buffer x_fft. |
- memcpy(x_fft_buf[0] + (*x_fft_buf_block_pos) * PART_LEN1, x_fft, |
- sizeof(float) * PART_LEN1); |
- memcpy(x_fft_buf[1] + (*x_fft_buf_block_pos) * PART_LEN1, &x_fft[PART_LEN1], |
- sizeof(float) * PART_LEN1); |
- |
- memset(s_fft, 0, sizeof(s_fft)); |
- |
- // Conditionally reset the echo subtraction filter if the filter has diverged |
- // significantly. |
- if (!aec->extended_filter_enabled && aec->extreme_filter_divergence) { |
- memset(aec->wfBuf, 0, sizeof(aec->wfBuf)); |
- aec->extreme_filter_divergence = 0; |
- } |
- |
- // Produce echo estimate s_fft. |
- WebRtcAec_FilterFar(num_partitions, *x_fft_buf_block_pos, x_fft_buf, |
- h_fft_buf, s_fft); |
- |
- // Compute the time-domain echo estimate s. |
- ScaledInverseFft(s_fft, s_extended, 2.0f, 0); |
- s = &s_extended[PART_LEN]; |
- |
- // Compute the time-domain echo prediction error. |
- for (i = 0; i < PART_LEN; ++i) { |
- e[i] = y[i] - s[i]; |
- } |
- |
- // Compute the frequency domain echo prediction error. |
- memset(e_extended, 0, sizeof(float) * PART_LEN); |
- memcpy(e_extended + PART_LEN, e, sizeof(float) * PART_LEN); |
- Fft(e_extended, e_fft); |
- |
- RTC_AEC_DEBUG_RAW_WRITE(aec->e_fft_file, &e_fft[0][0], |
- sizeof(e_fft[0][0]) * PART_LEN1 * 2); |
- |
- // Scale error signal inversely with far power. |
- WebRtcAec_ScaleErrorSignal(extended_filter_enabled, normal_mu, |
- normal_error_threshold, x_pow, e_fft); |
- WebRtcAec_FilterAdaptation(num_partitions, *x_fft_buf_block_pos, x_fft_buf, |
- e_fft, h_fft_buf); |
- memcpy(echo_subtractor_output, e, sizeof(float) * PART_LEN); |
-} |
- |
-static void EchoSuppression(AecCore* aec, |
- float farend[PART_LEN2], |
- float* echo_subtractor_output, |
- float* output, |
- float* const* outputH) { |
- float efw[2][PART_LEN1]; |
- float xfw[2][PART_LEN1]; |
- float dfw[2][PART_LEN1]; |
- float comfortNoiseHband[2][PART_LEN1]; |
- float fft[PART_LEN2]; |
- float nlpGainHband; |
- int i; |
- size_t j; |
- |
- // Coherence and non-linear filter |
- float cohde[PART_LEN1], cohxd[PART_LEN1]; |
- float hNlDeAvg, hNlXdAvg; |
- float hNl[PART_LEN1]; |
- float hNlPref[kPrefBandSize]; |
- float hNlFb = 0, hNlFbLow = 0; |
- const float prefBandQuant = 0.75f, prefBandQuantLow = 0.5f; |
- const int prefBandSize = kPrefBandSize / aec->mult; |
- const int minPrefBand = 4 / aec->mult; |
- // Power estimate smoothing coefficients. |
- const float* min_overdrive = aec->extended_filter_enabled |
- ? kExtendedMinOverDrive |
- : kNormalMinOverDrive; |
- |
- // Filter energy |
- const int delayEstInterval = 10 * aec->mult; |
- |
- float* xfw_ptr = NULL; |
- |
- // Update eBuf with echo subtractor output. |
- memcpy(aec->eBuf + PART_LEN, echo_subtractor_output, |
- sizeof(float) * PART_LEN); |
- |
- // Analysis filter banks for the echo suppressor. |
- // Windowed near-end ffts. |
- WindowData(fft, aec->dBuf); |
- aec_rdft_forward_128(fft); |
- StoreAsComplex(fft, dfw); |
- |
- // Windowed echo suppressor output ffts. |
- WindowData(fft, aec->eBuf); |
- aec_rdft_forward_128(fft); |
- StoreAsComplex(fft, efw); |
- |
- // NLP |
- |
- // Convert far-end partition to the frequency domain with windowing. |
- WindowData(fft, farend); |
- Fft(fft, xfw); |
- xfw_ptr = &xfw[0][0]; |
- |
- // Buffer far. |
- memcpy(aec->xfwBuf, xfw_ptr, sizeof(float) * 2 * PART_LEN1); |
- |
- aec->delayEstCtr++; |
- if (aec->delayEstCtr == delayEstInterval) { |
- aec->delayEstCtr = 0; |
- aec->delayIdx = WebRtcAec_PartitionDelay(aec); |
- } |
- |
- // Use delayed far. |
- memcpy(xfw, aec->xfwBuf + aec->delayIdx * PART_LEN1, |
- sizeof(xfw[0][0]) * 2 * PART_LEN1); |
- |
- WebRtcAec_SubbandCoherence(aec, efw, dfw, xfw, fft, cohde, cohxd, |
- &aec->extreme_filter_divergence); |
- |
- // Select the microphone signal as output if the filter is deemed to have |
- // diverged. |
- if (aec->divergeState) { |
- memcpy(efw, dfw, sizeof(efw[0][0]) * 2 * PART_LEN1); |
- } |
- |
- hNlXdAvg = 0; |
- for (i = minPrefBand; i < prefBandSize + minPrefBand; i++) { |
- hNlXdAvg += cohxd[i]; |
- } |
- hNlXdAvg /= prefBandSize; |
- hNlXdAvg = 1 - hNlXdAvg; |
- |
- hNlDeAvg = 0; |
- for (i = minPrefBand; i < prefBandSize + minPrefBand; i++) { |
- hNlDeAvg += cohde[i]; |
- } |
- hNlDeAvg /= prefBandSize; |
- |
- if (hNlXdAvg < 0.75f && hNlXdAvg < aec->hNlXdAvgMin) { |
- aec->hNlXdAvgMin = hNlXdAvg; |
- } |
- |
- if (hNlDeAvg > 0.98f && hNlXdAvg > 0.9f) { |
- aec->stNearState = 1; |
- } else if (hNlDeAvg < 0.95f || hNlXdAvg < 0.8f) { |
- aec->stNearState = 0; |
- } |
- |
- if (aec->hNlXdAvgMin == 1) { |
- aec->echoState = 0; |
- aec->overDrive = min_overdrive[aec->nlp_mode]; |
- |
- if (aec->stNearState == 1) { |
- memcpy(hNl, cohde, sizeof(hNl)); |
- hNlFb = hNlDeAvg; |
- hNlFbLow = hNlDeAvg; |
- } else { |
- for (i = 0; i < PART_LEN1; i++) { |
- hNl[i] = 1 - cohxd[i]; |
- } |
- hNlFb = hNlXdAvg; |
- hNlFbLow = hNlXdAvg; |
- } |
- } else { |
- if (aec->stNearState == 1) { |
- aec->echoState = 0; |
- memcpy(hNl, cohde, sizeof(hNl)); |
- hNlFb = hNlDeAvg; |
- hNlFbLow = hNlDeAvg; |
- } else { |
- aec->echoState = 1; |
- for (i = 0; i < PART_LEN1; i++) { |
- hNl[i] = WEBRTC_SPL_MIN(cohde[i], 1 - cohxd[i]); |
- } |
- |
- // Select an order statistic from the preferred bands. |
- // TODO: Using quicksort now, but a selection algorithm may be preferred. |
- memcpy(hNlPref, &hNl[minPrefBand], sizeof(float) * prefBandSize); |
- qsort(hNlPref, prefBandSize, sizeof(float), CmpFloat); |
- hNlFb = hNlPref[(int)floor(prefBandQuant * (prefBandSize - 1))]; |
- hNlFbLow = hNlPref[(int)floor(prefBandQuantLow * (prefBandSize - 1))]; |
- } |
- } |
- |
- // Track the local filter minimum to determine suppression overdrive. |
- if (hNlFbLow < 0.6f && hNlFbLow < aec->hNlFbLocalMin) { |
- aec->hNlFbLocalMin = hNlFbLow; |
- aec->hNlFbMin = hNlFbLow; |
- aec->hNlNewMin = 1; |
- aec->hNlMinCtr = 0; |
- } |
- aec->hNlFbLocalMin = |
- WEBRTC_SPL_MIN(aec->hNlFbLocalMin + 0.0008f / aec->mult, 1); |
- aec->hNlXdAvgMin = WEBRTC_SPL_MIN(aec->hNlXdAvgMin + 0.0006f / aec->mult, 1); |
- |
- if (aec->hNlNewMin == 1) { |
- aec->hNlMinCtr++; |
- } |
- if (aec->hNlMinCtr == 2) { |
- aec->hNlNewMin = 0; |
- aec->hNlMinCtr = 0; |
- aec->overDrive = |
- WEBRTC_SPL_MAX(kTargetSupp[aec->nlp_mode] / |
- ((float)log(aec->hNlFbMin + 1e-10f) + 1e-10f), |
- min_overdrive[aec->nlp_mode]); |
- } |
- |
- // Smooth the overdrive. |
- if (aec->overDrive < aec->overDriveSm) { |
- aec->overDriveSm = 0.99f * aec->overDriveSm + 0.01f * aec->overDrive; |
- } else { |
- aec->overDriveSm = 0.9f * aec->overDriveSm + 0.1f * aec->overDrive; |
- } |
- |
- WebRtcAec_OverdriveAndSuppress(aec, hNl, hNlFb, efw); |
- |
- // Add comfort noise. |
- WebRtcAec_ComfortNoise(aec, efw, comfortNoiseHband, aec->noisePow, hNl); |
- |
- // Inverse error fft. |
- ScaledInverseFft(efw, fft, 2.0f, 1); |
- |
- // TODO(bjornv): Investigate how to take the windowing below into account if |
- // needed. |
- if (aec->metricsMode == 1) { |
- // Note that we have a scaling by two in the time domain |eBuf|. |
- // In addition the time domain signal is windowed before transformation, |
- // losing half the energy on the average. We take care of the first |
- // scaling only in UpdateMetrics(). |
- UpdateLevel(&aec->nlpoutlevel, CalculatePower(fft, PART_LEN2)); |
- } |
- |
- // Overlap and add to obtain output. |
- for (i = 0; i < PART_LEN; i++) { |
- output[i] = (fft[i] * WebRtcAec_sqrtHanning[i] + |
- aec->outBuf[i] * WebRtcAec_sqrtHanning[PART_LEN - i]); |
- |
- // Saturate output to keep it in the allowed range. |
- output[i] = |
- WEBRTC_SPL_SAT(WEBRTC_SPL_WORD16_MAX, output[i], WEBRTC_SPL_WORD16_MIN); |
- } |
- memcpy(aec->outBuf, &fft[PART_LEN], PART_LEN * sizeof(aec->outBuf[0])); |
- |
- // For H band |
- if (aec->num_bands > 1) { |
- // H band gain |
- // average nlp over low band: average over second half of freq spectrum |
- // (4->8khz) |
- GetHighbandGain(hNl, &nlpGainHband); |
- |
- // Inverse comfort_noise |
- ScaledInverseFft(comfortNoiseHband, fft, 2.0f, 0); |
- |
- // compute gain factor |
- for (j = 0; j < aec->num_bands - 1; ++j) { |
- for (i = 0; i < PART_LEN; i++) { |
- outputH[j][i] = aec->dBufH[j][i] * nlpGainHband; |
- } |
- } |
- |
- // Add some comfort noise where Hband is attenuated. |
- for (i = 0; i < PART_LEN; i++) { |
- outputH[0][i] += cnScaleHband * fft[i]; |
- } |
- |
- // Saturate output to keep it in the allowed range. |
- for (j = 0; j < aec->num_bands - 1; ++j) { |
- for (i = 0; i < PART_LEN; i++) { |
- outputH[j][i] = WEBRTC_SPL_SAT(WEBRTC_SPL_WORD16_MAX, outputH[j][i], |
- WEBRTC_SPL_WORD16_MIN); |
- } |
- } |
- } |
- |
- // Copy the current block to the old position. |
- memcpy(aec->dBuf, aec->dBuf + PART_LEN, sizeof(float) * PART_LEN); |
- memcpy(aec->eBuf, aec->eBuf + PART_LEN, sizeof(float) * PART_LEN); |
- |
- // Copy the current block to the old position for H band |
- for (j = 0; j < aec->num_bands - 1; ++j) { |
- memcpy(aec->dBufH[j], aec->dBufH[j] + PART_LEN, sizeof(float) * PART_LEN); |
- } |
- |
- memmove(aec->xfwBuf + PART_LEN1, aec->xfwBuf, |
- sizeof(aec->xfwBuf) - sizeof(complex_t) * PART_LEN1); |
-} |
- |
-static void ProcessBlock(AecCore* aec) { |
- size_t i; |
- |
- float fft[PART_LEN2]; |
- float x_fft[2][PART_LEN1]; |
- float df[2][PART_LEN1]; |
- float far_spectrum = 0.0f; |
- float near_spectrum = 0.0f; |
- float abs_far_spectrum[PART_LEN1]; |
- float abs_near_spectrum[PART_LEN1]; |
- |
- const float gPow[2] = {0.9f, 0.1f}; |
- |
- // Noise estimate constants. |
- const int noiseInitBlocks = 500 * aec->mult; |
- const float step = 0.1f; |
- const float ramp = 1.0002f; |
- const float gInitNoise[2] = {0.999f, 0.001f}; |
- |
- float nearend[PART_LEN]; |
- float* nearend_ptr = NULL; |
- float farend[PART_LEN2]; |
- float* farend_ptr = NULL; |
- float echo_subtractor_output[PART_LEN]; |
- float output[PART_LEN]; |
- float outputH[NUM_HIGH_BANDS_MAX][PART_LEN]; |
- float* outputH_ptr[NUM_HIGH_BANDS_MAX]; |
- float* x_fft_ptr = NULL; |
- |
- for (i = 0; i < NUM_HIGH_BANDS_MAX; ++i) { |
- outputH_ptr[i] = outputH[i]; |
- } |
- |
- // Concatenate old and new nearend blocks. |
- for (i = 0; i < aec->num_bands - 1; ++i) { |
- WebRtc_ReadBuffer(aec->nearFrBufH[i], (void**)&nearend_ptr, nearend, |
- PART_LEN); |
- memcpy(aec->dBufH[i] + PART_LEN, nearend_ptr, sizeof(nearend)); |
- } |
- WebRtc_ReadBuffer(aec->nearFrBuf, (void**)&nearend_ptr, nearend, PART_LEN); |
- memcpy(aec->dBuf + PART_LEN, nearend_ptr, sizeof(nearend)); |
- |
- // We should always have at least one element stored in |far_buf|. |
- assert(WebRtc_available_read(aec->far_time_buf) > 0); |
- WebRtc_ReadBuffer(aec->far_time_buf, (void**)&farend_ptr, farend, 1); |
- |
-#ifdef WEBRTC_AEC_DEBUG_DUMP |
- { |
- // TODO(minyue): |farend_ptr| starts from buffered samples. This will be |
- // modified when |aec->far_time_buf| is revised. |
- RTC_AEC_DEBUG_WAV_WRITE(aec->farFile, &farend_ptr[PART_LEN], PART_LEN); |
- |
- RTC_AEC_DEBUG_WAV_WRITE(aec->nearFile, nearend_ptr, PART_LEN); |
- } |
-#endif |
- |
- if (aec->metricsMode == 1) { |
- // Update power levels |
- UpdateLevel(&aec->farlevel, |
- CalculatePower(&farend_ptr[PART_LEN], PART_LEN)); |
- UpdateLevel(&aec->nearlevel, CalculatePower(nearend_ptr, PART_LEN)); |
- } |
- |
- // Convert far-end signal to the frequency domain. |
- memcpy(fft, farend_ptr, sizeof(float) * PART_LEN2); |
- Fft(fft, x_fft); |
- x_fft_ptr = &x_fft[0][0]; |
- |
- // Near fft |
- memcpy(fft, aec->dBuf, sizeof(float) * PART_LEN2); |
- Fft(fft, df); |
- |
- // Power smoothing |
- for (i = 0; i < PART_LEN1; i++) { |
- far_spectrum = (x_fft_ptr[i] * x_fft_ptr[i]) + |
- (x_fft_ptr[PART_LEN1 + i] * x_fft_ptr[PART_LEN1 + i]); |
- aec->xPow[i] = |
- gPow[0] * aec->xPow[i] + gPow[1] * aec->num_partitions * far_spectrum; |
- // Calculate absolute spectra |
- abs_far_spectrum[i] = sqrtf(far_spectrum); |
- |
- near_spectrum = df[0][i] * df[0][i] + df[1][i] * df[1][i]; |
- aec->dPow[i] = gPow[0] * aec->dPow[i] + gPow[1] * near_spectrum; |
- // Calculate absolute spectra |
- abs_near_spectrum[i] = sqrtf(near_spectrum); |
- } |
- |
- // Estimate noise power. Wait until dPow is more stable. |
- if (aec->noiseEstCtr > 50) { |
- for (i = 0; i < PART_LEN1; i++) { |
- if (aec->dPow[i] < aec->dMinPow[i]) { |
- aec->dMinPow[i] = |
- (aec->dPow[i] + step * (aec->dMinPow[i] - aec->dPow[i])) * ramp; |
- } else { |
- aec->dMinPow[i] *= ramp; |
- } |
- } |
- } |
- |
- // Smooth increasing noise power from zero at the start, |
- // to avoid a sudden burst of comfort noise. |
- if (aec->noiseEstCtr < noiseInitBlocks) { |
- aec->noiseEstCtr++; |
- for (i = 0; i < PART_LEN1; i++) { |
- if (aec->dMinPow[i] > aec->dInitMinPow[i]) { |
- aec->dInitMinPow[i] = gInitNoise[0] * aec->dInitMinPow[i] + |
- gInitNoise[1] * aec->dMinPow[i]; |
- } else { |
- aec->dInitMinPow[i] = aec->dMinPow[i]; |
- } |
- } |
- aec->noisePow = aec->dInitMinPow; |
- } else { |
- aec->noisePow = aec->dMinPow; |
- } |
- |
- // Block wise delay estimation used for logging |
- if (aec->delay_logging_enabled) { |
- if (WebRtc_AddFarSpectrumFloat(aec->delay_estimator_farend, |
- abs_far_spectrum, PART_LEN1) == 0) { |
- int delay_estimate = WebRtc_DelayEstimatorProcessFloat( |
- aec->delay_estimator, abs_near_spectrum, PART_LEN1); |
- if (delay_estimate >= 0) { |
- // Update delay estimate buffer. |
- aec->delay_histogram[delay_estimate]++; |
- aec->num_delay_values++; |
- } |
- if (aec->delay_metrics_delivered == 1 && |
- aec->num_delay_values >= kDelayMetricsAggregationWindow) { |
- UpdateDelayMetrics(aec); |
- } |
- } |
- } |
- |
- // Perform echo subtraction. |
- EchoSubtraction(aec, aec->num_partitions, aec->extended_filter_enabled, |
- aec->normal_mu, aec->normal_error_threshold, &x_fft[0][0], |
- &aec->xfBufBlockPos, aec->xfBuf, nearend_ptr, aec->xPow, |
- aec->wfBuf, echo_subtractor_output); |
- |
- RTC_AEC_DEBUG_WAV_WRITE(aec->outLinearFile, echo_subtractor_output, PART_LEN); |
- |
- if (aec->metricsMode == 1) { |
- UpdateLevel(&aec->linoutlevel, |
- CalculatePower(echo_subtractor_output, PART_LEN)); |
- } |
- |
- // Perform echo suppression. |
- EchoSuppression(aec, farend_ptr, echo_subtractor_output, output, outputH_ptr); |
- |
- if (aec->metricsMode == 1) { |
- UpdateMetrics(aec); |
- } |
- |
- // Store the output block. |
- WebRtc_WriteBuffer(aec->outFrBuf, output, PART_LEN); |
- // For high bands |
- for (i = 0; i < aec->num_bands - 1; ++i) { |
- WebRtc_WriteBuffer(aec->outFrBufH[i], outputH[i], PART_LEN); |
- } |
- |
- RTC_AEC_DEBUG_WAV_WRITE(aec->outFile, output, PART_LEN); |
-} |
- |
-AecCore* WebRtcAec_CreateAec() { |
- int i; |
- AecCore* aec = malloc(sizeof(AecCore)); |
- if (!aec) { |
- return NULL; |
- } |
- |
- aec->nearFrBuf = WebRtc_CreateBuffer(FRAME_LEN + PART_LEN, sizeof(float)); |
- if (!aec->nearFrBuf) { |
- WebRtcAec_FreeAec(aec); |
- return NULL; |
- } |
- |
- aec->outFrBuf = WebRtc_CreateBuffer(FRAME_LEN + PART_LEN, sizeof(float)); |
- if (!aec->outFrBuf) { |
- WebRtcAec_FreeAec(aec); |
- return NULL; |
- } |
- |
- for (i = 0; i < NUM_HIGH_BANDS_MAX; ++i) { |
- aec->nearFrBufH[i] = |
- WebRtc_CreateBuffer(FRAME_LEN + PART_LEN, sizeof(float)); |
- if (!aec->nearFrBufH[i]) { |
- WebRtcAec_FreeAec(aec); |
- return NULL; |
- } |
- aec->outFrBufH[i] = |
- WebRtc_CreateBuffer(FRAME_LEN + PART_LEN, sizeof(float)); |
- if (!aec->outFrBufH[i]) { |
- WebRtcAec_FreeAec(aec); |
- return NULL; |
- } |
- } |
- |
- // Create far-end buffers. |
- // For bit exactness with legacy code, each element in |far_time_buf| is |
- // supposed to contain |PART_LEN2| samples with an overlap of |PART_LEN| |
- // samples from the last frame. |
- // TODO(minyue): reduce |far_time_buf| to non-overlapped |PART_LEN| samples. |
- aec->far_time_buf = |
- WebRtc_CreateBuffer(kBufSizePartitions, sizeof(float) * PART_LEN2); |
- if (!aec->far_time_buf) { |
- WebRtcAec_FreeAec(aec); |
- return NULL; |
- } |
- |
-#ifdef WEBRTC_AEC_DEBUG_DUMP |
- aec->instance_index = webrtc_aec_instance_count; |
- |
- aec->farFile = aec->nearFile = aec->outFile = aec->outLinearFile = NULL; |
- aec->debug_dump_count = 0; |
-#endif |
- aec->delay_estimator_farend = |
- WebRtc_CreateDelayEstimatorFarend(PART_LEN1, kHistorySizeBlocks); |
- if (aec->delay_estimator_farend == NULL) { |
- WebRtcAec_FreeAec(aec); |
- return NULL; |
- } |
- // We create the delay_estimator with the same amount of maximum lookahead as |
- // the delay history size (kHistorySizeBlocks) for symmetry reasons. |
- aec->delay_estimator = WebRtc_CreateDelayEstimator( |
- aec->delay_estimator_farend, kHistorySizeBlocks); |
- if (aec->delay_estimator == NULL) { |
- WebRtcAec_FreeAec(aec); |
- return NULL; |
- } |
-#ifdef WEBRTC_ANDROID |
- aec->delay_agnostic_enabled = 1; // DA-AEC enabled by default. |
- // DA-AEC assumes the system is causal from the beginning and will self adjust |
- // the lookahead when shifting is required. |
- WebRtc_set_lookahead(aec->delay_estimator, 0); |
-#else |
- aec->delay_agnostic_enabled = 0; |
- WebRtc_set_lookahead(aec->delay_estimator, kLookaheadBlocks); |
-#endif |
- aec->extended_filter_enabled = 0; |
- aec->next_generation_aec_enabled = 0; |
- |
- // Assembly optimization |
- WebRtcAec_FilterFar = FilterFar; |
- WebRtcAec_ScaleErrorSignal = ScaleErrorSignal; |
- WebRtcAec_FilterAdaptation = FilterAdaptation; |
- WebRtcAec_OverdriveAndSuppress = OverdriveAndSuppress; |
- WebRtcAec_ComfortNoise = ComfortNoise; |
- WebRtcAec_SubbandCoherence = SubbandCoherence; |
- WebRtcAec_StoreAsComplex = StoreAsComplex; |
- WebRtcAec_PartitionDelay = PartitionDelay; |
- WebRtcAec_WindowData = WindowData; |
- |
-#if defined(WEBRTC_ARCH_X86_FAMILY) |
- if (WebRtc_GetCPUInfo(kSSE2)) { |
- WebRtcAec_InitAec_SSE2(); |
- } |
-#endif |
- |
-#if defined(MIPS_FPU_LE) |
- WebRtcAec_InitAec_mips(); |
-#endif |
- |
-#if defined(WEBRTC_HAS_NEON) |
- WebRtcAec_InitAec_neon(); |
-#elif defined(WEBRTC_DETECT_NEON) |
- if ((WebRtc_GetCPUFeaturesARM() & kCPUFeatureNEON) != 0) { |
- WebRtcAec_InitAec_neon(); |
- } |
-#endif |
- |
- aec_rdft_init(); |
- |
- return aec; |
-} |
- |
-void WebRtcAec_FreeAec(AecCore* aec) { |
- int i; |
- if (aec == NULL) { |
- return; |
- } |
- |
- WebRtc_FreeBuffer(aec->nearFrBuf); |
- WebRtc_FreeBuffer(aec->outFrBuf); |
- |
- for (i = 0; i < NUM_HIGH_BANDS_MAX; ++i) { |
- WebRtc_FreeBuffer(aec->nearFrBufH[i]); |
- WebRtc_FreeBuffer(aec->outFrBufH[i]); |
- } |
- |
- WebRtc_FreeBuffer(aec->far_time_buf); |
- |
- RTC_AEC_DEBUG_WAV_CLOSE(aec->farFile); |
- RTC_AEC_DEBUG_WAV_CLOSE(aec->nearFile); |
- RTC_AEC_DEBUG_WAV_CLOSE(aec->outFile); |
- RTC_AEC_DEBUG_WAV_CLOSE(aec->outLinearFile); |
- RTC_AEC_DEBUG_RAW_CLOSE(aec->e_fft_file); |
- |
- WebRtc_FreeDelayEstimator(aec->delay_estimator); |
- WebRtc_FreeDelayEstimatorFarend(aec->delay_estimator_farend); |
- |
- free(aec); |
-} |
- |
-int WebRtcAec_InitAec(AecCore* aec, int sampFreq) { |
- int i; |
- |
- aec->sampFreq = sampFreq; |
- |
- if (sampFreq == 8000) { |
- aec->normal_mu = 0.6f; |
- aec->normal_error_threshold = 2e-6f; |
- aec->num_bands = 1; |
- } else { |
- aec->normal_mu = 0.5f; |
- aec->normal_error_threshold = 1.5e-6f; |
- aec->num_bands = (size_t)(sampFreq / 16000); |
- } |
- |
- WebRtc_InitBuffer(aec->nearFrBuf); |
- WebRtc_InitBuffer(aec->outFrBuf); |
- for (i = 0; i < NUM_HIGH_BANDS_MAX; ++i) { |
- WebRtc_InitBuffer(aec->nearFrBufH[i]); |
- WebRtc_InitBuffer(aec->outFrBufH[i]); |
- } |
- |
- // Initialize far-end buffers. |
- WebRtc_InitBuffer(aec->far_time_buf); |
- |
-#ifdef WEBRTC_AEC_DEBUG_DUMP |
- { |
- int process_rate = sampFreq > 16000 ? 16000 : sampFreq; |
- RTC_AEC_DEBUG_WAV_REOPEN("aec_far", aec->instance_index, |
- aec->debug_dump_count, process_rate, |
- &aec->farFile); |
- RTC_AEC_DEBUG_WAV_REOPEN("aec_near", aec->instance_index, |
- aec->debug_dump_count, process_rate, |
- &aec->nearFile); |
- RTC_AEC_DEBUG_WAV_REOPEN("aec_out", aec->instance_index, |
- aec->debug_dump_count, process_rate, |
- &aec->outFile); |
- RTC_AEC_DEBUG_WAV_REOPEN("aec_out_linear", aec->instance_index, |
- aec->debug_dump_count, process_rate, |
- &aec->outLinearFile); |
- } |
- |
- RTC_AEC_DEBUG_RAW_OPEN("aec_e_fft", aec->debug_dump_count, &aec->e_fft_file); |
- |
- ++aec->debug_dump_count; |
-#endif |
- aec->system_delay = 0; |
- |
- if (WebRtc_InitDelayEstimatorFarend(aec->delay_estimator_farend) != 0) { |
- return -1; |
- } |
- if (WebRtc_InitDelayEstimator(aec->delay_estimator) != 0) { |
- return -1; |
- } |
- aec->delay_logging_enabled = 0; |
- aec->delay_metrics_delivered = 0; |
- memset(aec->delay_histogram, 0, sizeof(aec->delay_histogram)); |
- aec->num_delay_values = 0; |
- aec->delay_median = -1; |
- aec->delay_std = -1; |
- aec->fraction_poor_delays = -1.0f; |
- |
- aec->signal_delay_correction = 0; |
- aec->previous_delay = -2; // (-2): Uninitialized. |
- aec->delay_correction_count = 0; |
- aec->shift_offset = kInitialShiftOffset; |
- aec->delay_quality_threshold = kDelayQualityThresholdMin; |
- |
- aec->num_partitions = kNormalNumPartitions; |
- |
- // Update the delay estimator with filter length. We use half the |
- // |num_partitions| to take the echo path into account. In practice we say |
- // that the echo has a duration of maximum half |num_partitions|, which is not |
- // true, but serves as a crude measure. |
- WebRtc_set_allowed_offset(aec->delay_estimator, aec->num_partitions / 2); |
- // TODO(bjornv): I currently hard coded the enable. Once we've established |
- // that AECM has no performance regression, robust_validation will be enabled |
- // all the time and the APIs to turn it on/off will be removed. Hence, remove |
- // this line then. |
- WebRtc_enable_robust_validation(aec->delay_estimator, 1); |
- aec->frame_count = 0; |
- |
- // Default target suppression mode. |
- aec->nlp_mode = 1; |
- |
- // Sampling frequency multiplier w.r.t. 8 kHz. |
- // In case of multiple bands we process the lower band in 16 kHz, hence the |
- // multiplier is always 2. |
- if (aec->num_bands > 1) { |
- aec->mult = 2; |
- } else { |
- aec->mult = (short)aec->sampFreq / 8000; |
- } |
- |
- aec->farBufWritePos = 0; |
- aec->farBufReadPos = 0; |
- |
- aec->inSamples = 0; |
- aec->outSamples = 0; |
- aec->knownDelay = 0; |
- |
- // Initialize buffers |
- memset(aec->dBuf, 0, sizeof(aec->dBuf)); |
- memset(aec->eBuf, 0, sizeof(aec->eBuf)); |
- // For H bands |
- for (i = 0; i < NUM_HIGH_BANDS_MAX; ++i) { |
- memset(aec->dBufH[i], 0, sizeof(aec->dBufH[i])); |
- } |
- |
- memset(aec->xPow, 0, sizeof(aec->xPow)); |
- memset(aec->dPow, 0, sizeof(aec->dPow)); |
- memset(aec->dInitMinPow, 0, sizeof(aec->dInitMinPow)); |
- aec->noisePow = aec->dInitMinPow; |
- aec->noiseEstCtr = 0; |
- |
- // Initial comfort noise power |
- for (i = 0; i < PART_LEN1; i++) { |
- aec->dMinPow[i] = 1.0e6f; |
- } |
- |
- // Holds the last block written to |
- aec->xfBufBlockPos = 0; |
- // TODO: Investigate need for these initializations. Deleting them doesn't |
- // change the output at all and yields 0.4% overall speedup. |
- memset(aec->xfBuf, 0, sizeof(complex_t) * kExtendedNumPartitions * PART_LEN1); |
- memset(aec->wfBuf, 0, sizeof(complex_t) * kExtendedNumPartitions * PART_LEN1); |
- memset(aec->sde, 0, sizeof(complex_t) * PART_LEN1); |
- memset(aec->sxd, 0, sizeof(complex_t) * PART_LEN1); |
- memset(aec->xfwBuf, 0, |
- sizeof(complex_t) * kExtendedNumPartitions * PART_LEN1); |
- memset(aec->se, 0, sizeof(float) * PART_LEN1); |
- |
- // To prevent numerical instability in the first block. |
- for (i = 0; i < PART_LEN1; i++) { |
- aec->sd[i] = 1; |
- } |
- for (i = 0; i < PART_LEN1; i++) { |
- aec->sx[i] = 1; |
- } |
- |
- memset(aec->hNs, 0, sizeof(aec->hNs)); |
- memset(aec->outBuf, 0, sizeof(float) * PART_LEN); |
- |
- aec->hNlFbMin = 1; |
- aec->hNlFbLocalMin = 1; |
- aec->hNlXdAvgMin = 1; |
- aec->hNlNewMin = 0; |
- aec->hNlMinCtr = 0; |
- aec->overDrive = 2; |
- aec->overDriveSm = 2; |
- aec->delayIdx = 0; |
- aec->stNearState = 0; |
- aec->echoState = 0; |
- aec->divergeState = 0; |
- |
- aec->seed = 777; |
- aec->delayEstCtr = 0; |
- |
- aec->extreme_filter_divergence = 0; |
- |
- // Metrics disabled by default |
- aec->metricsMode = 0; |
- InitMetrics(aec); |
- |
- return 0; |
-} |
- |
-// For bit exactness with a legacy code, |farend| is supposed to contain |
-// |PART_LEN2| samples with an overlap of |PART_LEN| samples from the last |
-// frame. |
-// TODO(minyue): reduce |farend| to non-overlapped |PART_LEN| samples. |
-void WebRtcAec_BufferFarendPartition(AecCore* aec, const float* farend) { |
- // Check if the buffer is full, and in that case flush the oldest data. |
- if (WebRtc_available_write(aec->far_time_buf) < 1) { |
- WebRtcAec_MoveFarReadPtr(aec, 1); |
- } |
- |
- WebRtc_WriteBuffer(aec->far_time_buf, farend, 1); |
-} |
- |
-int WebRtcAec_MoveFarReadPtr(AecCore* aec, int elements) { |
- int elements_moved = WebRtc_MoveReadPtr(aec->far_time_buf, elements); |
- aec->system_delay -= elements_moved * PART_LEN; |
- return elements_moved; |
-} |
- |
-void WebRtcAec_ProcessFrames(AecCore* aec, |
- const float* const* nearend, |
- size_t num_bands, |
- size_t num_samples, |
- int knownDelay, |
- float* const* out) { |
- size_t i, j; |
- int out_elements = 0; |
- |
- aec->frame_count++; |
- // For each frame the process is as follows: |
- // 1) If the system_delay indicates on being too small for processing a |
- // frame we stuff the buffer with enough data for 10 ms. |
- // 2 a) Adjust the buffer to the system delay, by moving the read pointer. |
- // b) Apply signal based delay correction, if we have detected poor AEC |
- // performance. |
- // 3) TODO(bjornv): Investigate if we need to add this: |
- // If we can't move read pointer due to buffer size limitations we |
- // flush/stuff the buffer. |
- // 4) Process as many partitions as possible. |
- // 5) Update the |system_delay| with respect to a full frame of FRAME_LEN |
- // samples. Even though we will have data left to process (we work with |
- // partitions) we consider updating a whole frame, since that's the |
- // amount of data we input and output in audio_processing. |
- // 6) Update the outputs. |
- |
- // The AEC has two different delay estimation algorithms built in. The |
- // first relies on delay input values from the user and the amount of |
- // shifted buffer elements is controlled by |knownDelay|. This delay will |
- // give a guess on how much we need to shift far-end buffers to align with |
- // the near-end signal. The other delay estimation algorithm uses the |
- // far- and near-end signals to find the offset between them. This one |
- // (called "signal delay") is then used to fine tune the alignment, or |
- // simply compensate for errors in the system based one. |
- // Note that the two algorithms operate independently. Currently, we only |
- // allow one algorithm to be turned on. |
- |
- assert(aec->num_bands == num_bands); |
- |
- for (j = 0; j < num_samples; j += FRAME_LEN) { |
- // TODO(bjornv): Change the near-end buffer handling to be the same as for |
- // far-end, that is, with a near_pre_buf. |
- // Buffer the near-end frame. |
- WebRtc_WriteBuffer(aec->nearFrBuf, &nearend[0][j], FRAME_LEN); |
- // For H band |
- for (i = 1; i < num_bands; ++i) { |
- WebRtc_WriteBuffer(aec->nearFrBufH[i - 1], &nearend[i][j], FRAME_LEN); |
- } |
- |
- // 1) At most we process |aec->mult|+1 partitions in 10 ms. Make sure we |
- // have enough far-end data for that by stuffing the buffer if the |
- // |system_delay| indicates others. |
- if (aec->system_delay < FRAME_LEN) { |
- // We don't have enough data so we rewind 10 ms. |
- WebRtcAec_MoveFarReadPtr(aec, -(aec->mult + 1)); |
- } |
- |
- if (!aec->delay_agnostic_enabled) { |
- // 2 a) Compensate for a possible change in the system delay. |
- |
- // TODO(bjornv): Investigate how we should round the delay difference; |
- // right now we know that incoming |knownDelay| is underestimated when |
- // it's less than |aec->knownDelay|. We therefore, round (-32) in that |
- // direction. In the other direction, we don't have this situation, but |
- // might flush one partition too little. This can cause non-causality, |
- // which should be investigated. Maybe, allow for a non-symmetric |
- // rounding, like -16. |
- int move_elements = (aec->knownDelay - knownDelay - 32) / PART_LEN; |
- int moved_elements = WebRtc_MoveReadPtr(aec->far_time_buf, move_elements); |
- aec->knownDelay -= moved_elements * PART_LEN; |
- } else { |
- // 2 b) Apply signal based delay correction. |
- int move_elements = SignalBasedDelayCorrection(aec); |
- int moved_elements = WebRtc_MoveReadPtr(aec->far_time_buf, move_elements); |
- int far_near_buffer_diff = |
- WebRtc_available_read(aec->far_time_buf) - |
- WebRtc_available_read(aec->nearFrBuf) / PART_LEN; |
- WebRtc_SoftResetDelayEstimator(aec->delay_estimator, moved_elements); |
- WebRtc_SoftResetDelayEstimatorFarend(aec->delay_estimator_farend, |
- moved_elements); |
- aec->signal_delay_correction += moved_elements; |
- // If we rely on reported system delay values only, a buffer underrun here |
- // can never occur since we've taken care of that in 1) above. Here, we |
- // apply signal based delay correction and can therefore end up with |
- // buffer underruns since the delay estimation can be wrong. We therefore |
- // stuff the buffer with enough elements if needed. |
- if (far_near_buffer_diff < 0) { |
- WebRtcAec_MoveFarReadPtr(aec, far_near_buffer_diff); |
- } |
- } |
- |
- // 4) Process as many blocks as possible. |
- while (WebRtc_available_read(aec->nearFrBuf) >= PART_LEN) { |
- ProcessBlock(aec); |
- } |
- |
- // 5) Update system delay with respect to the entire frame. |
- aec->system_delay -= FRAME_LEN; |
- |
- // 6) Update output frame. |
- // Stuff the out buffer if we have less than a frame to output. |
- // This should only happen for the first frame. |
- out_elements = (int)WebRtc_available_read(aec->outFrBuf); |
- if (out_elements < FRAME_LEN) { |
- WebRtc_MoveReadPtr(aec->outFrBuf, out_elements - FRAME_LEN); |
- for (i = 0; i < num_bands - 1; ++i) { |
- WebRtc_MoveReadPtr(aec->outFrBufH[i], out_elements - FRAME_LEN); |
- } |
- } |
- // Obtain an output frame. |
- WebRtc_ReadBuffer(aec->outFrBuf, NULL, &out[0][j], FRAME_LEN); |
- // For H bands. |
- for (i = 1; i < num_bands; ++i) { |
- WebRtc_ReadBuffer(aec->outFrBufH[i - 1], NULL, &out[i][j], FRAME_LEN); |
- } |
- } |
-} |
- |
-int WebRtcAec_GetDelayMetricsCore(AecCore* self, |
- int* median, |
- int* std, |
- float* fraction_poor_delays) { |
- assert(self != NULL); |
- assert(median != NULL); |
- assert(std != NULL); |
- |
- if (self->delay_logging_enabled == 0) { |
- // Logging disabled. |
- return -1; |
- } |
- |
- if (self->delay_metrics_delivered == 0) { |
- UpdateDelayMetrics(self); |
- self->delay_metrics_delivered = 1; |
- } |
- *median = self->delay_median; |
- *std = self->delay_std; |
- *fraction_poor_delays = self->fraction_poor_delays; |
- |
- return 0; |
-} |
- |
-int WebRtcAec_echo_state(AecCore* self) { |
- return self->echoState; |
-} |
- |
-void WebRtcAec_GetEchoStats(AecCore* self, |
- Stats* erl, |
- Stats* erle, |
- Stats* a_nlp) { |
- assert(erl != NULL); |
- assert(erle != NULL); |
- assert(a_nlp != NULL); |
- *erl = self->erl; |
- *erle = self->erle; |
- *a_nlp = self->aNlp; |
-} |
- |
-void WebRtcAec_SetConfigCore(AecCore* self, |
- int nlp_mode, |
- int metrics_mode, |
- int delay_logging) { |
- assert(nlp_mode >= 0 && nlp_mode < 3); |
- self->nlp_mode = nlp_mode; |
- self->metricsMode = metrics_mode; |
- if (self->metricsMode) { |
- InitMetrics(self); |
- } |
- // Turn on delay logging if it is either set explicitly or if delay agnostic |
- // AEC is enabled (which requires delay estimates). |
- self->delay_logging_enabled = delay_logging || self->delay_agnostic_enabled; |
- if (self->delay_logging_enabled) { |
- memset(self->delay_histogram, 0, sizeof(self->delay_histogram)); |
- } |
-} |
- |
-void WebRtcAec_enable_delay_agnostic(AecCore* self, int enable) { |
- self->delay_agnostic_enabled = enable; |
-} |
- |
-int WebRtcAec_delay_agnostic_enabled(AecCore* self) { |
- return self->delay_agnostic_enabled; |
-} |
- |
-void WebRtcAec_enable_next_generation_aec(AecCore* self, int enable) { |
- self->next_generation_aec_enabled = (enable != 0); |
-} |
- |
-int WebRtcAec_next_generation_aec_enabled(AecCore* self) { |
- assert(self->next_generation_aec_enabled == 0 || |
- self->next_generation_aec_enabled == 1); |
- return self->next_generation_aec_enabled; |
-} |
- |
- |
-void WebRtcAec_enable_extended_filter(AecCore* self, int enable) { |
- self->extended_filter_enabled = enable; |
- self->num_partitions = enable ? kExtendedNumPartitions : kNormalNumPartitions; |
- // Update the delay estimator with filter length. See InitAEC() for details. |
- WebRtc_set_allowed_offset(self->delay_estimator, self->num_partitions / 2); |
-} |
- |
-int WebRtcAec_extended_filter_enabled(AecCore* self) { |
- return self->extended_filter_enabled; |
-} |
- |
-int WebRtcAec_system_delay(AecCore* self) { |
- return self->system_delay; |
-} |
- |
-void WebRtcAec_SetSystemDelay(AecCore* self, int delay) { |
- assert(delay >= 0); |
- self->system_delay = delay; |
-} |