OLD | NEW |
---|---|
1 /* | 1 /* |
2 * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved. | 2 * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved. |
3 * | 3 * |
4 * Use of this source code is governed by a BSD-style license | 4 * Use of this source code is governed by a BSD-style license |
5 * that can be found in the LICENSE file in the root of the source | 5 * that can be found in the LICENSE file in the root of the source |
6 * tree. An additional intellectual property rights grant can be found | 6 * tree. An additional intellectual property rights grant can be found |
7 * in the file PATENTS. All contributing project authors may | 7 * in the file PATENTS. All contributing project authors may |
8 * be found in the AUTHORS file in the root of the source tree. | 8 * be found in the AUTHORS file in the root of the source tree. |
9 */ | 9 */ |
10 | 10 |
(...skipping 28 matching lines...) Expand all Loading... | |
39 SyncBuffer* sync_buffer); | 39 SyncBuffer* sync_buffer); |
40 virtual ~Merge() {} | 40 virtual ~Merge() {} |
41 | 41 |
42 // The main method to produce the audio data. The decoded data is supplied in | 42 // The main method to produce the audio data. The decoded data is supplied in |
43 // |input|, having |input_length| samples in total for all channels | 43 // |input|, having |input_length| samples in total for all channels |
44 // (interleaved). The result is written to |output|. The number of channels | 44 // (interleaved). The result is written to |output|. The number of channels |
45 // allocated in |output| defines the number of channels that will be used when | 45 // allocated in |output| defines the number of channels that will be used when |
46 // de-interleaving |input|. The values in |external_mute_factor_array| (Q14) | 46 // de-interleaving |input|. The values in |external_mute_factor_array| (Q14) |
47 // will be used to scale the audio, and is updated in the process. The array | 47 // will be used to scale the audio, and is updated in the process. The array |
48 // must have |num_channels_| elements. | 48 // must have |num_channels_| elements. |
49 virtual int Process(int16_t* input, size_t input_length, | 49 virtual size_t Process(int16_t* input, size_t input_length, |
50 int16_t* external_mute_factor_array, | 50 int16_t* external_mute_factor_array, |
51 AudioMultiVector* output); | 51 AudioMultiVector* output); |
52 | 52 |
53 virtual int RequiredFutureSamples(); | 53 virtual size_t RequiredFutureSamples(); |
54 | 54 |
55 protected: | 55 protected: |
56 const int fs_hz_; | 56 const int fs_hz_; |
57 const size_t num_channels_; | 57 const size_t num_channels_; |
58 | 58 |
59 private: | 59 private: |
60 static const int kMaxSampleRate = 48000; | 60 static const int kMaxSampleRate = 48000; |
61 static const int kExpandDownsampLength = 100; | 61 static const size_t kExpandDownsampLength = 100; |
62 static const int kInputDownsampLength = 40; | 62 static const size_t kInputDownsampLength = 40; |
63 static const int kMaxCorrelationLength = 60; | 63 static const size_t kMaxCorrelationLength = 60; |
64 | 64 |
65 // Calls |expand_| to get more expansion data to merge with. The data is | 65 // Calls |expand_| to get more expansion data to merge with. The data is |
66 // written to |expanded_signal_|. Returns the length of the expanded data, | 66 // written to |expanded_signal_|. Returns the length of the expanded data, |
67 // while |expand_period| will be the number of samples in one expansion period | 67 // while |expand_period| will be the number of samples in one expansion period |
68 // (typically one pitch period). The value of |old_length| will be the number | 68 // (typically one pitch period). The value of |old_length| will be the number |
69 // of samples that were taken from the |sync_buffer_|. | 69 // of samples that were taken from the |sync_buffer_|. |
70 int GetExpandedSignal(int* old_length, int* expand_period); | 70 size_t GetExpandedSignal(size_t* old_length, size_t* expand_period); |
71 | 71 |
72 // Analyzes |input| and |expanded_signal| to find maximum values. Returns | 72 // Analyzes |input| and |expanded_signal| to find maximum values. Returns |
73 // a muting factor (Q14) to be used on the new data. | 73 // a muting factor (Q14) to be used on the new data. |
74 int16_t SignalScaling(const int16_t* input, int input_length, | 74 int16_t SignalScaling(const int16_t* input, size_t input_length, |
75 const int16_t* expanded_signal, | 75 const int16_t* expanded_signal, |
76 int16_t* expanded_max, int16_t* input_max) const; | 76 int16_t* expanded_max, int16_t* input_max) const; |
77 | 77 |
78 // Downsamples |input| (|input_length| samples) and |expanded_signal| to | 78 // Downsamples |input| (|input_length| samples) and |expanded_signal| to |
79 // 4 kHz sample rate. The downsampled signals are written to | 79 // 4 kHz sample rate. The downsampled signals are written to |
80 // |input_downsampled_| and |expanded_downsampled_|, respectively. | 80 // |input_downsampled_| and |expanded_downsampled_|, respectively. |
81 void Downsample(const int16_t* input, int input_length, | 81 void Downsample(const int16_t* input, size_t input_length, |
82 const int16_t* expanded_signal, int expanded_length); | 82 const int16_t* expanded_signal, size_t expanded_length); |
83 | 83 |
84 // Calculates cross-correlation between |input_downsampled_| and | 84 // Calculates cross-correlation between |input_downsampled_| and |
85 // |expanded_downsampled_|, and finds the correlation maximum. The maximizing | 85 // |expanded_downsampled_|, and finds the correlation maximum. The maximizing |
86 // lag is returned. | 86 // lag is returned. |
87 int16_t CorrelateAndPeakSearch(int16_t expanded_max, int16_t input_max, | 87 size_t CorrelateAndPeakSearch(int16_t expanded_max, int16_t input_max, |
88 int start_position, int input_length, | 88 size_t start_position, size_t input_length, |
89 int expand_period) const; | 89 size_t expand_period) const; |
90 | 90 |
91 const int fs_mult_; // fs_hz_ / 8000. | 91 const size_t fs_mult_; // fs_hz_ / 8000. |
hlundin-webrtc
2015/08/10 11:30:01
Not size_t
| |
92 const int timestamps_per_call_; | 92 const size_t timestamps_per_call_; |
93 Expand* expand_; | 93 Expand* expand_; |
94 SyncBuffer* sync_buffer_; | 94 SyncBuffer* sync_buffer_; |
95 int16_t expanded_downsampled_[kExpandDownsampLength]; | 95 int16_t expanded_downsampled_[kExpandDownsampLength]; |
96 int16_t input_downsampled_[kInputDownsampLength]; | 96 int16_t input_downsampled_[kInputDownsampLength]; |
97 AudioMultiVector expanded_; | 97 AudioMultiVector expanded_; |
98 | 98 |
99 DISALLOW_COPY_AND_ASSIGN(Merge); | 99 DISALLOW_COPY_AND_ASSIGN(Merge); |
100 }; | 100 }; |
101 | 101 |
102 } // namespace webrtc | 102 } // namespace webrtc |
103 #endif // WEBRTC_MODULES_AUDIO_CODING_NETEQ_MERGE_H_ | 103 #endif // WEBRTC_MODULES_AUDIO_CODING_NETEQ_MERGE_H_ |
OLD | NEW |