| Index: webrtc/common_audio/vad/vad_core.c
|
| diff --git a/webrtc/common_audio/vad/vad_core.c b/webrtc/common_audio/vad/vad_core.c
|
| index d246a4d954e6164084aa4b7b903e7dfaf8f93308..51797eed549440f0b10f5dea7312792b2f4bb819 100644
|
| --- a/webrtc/common_audio/vad/vad_core.c
|
| +++ b/webrtc/common_audio/vad/vad_core.c
|
| @@ -122,7 +122,7 @@ static int32_t WeightedAverage(int16_t* data, int16_t offset,
|
| //
|
| // - returns : the VAD decision (0 - noise, 1 - speech).
|
| static int16_t GmmProbability(VadInstT* self, int16_t* features,
|
| - int16_t total_power, int frame_length) {
|
| + int16_t total_power, size_t frame_length) {
|
| int channel, k;
|
| int16_t feature_minimum;
|
| int16_t h0, h1;
|
| @@ -596,16 +596,16 @@ int WebRtcVad_set_mode_core(VadInstT* self, int mode) {
|
| // probability for both speech and background noise.
|
|
|
| int WebRtcVad_CalcVad48khz(VadInstT* inst, const int16_t* speech_frame,
|
| - int frame_length) {
|
| + size_t frame_length) {
|
| int vad;
|
| - int i;
|
| + size_t i;
|
| int16_t speech_nb[240]; // 30 ms in 8 kHz.
|
| // |tmp_mem| is a temporary memory used by resample function, length is
|
| // frame length in 10 ms (480 samples) + 256 extra.
|
| int32_t tmp_mem[480 + 256] = { 0 };
|
| - const int kFrameLen10ms48khz = 480;
|
| - const int kFrameLen10ms8khz = 80;
|
| - int num_10ms_frames = frame_length / kFrameLen10ms48khz;
|
| + const size_t kFrameLen10ms48khz = 480;
|
| + const size_t kFrameLen10ms8khz = 80;
|
| + size_t num_10ms_frames = frame_length / kFrameLen10ms48khz;
|
|
|
| for (i = 0; i < num_10ms_frames; i++) {
|
| WebRtcSpl_Resample48khzTo8khz(speech_frame,
|
| @@ -621,9 +621,10 @@ int WebRtcVad_CalcVad48khz(VadInstT* inst, const int16_t* speech_frame,
|
| }
|
|
|
| int WebRtcVad_CalcVad32khz(VadInstT* inst, const int16_t* speech_frame,
|
| - int frame_length)
|
| + size_t frame_length)
|
| {
|
| - int len, vad;
|
| + size_t len;
|
| + int vad;
|
| int16_t speechWB[480]; // Downsampled speech frame: 960 samples (30ms in SWB)
|
| int16_t speechNB[240]; // Downsampled speech frame: 480 samples (30ms in WB)
|
|
|
| @@ -643,9 +644,10 @@ int WebRtcVad_CalcVad32khz(VadInstT* inst, const int16_t* speech_frame,
|
| }
|
|
|
| int WebRtcVad_CalcVad16khz(VadInstT* inst, const int16_t* speech_frame,
|
| - int frame_length)
|
| + size_t frame_length)
|
| {
|
| - int len, vad;
|
| + size_t len;
|
| + int vad;
|
| int16_t speechNB[240]; // Downsampled speech frame: 480 samples (30ms in WB)
|
|
|
| // Wideband: Downsample signal before doing VAD
|
| @@ -659,7 +661,7 @@ int WebRtcVad_CalcVad16khz(VadInstT* inst, const int16_t* speech_frame,
|
| }
|
|
|
| int WebRtcVad_CalcVad8khz(VadInstT* inst, const int16_t* speech_frame,
|
| - int frame_length)
|
| + size_t frame_length)
|
| {
|
| int16_t feature_vector[kNumChannels], total_power;
|
|
|
|
|