| Index: webrtc/modules/audio_processing/audio_buffer.cc
|
| diff --git a/webrtc/modules/audio_processing/audio_buffer.cc b/webrtc/modules/audio_processing/audio_buffer.cc
|
| index 579a5c24904aab28a857f9d7beca753888825404..5f90e0f54776f9e34c0e1c4f7ab918367dcd8000 100644
|
| --- a/webrtc/modules/audio_processing/audio_buffer.cc
|
| +++ b/webrtc/modules/audio_processing/audio_buffer.cc
|
| @@ -394,13 +394,14 @@ void AudioBuffer::DeinterleaveFrom(AudioFrame* frame) {
|
| } else {
|
| deinterleaved = input_buffer_->ibuf()->channels();
|
| }
|
| + // TODO(yujo): handle muted frames more efficiently.
|
| if (num_proc_channels_ == 1) {
|
| // Downmix and deinterleave simultaneously.
|
| - DownmixInterleavedToMono(frame->data_, input_num_frames_,
|
| + DownmixInterleavedToMono(frame->data(), input_num_frames_,
|
| num_input_channels_, deinterleaved[0]);
|
| } else {
|
| RTC_DCHECK_EQ(num_proc_channels_, num_input_channels_);
|
| - Deinterleave(frame->data_,
|
| + Deinterleave(frame->data(),
|
| input_num_frames_,
|
| num_proc_channels_,
|
| deinterleaved);
|
| @@ -437,12 +438,13 @@ void AudioBuffer::InterleaveTo(AudioFrame* frame, bool data_changed) const {
|
| data_ptr = output_buffer_.get();
|
| }
|
|
|
| + // TODO(yujo): handle muted frames more efficiently.
|
| if (frame->num_channels_ == num_channels_) {
|
| Interleave(data_ptr->ibuf()->channels(), output_num_frames_, num_channels_,
|
| - frame->data_);
|
| + frame->mutable_data());
|
| } else {
|
| UpmixMonoToInterleaved(data_ptr->ibuf()->channels()[0], output_num_frames_,
|
| - frame->num_channels_, frame->data_);
|
| + frame->num_channels_, frame->mutable_data());
|
| }
|
| }
|
|
|
|
|