| OLD | NEW |
| 1 /* | 1 /* |
| 2 * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved. | 2 * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved. |
| 3 * | 3 * |
| 4 * Use of this source code is governed by a BSD-style license | 4 * Use of this source code is governed by a BSD-style license |
| 5 * that can be found in the LICENSE file in the root of the source | 5 * that can be found in the LICENSE file in the root of the source |
| 6 * tree. An additional intellectual property rights grant can be found | 6 * tree. An additional intellectual property rights grant can be found |
| 7 * in the file PATENTS. All contributing project authors may | 7 * in the file PATENTS. All contributing project authors may |
| 8 * be found in the AUTHORS file in the root of the source tree. | 8 * be found in the AUTHORS file in the root of the source tree. |
| 9 */ | 9 */ |
| 10 | 10 |
| 11 #include "webrtc/modules/interface/module_common_types.h" | 11 #include "webrtc/modules/interface/module_common_types.h" |
| 12 #include "webrtc/modules/utility/interface/audio_frame_operations.h" | 12 #include "webrtc/modules/utility/interface/audio_frame_operations.h" |
| 13 | 13 |
| 14 namespace webrtc { | 14 namespace webrtc { |
| 15 | 15 |
| 16 void AudioFrameOperations::MonoToStereo(const int16_t* src_audio, | 16 void AudioFrameOperations::MonoToStereo(const int16_t* src_audio, |
| 17 int samples_per_channel, | 17 size_t samples_per_channel, |
| 18 int16_t* dst_audio) { | 18 int16_t* dst_audio) { |
| 19 for (int i = 0; i < samples_per_channel; i++) { | 19 for (size_t i = 0; i < samples_per_channel; i++) { |
| 20 dst_audio[2 * i] = src_audio[i]; | 20 dst_audio[2 * i] = src_audio[i]; |
| 21 dst_audio[2 * i + 1] = src_audio[i]; | 21 dst_audio[2 * i + 1] = src_audio[i]; |
| 22 } | 22 } |
| 23 } | 23 } |
| 24 | 24 |
| 25 int AudioFrameOperations::MonoToStereo(AudioFrame* frame) { | 25 int AudioFrameOperations::MonoToStereo(AudioFrame* frame) { |
| 26 if (frame->num_channels_ != 1) { | 26 if (frame->num_channels_ != 1) { |
| 27 return -1; | 27 return -1; |
| 28 } | 28 } |
| 29 if ((frame->samples_per_channel_ * 2) >= AudioFrame::kMaxDataSizeSamples) { | 29 if ((frame->samples_per_channel_ * 2) >= AudioFrame::kMaxDataSizeSamples) { |
| 30 // Not enough memory to expand from mono to stereo. | 30 // Not enough memory to expand from mono to stereo. |
| 31 return -1; | 31 return -1; |
| 32 } | 32 } |
| 33 | 33 |
| 34 int16_t data_copy[AudioFrame::kMaxDataSizeSamples]; | 34 int16_t data_copy[AudioFrame::kMaxDataSizeSamples]; |
| 35 memcpy(data_copy, frame->data_, | 35 memcpy(data_copy, frame->data_, |
| 36 sizeof(int16_t) * frame->samples_per_channel_); | 36 sizeof(int16_t) * frame->samples_per_channel_); |
| 37 MonoToStereo(data_copy, frame->samples_per_channel_, frame->data_); | 37 MonoToStereo(data_copy, frame->samples_per_channel_, frame->data_); |
| 38 frame->num_channels_ = 2; | 38 frame->num_channels_ = 2; |
| 39 | 39 |
| 40 return 0; | 40 return 0; |
| 41 } | 41 } |
| 42 | 42 |
| 43 void AudioFrameOperations::StereoToMono(const int16_t* src_audio, | 43 void AudioFrameOperations::StereoToMono(const int16_t* src_audio, |
| 44 int samples_per_channel, | 44 size_t samples_per_channel, |
| 45 int16_t* dst_audio) { | 45 int16_t* dst_audio) { |
| 46 for (int i = 0; i < samples_per_channel; i++) { | 46 for (size_t i = 0; i < samples_per_channel; i++) { |
| 47 dst_audio[i] = (src_audio[2 * i] + src_audio[2 * i + 1]) >> 1; | 47 dst_audio[i] = (src_audio[2 * i] + src_audio[2 * i + 1]) >> 1; |
| 48 } | 48 } |
| 49 } | 49 } |
| 50 | 50 |
| 51 int AudioFrameOperations::StereoToMono(AudioFrame* frame) { | 51 int AudioFrameOperations::StereoToMono(AudioFrame* frame) { |
| 52 if (frame->num_channels_ != 2) { | 52 if (frame->num_channels_ != 2) { |
| 53 return -1; | 53 return -1; |
| 54 } | 54 } |
| 55 | 55 |
| 56 StereoToMono(frame->data_, frame->samples_per_channel_, frame->data_); | 56 StereoToMono(frame->data_, frame->samples_per_channel_, frame->data_); |
| 57 frame->num_channels_ = 1; | 57 frame->num_channels_ = 1; |
| 58 | 58 |
| 59 return 0; | 59 return 0; |
| 60 } | 60 } |
| 61 | 61 |
| 62 void AudioFrameOperations::SwapStereoChannels(AudioFrame* frame) { | 62 void AudioFrameOperations::SwapStereoChannels(AudioFrame* frame) { |
| 63 if (frame->num_channels_ != 2) return; | 63 if (frame->num_channels_ != 2) return; |
| 64 | 64 |
| 65 for (int i = 0; i < frame->samples_per_channel_ * 2; i += 2) { | 65 for (size_t i = 0; i < frame->samples_per_channel_ * 2; i += 2) { |
| 66 int16_t temp_data = frame->data_[i]; | 66 int16_t temp_data = frame->data_[i]; |
| 67 frame->data_[i] = frame->data_[i + 1]; | 67 frame->data_[i] = frame->data_[i + 1]; |
| 68 frame->data_[i + 1] = temp_data; | 68 frame->data_[i + 1] = temp_data; |
| 69 } | 69 } |
| 70 } | 70 } |
| 71 | 71 |
| 72 void AudioFrameOperations::Mute(AudioFrame& frame) { | 72 void AudioFrameOperations::Mute(AudioFrame& frame) { |
| 73 memset(frame.data_, 0, sizeof(int16_t) * | 73 memset(frame.data_, 0, sizeof(int16_t) * |
| 74 frame.samples_per_channel_ * frame.num_channels_); | 74 frame.samples_per_channel_ * frame.num_channels_); |
| 75 } | 75 } |
| 76 | 76 |
| 77 int AudioFrameOperations::Scale(float left, float right, AudioFrame& frame) { | 77 int AudioFrameOperations::Scale(float left, float right, AudioFrame& frame) { |
| 78 if (frame.num_channels_ != 2) { | 78 if (frame.num_channels_ != 2) { |
| 79 return -1; | 79 return -1; |
| 80 } | 80 } |
| 81 | 81 |
| 82 for (int i = 0; i < frame.samples_per_channel_; i++) { | 82 for (size_t i = 0; i < frame.samples_per_channel_; i++) { |
| 83 frame.data_[2 * i] = | 83 frame.data_[2 * i] = |
| 84 static_cast<int16_t>(left * frame.data_[2 * i]); | 84 static_cast<int16_t>(left * frame.data_[2 * i]); |
| 85 frame.data_[2 * i + 1] = | 85 frame.data_[2 * i + 1] = |
| 86 static_cast<int16_t>(right * frame.data_[2 * i + 1]); | 86 static_cast<int16_t>(right * frame.data_[2 * i + 1]); |
| 87 } | 87 } |
| 88 return 0; | 88 return 0; |
| 89 } | 89 } |
| 90 | 90 |
| 91 int AudioFrameOperations::ScaleWithSat(float scale, AudioFrame& frame) { | 91 int AudioFrameOperations::ScaleWithSat(float scale, AudioFrame& frame) { |
| 92 int32_t temp_data = 0; | 92 int32_t temp_data = 0; |
| 93 | 93 |
| 94 // Ensure that the output result is saturated [-32768, +32767]. | 94 // Ensure that the output result is saturated [-32768, +32767]. |
| 95 for (int i = 0; i < frame.samples_per_channel_ * frame.num_channels_; | 95 for (size_t i = 0; i < frame.samples_per_channel_ * frame.num_channels_; |
| 96 i++) { | 96 i++) { |
| 97 temp_data = static_cast<int32_t>(scale * frame.data_[i]); | 97 temp_data = static_cast<int32_t>(scale * frame.data_[i]); |
| 98 if (temp_data < -32768) { | 98 if (temp_data < -32768) { |
| 99 frame.data_[i] = -32768; | 99 frame.data_[i] = -32768; |
| 100 } else if (temp_data > 32767) { | 100 } else if (temp_data > 32767) { |
| 101 frame.data_[i] = 32767; | 101 frame.data_[i] = 32767; |
| 102 } else { | 102 } else { |
| 103 frame.data_[i] = static_cast<int16_t>(temp_data); | 103 frame.data_[i] = static_cast<int16_t>(temp_data); |
| 104 } | 104 } |
| 105 } | 105 } |
| 106 return 0; | 106 return 0; |
| 107 } | 107 } |
| 108 | 108 |
| 109 } // namespace webrtc | 109 } // namespace webrtc |
| OLD | NEW |