Index: webrtc/modules/audio_mixer/frame_combiner_unittest.cc |
diff --git a/webrtc/modules/audio_mixer/frame_combiner_unittest.cc b/webrtc/modules/audio_mixer/frame_combiner_unittest.cc |
index 13c66012f99f27c4fd30d820e19c0166fcf3d1f6..80b2798213ec9c840f8ad829bbc12f356b3ea60d 100644 |
--- a/webrtc/modules/audio_mixer/frame_combiner_unittest.cc |
+++ b/webrtc/modules/audio_mixer/frame_combiner_unittest.cc |
@@ -95,9 +95,11 @@ TEST(FrameCombiner, CombiningZeroFramesShouldProduceSilence) { |
combiner.Combine(frames_to_combine, number_of_channels, rate, |
&audio_frame_for_mixing); |
+ const int16_t* audio_frame_for_mixing_data = |
+ audio_frame_for_mixing.data(); |
const std::vector<int16_t> mixed_data( |
- audio_frame_for_mixing.data_, |
- audio_frame_for_mixing.data_ + number_of_channels * rate / 100); |
+ audio_frame_for_mixing_data, |
+ audio_frame_for_mixing_data + number_of_channels * rate / 100); |
const std::vector<int16_t> expected(number_of_channels * rate / 100, 0); |
EXPECT_EQ(mixed_data, expected); |
@@ -112,15 +114,17 @@ TEST(FrameCombiner, CombiningOneFrameShouldNotChangeFrame) { |
SCOPED_TRACE(ProduceDebugText(rate, number_of_channels, 1)); |
SetUpFrames(rate, number_of_channels); |
- std::iota(frame1.data_, frame1.data_ + number_of_channels * rate / 100, |
- 0); |
+ int16_t* frame1_data = frame1.mutable_data(); |
+ std::iota(frame1_data, frame1_data + number_of_channels * rate / 100, 0); |
const std::vector<AudioFrame*> frames_to_combine = {&frame1}; |
combiner.Combine(frames_to_combine, number_of_channels, rate, |
&audio_frame_for_mixing); |
+ const int16_t* audio_frame_for_mixing_data = |
+ audio_frame_for_mixing.data(); |
const std::vector<int16_t> mixed_data( |
- audio_frame_for_mixing.data_, |
- audio_frame_for_mixing.data_ + number_of_channels * rate / 100); |
+ audio_frame_for_mixing_data, |
+ audio_frame_for_mixing_data + number_of_channels * rate / 100); |
std::vector<int16_t> expected(number_of_channels * rate / 100); |
std::iota(expected.begin(), expected.end(), 0); |