| Index: webrtc/modules/audio_processing/audio_processing_unittest.cc | 
| diff --git a/webrtc/modules/audio_processing/audio_processing_unittest.cc b/webrtc/modules/audio_processing/audio_processing_unittest.cc | 
| index 4d97904c6869c25febb77933fc73fbf1760ddf7d..f37d04a56923b5b9756d3d3f3e23e2053b55f749 100644 | 
| --- a/webrtc/modules/audio_processing/audio_processing_unittest.cc | 
| +++ b/webrtc/modules/audio_processing/audio_processing_unittest.cc | 
| @@ -87,7 +87,7 @@ void ConvertToFloat(const int16_t* int_data, ChannelBuffer<float>* cb) { | 
| } | 
|  | 
| void ConvertToFloat(const AudioFrame& frame, ChannelBuffer<float>* cb) { | 
| -  ConvertToFloat(frame.data_, cb); | 
| +  ConvertToFloat(frame.data(), cb); | 
| } | 
|  | 
| // Number of channels including the keyboard channel. | 
| @@ -127,31 +127,34 @@ void CopyLeftToRightChannel(int16_t* stereo, size_t samples_per_channel) { | 
| } | 
| } | 
|  | 
| -void VerifyChannelsAreEqual(int16_t* stereo, size_t samples_per_channel) { | 
| +void VerifyChannelsAreEqual(const int16_t* stereo, size_t samples_per_channel) { | 
| for (size_t i = 0; i < samples_per_channel; i++) { | 
| EXPECT_EQ(stereo[i * 2 + 1], stereo[i * 2]); | 
| } | 
| } | 
|  | 
| void SetFrameTo(AudioFrame* frame, int16_t value) { | 
| +  int16_t* frame_data = frame->mutable_data(); | 
| for (size_t i = 0; i < frame->samples_per_channel_ * frame->num_channels_; | 
| ++i) { | 
| -    frame->data_[i] = value; | 
| +    frame_data[i] = value; | 
| } | 
| } | 
|  | 
| void SetFrameTo(AudioFrame* frame, int16_t left, int16_t right) { | 
| ASSERT_EQ(2u, frame->num_channels_); | 
| +  int16_t* frame_data = frame->mutable_data(); | 
| for (size_t i = 0; i < frame->samples_per_channel_ * 2; i += 2) { | 
| -    frame->data_[i] = left; | 
| -    frame->data_[i + 1] = right; | 
| +    frame_data[i] = left; | 
| +    frame_data[i + 1] = right; | 
| } | 
| } | 
|  | 
| void ScaleFrame(AudioFrame* frame, float scale) { | 
| +  int16_t* frame_data = frame->mutable_data(); | 
| for (size_t i = 0; i < frame->samples_per_channel_ * frame->num_channels_; | 
| ++i) { | 
| -    frame->data_[i] = FloatS16ToS16(frame->data_[i] * scale); | 
| +    frame_data[i] = FloatS16ToS16(frame_data[i] * scale); | 
| } | 
| } | 
|  | 
| @@ -162,7 +165,7 @@ bool FrameDataAreEqual(const AudioFrame& frame1, const AudioFrame& frame2) { | 
| if (frame1.num_channels_ != frame2.num_channels_) { | 
| return false; | 
| } | 
| -  if (memcmp(frame1.data_, frame2.data_, | 
| +  if (memcmp(frame1.data(), frame2.data(), | 
| frame1.samples_per_channel_ * frame1.num_channels_ * | 
| sizeof(int16_t))) { | 
| return false; | 
| @@ -205,9 +208,10 @@ T AbsValue(T a) { | 
|  | 
| int16_t MaxAudioFrame(const AudioFrame& frame) { | 
| const size_t length = frame.samples_per_channel_ * frame.num_channels_; | 
| -  int16_t max_data = AbsValue(frame.data_[0]); | 
| +  const int16_t* frame_data = frame.data(); | 
| +  int16_t max_data = AbsValue(frame_data[0]); | 
| for (size_t i = 1; i < length; i++) { | 
| -    max_data = std::max(max_data, AbsValue(frame.data_[i])); | 
| +    max_data = std::max(max_data, AbsValue(frame_data[i])); | 
| } | 
|  | 
| return max_data; | 
| @@ -534,7 +538,7 @@ bool ApmTest::ReadFrame(FILE* file, AudioFrame* frame, | 
| ChannelBuffer<float>* cb) { | 
| // The files always contain stereo audio. | 
| size_t frame_size = frame->samples_per_channel_ * 2; | 
| -  size_t read_count = fread(frame->data_, | 
| +  size_t read_count = fread(frame->mutable_data(), | 
| sizeof(int16_t), | 
| frame_size, | 
| file); | 
| @@ -545,7 +549,7 @@ bool ApmTest::ReadFrame(FILE* file, AudioFrame* frame, | 
| } | 
|  | 
| if (frame->num_channels_ == 1) { | 
| -    MixStereoToMono(frame->data_, frame->data_, | 
| +    MixStereoToMono(frame->data(), frame->mutable_data(), | 
| frame->samples_per_channel_); | 
| } | 
|  | 
| @@ -1603,11 +1607,13 @@ TEST_F(ApmTest, IdenticalInputChannelsResultInIdenticalOutputChannels) { | 
| ASSERT_EQ(0, feof(far_file_)); | 
| ASSERT_EQ(0, feof(near_file_)); | 
| while (ReadFrame(far_file_, revframe_) && ReadFrame(near_file_, frame_)) { | 
| -      CopyLeftToRightChannel(revframe_->data_, revframe_->samples_per_channel_); | 
| +      CopyLeftToRightChannel(revframe_->mutable_data(), | 
| +                             revframe_->samples_per_channel_); | 
|  | 
| ASSERT_EQ(kNoErr, apm_->ProcessReverseStream(revframe_)); | 
|  | 
| -      CopyLeftToRightChannel(frame_->data_, frame_->samples_per_channel_); | 
| +      CopyLeftToRightChannel(frame_->mutable_data(), | 
| +                             frame_->samples_per_channel_); | 
| frame_->vad_activity_ = AudioFrame::kVadUnknown; | 
|  | 
| ASSERT_EQ(kNoErr, apm_->set_stream_delay_ms(0)); | 
| @@ -1617,7 +1623,7 @@ TEST_F(ApmTest, IdenticalInputChannelsResultInIdenticalOutputChannels) { | 
| ASSERT_EQ(kNoErr, apm_->ProcessStream(frame_)); | 
| analog_level = apm_->gain_control()->stream_analog_level(); | 
|  | 
| -      VerifyChannelsAreEqual(frame_->data_, frame_->samples_per_channel_); | 
| +      VerifyChannelsAreEqual(frame_->data(), frame_->samples_per_channel_); | 
| } | 
| rewind(far_file_); | 
| rewind(near_file_); | 
| @@ -1749,7 +1755,7 @@ void ApmTest::ProcessDebugDump(const std::string& in_filename, | 
| msg.channel(i).size()); | 
| } | 
| } else { | 
| -        memcpy(revframe_->data_, msg.data().data(), msg.data().size()); | 
| +        memcpy(revframe_->mutable_data(), msg.data().data(), msg.data().size()); | 
| if (format == kFloatFormat) { | 
| // We're using an int16 input file; convert to float. | 
| ConvertToFloat(*revframe_, revfloat_cb_.get()); | 
| @@ -1780,7 +1786,8 @@ void ApmTest::ProcessDebugDump(const std::string& in_filename, | 
| msg.input_channel(i).size()); | 
| } | 
| } else { | 
| -        memcpy(frame_->data_, msg.input_data().data(), msg.input_data().size()); | 
| +        memcpy(frame_->mutable_data(), msg.input_data().data(), | 
| +               msg.input_data().size()); | 
| if (format == kFloatFormat) { | 
| // We're using an int16 input file; convert to float. | 
| ConvertToFloat(*frame_, float_cb_.get()); | 
| @@ -1989,7 +1996,7 @@ TEST_F(ApmTest, FloatAndIntInterfacesGiveSimilarResults) { | 
| EXPECT_NOERR(fapm->gain_control()->set_stream_analog_level(analog_level)); | 
|  | 
| EXPECT_NOERR(apm_->ProcessStream(frame_)); | 
| -      Deinterleave(frame_->data_, samples_per_channel, num_output_channels, | 
| +      Deinterleave(frame_->data(), samples_per_channel, num_output_channels, | 
| output_int16.channels()); | 
|  | 
| EXPECT_NOERR(fapm->ProcessStream( | 
| @@ -2153,7 +2160,7 @@ TEST_F(ApmTest, Process) { | 
| ns_speech_prob_average += apm_->noise_suppression()->speech_probability(); | 
|  | 
| size_t frame_size = frame_->samples_per_channel_ * frame_->num_channels_; | 
| -      size_t write_count = fwrite(frame_->data_, | 
| +      size_t write_count = fwrite(frame_->data(), | 
| sizeof(int16_t), | 
| frame_size, | 
| out_file_); | 
|  |