Index: webrtc/tools/agc/activity_metric.cc |
diff --git a/webrtc/tools/agc/activity_metric.cc b/webrtc/tools/agc/activity_metric.cc |
index a98730a3422bd1c3966f9bae0da7237a188eee7a..d95572d21429bd8d1b14ba1d3550f2d2b7cffc02 100644 |
--- a/webrtc/tools/agc/activity_metric.cc |
+++ b/webrtc/tools/agc/activity_metric.cc |
@@ -62,11 +62,12 @@ static void DitherSilence(AudioFrame* frame) { |
const double sum_squared_silence = kRmsSilence * kRmsSilence * |
frame->samples_per_channel_; |
double sum_squared = 0; |
+ int16_t* frame_data = frame->mutable_data(); |
for (size_t n = 0; n < frame->samples_per_channel_; n++) |
- sum_squared += frame->data_[n] * frame->data_[n]; |
+ sum_squared += frame_data[n] * frame_data[n]; |
if (sum_squared <= sum_squared_silence) { |
for (size_t n = 0; n < frame->samples_per_channel_; n++) |
- frame->data_[n] = (rand() & 0xF) - 8; // NOLINT: ignore non-threadsafe. |
+ frame_data[n] = (rand() & 0xF) - 8; // NOLINT: ignore non-threadsafe. |
} |
} |
@@ -103,10 +104,11 @@ class AgcStat { |
return -1; |
video_vad_[video_index_++] = p_video; |
AudioFeatures features; |
+ const int16_t* frame_data = frame.data(); |
audio_processing_->ExtractFeatures( |
- frame.data_, frame.samples_per_channel_, &features); |
+ frame_data, frame.samples_per_channel_, &features); |
if (FLAGS_standalone_vad) { |
- standalone_vad_->AddAudio(frame.data_, |
+ standalone_vad_->AddAudio(frame_data, |
frame.samples_per_channel_); |
} |
if (features.num_frames > 0) { |
@@ -251,7 +253,7 @@ void void_main(int argc, char* argv[]) { |
bool in_false_positive_region = false; |
int total_false_positive_duration = 0; |
bool video_adapted = false; |
- while (kSamplesToRead == fread(frame.data_, sizeof(int16_t), |
+ while (kSamplesToRead == fread(frame.mutable_data(), sizeof(int16_t), |
kSamplesToRead, pcm_fid)) { |
assert(true_vad_index < kMaxNumFrames); |
ASSERT_EQ(1u, fread(&true_vad[true_vad_index], sizeof(*true_vad), 1, |