| Index: webrtc/tools/event_log_visualizer/analyzer.cc
|
| diff --git a/webrtc/tools/event_log_visualizer/analyzer.cc b/webrtc/tools/event_log_visualizer/analyzer.cc
|
| index 5c0433a35807a2fb62d4e22b8f671da1d13ccb74..0125914e57a8f6963f72d666b1c21d02a9ae9e32 100644
|
| --- a/webrtc/tools/event_log_visualizer/analyzer.cc
|
| +++ b/webrtc/tools/event_log_visualizer/analyzer.cc
|
| @@ -451,6 +451,10 @@ EventLogAnalyzer::EventLogAnalyzer(const ParsedRtcEventLog& log)
|
| break;
|
| }
|
| case ParsedRtcEventLog::AUDIO_NETWORK_ADAPTATION_EVENT: {
|
| + AudioNetworkAdaptationEvent ana_event;
|
| + ana_event.timestamp = parsed_log_.GetTimestamp(i);
|
| + parsed_log_.GetAudioNetworkAdaptation(i, &ana_event.config);
|
| + audio_network_adaptation_events_.push_back(ana_event);
|
| break;
|
| }
|
| case ParsedRtcEventLog::UNKNOWN_EVENT: {
|
| @@ -532,6 +536,21 @@ std::string EventLogAnalyzer::GetStreamName(StreamId stream_id) const {
|
| return name.str();
|
| }
|
|
|
| +void EventLogAnalyzer::FillAudioEncoderTimeSeries(
|
| + Plot* plot,
|
| + rtc::FunctionView<rtc::Optional<float>(
|
| + const AudioNetworkAdaptationEvent& ana_event)> get_y) const {
|
| + plot->series_list_.push_back(TimeSeries());
|
| + plot->series_list_.back().style = LINE_DOT_GRAPH;
|
| + for (auto& ana_event : audio_network_adaptation_events_) {
|
| + rtc::Optional<float> y = get_y(ana_event);
|
| + if (y) {
|
| + float x = static_cast<float>(ana_event.timestamp - begin_time_) / 1000000;
|
| + plot->series_list_.back().points.emplace_back(x, *y);
|
| + }
|
| + }
|
| +}
|
| +
|
| void EventLogAnalyzer::CreatePacketGraph(PacketDirection desired_direction,
|
| Plot* plot) {
|
| for (auto& kv : rtp_packets_) {
|
| @@ -1275,5 +1294,92 @@ void EventLogAnalyzer::CreateTimestampGraph(Plot* plot) {
|
| plot->SetSuggestedYAxis(0, 1, "Timestamp (90khz)", kBottomMargin, kTopMargin);
|
| plot->SetTitle("Timestamps");
|
| }
|
| +
|
| +void EventLogAnalyzer::CreateAudioEncoderTargetBitrateGraph(Plot* plot) {
|
| + FillAudioEncoderTimeSeries(
|
| + plot, [](const AudioNetworkAdaptationEvent& ana_event) {
|
| + if (ana_event.config.bitrate_bps)
|
| + return rtc::Optional<float>(
|
| + static_cast<float>(*ana_event.config.bitrate_bps));
|
| + return rtc::Optional<float>();
|
| + });
|
| + plot->series_list_.back().label = "Audio encoder target bitrate";
|
| + plot->SetXAxis(0, call_duration_s_, "Time (s)", kLeftMargin, kRightMargin);
|
| + plot->SetSuggestedYAxis(0, 1, "Bitrate (bps)", kBottomMargin, kTopMargin);
|
| + plot->SetTitle("Reported audio encoder target bitrate");
|
| +}
|
| +
|
| +void EventLogAnalyzer::CreateAudioEncoderFrameLengthGraph(Plot* plot) {
|
| + FillAudioEncoderTimeSeries(
|
| + plot, [](const AudioNetworkAdaptationEvent& ana_event) {
|
| + if (ana_event.config.frame_length_ms)
|
| + return rtc::Optional<float>(
|
| + static_cast<float>(*ana_event.config.frame_length_ms));
|
| + return rtc::Optional<float>();
|
| + });
|
| + plot->series_list_.back().label = "Audio encoder frame length";
|
| + plot->SetXAxis(0, call_duration_s_, "Time (s)", kLeftMargin, kRightMargin);
|
| + plot->SetSuggestedYAxis(0, 1, "Frame length (ms)", kBottomMargin, kTopMargin);
|
| + plot->SetTitle("Reported audio encoder frame length");
|
| +}
|
| +
|
| +void EventLogAnalyzer::CreateAudioEncoderUplinkPacketLossFractionGraph(
|
| + Plot* plot) {
|
| + FillAudioEncoderTimeSeries(
|
| + plot, [&](const AudioNetworkAdaptationEvent& ana_event) {
|
| + if (ana_event.config.uplink_packet_loss_fraction)
|
| + return rtc::Optional<float>(static_cast<float>(
|
| + *ana_event.config.uplink_packet_loss_fraction));
|
| + return rtc::Optional<float>();
|
| + });
|
| + plot->series_list_.back().label = "Audio encoder uplink packet loss fraction";
|
| + plot->SetXAxis(0, call_duration_s_, "Time (s)", kLeftMargin, kRightMargin);
|
| + plot->SetSuggestedYAxis(0, 10, "Percent lost packets", kBottomMargin,
|
| + kTopMargin);
|
| + plot->SetTitle("Reported audio encoder lost packets");
|
| +}
|
| +
|
| +void EventLogAnalyzer::CreateAudioEncoderEnableFecGraph(Plot* plot) {
|
| + FillAudioEncoderTimeSeries(
|
| + plot, [&](const AudioNetworkAdaptationEvent& ana_event) {
|
| + if (ana_event.config.enable_fec)
|
| + return rtc::Optional<float>(
|
| + static_cast<float>(*ana_event.config.enable_fec));
|
| + return rtc::Optional<float>();
|
| + });
|
| + plot->series_list_.back().label = "Audio encoder FEC";
|
| + plot->SetXAxis(0, call_duration_s_, "Time (s)", kLeftMargin, kRightMargin);
|
| + plot->SetSuggestedYAxis(0, 1, "FEC (false/true)", kBottomMargin, kTopMargin);
|
| + plot->SetTitle("Reported audio encoder FEC");
|
| +}
|
| +
|
| +void EventLogAnalyzer::CreateAudioEncoderEnableDtxGraph(Plot* plot) {
|
| + FillAudioEncoderTimeSeries(
|
| + plot, [&](const AudioNetworkAdaptationEvent& ana_event) {
|
| + if (ana_event.config.enable_dtx)
|
| + return rtc::Optional<float>(
|
| + static_cast<float>(*ana_event.config.enable_dtx));
|
| + return rtc::Optional<float>();
|
| + });
|
| + plot->series_list_.back().label = "Audio encoder DTX";
|
| + plot->SetXAxis(0, call_duration_s_, "Time (s)", kLeftMargin, kRightMargin);
|
| + plot->SetSuggestedYAxis(0, 1, "DTX (false/true)", kBottomMargin, kTopMargin);
|
| + plot->SetTitle("Reported audio encoder DTX");
|
| +}
|
| +
|
| +void EventLogAnalyzer::CreateAudioEncoderNumChannelsGraph(Plot* plot) {
|
| + FillAudioEncoderTimeSeries(
|
| + plot, [&](const AudioNetworkAdaptationEvent& ana_event) {
|
| + if (ana_event.config.num_channels)
|
| + return rtc::Optional<float>(
|
| + static_cast<float>(*ana_event.config.num_channels));
|
| + return rtc::Optional<float>();
|
| + });
|
| + plot->series_list_.back().label = "Audio encoder number of channels";
|
| + plot->SetXAxis(0, call_duration_s_, "Time (s)", kLeftMargin, kRightMargin);
|
| + plot->SetSuggestedYAxis(0, 1, "Number of channels (1 (mono)/2 (stereo))",
|
| + kBottomMargin, kTopMargin);
|
| + plot->SetTitle("Reported audio encoder number of channels");
|
| +}
|
| } // namespace plotting
|
| } // namespace webrtc
|
|
|