| Index: webrtc/modules/video_coding/jitter_buffer.cc
|
| diff --git a/webrtc/modules/video_coding/jitter_buffer.cc b/webrtc/modules/video_coding/jitter_buffer.cc
|
| index a1142bb49d7ec935593eee508b02a48cc9225739..663bf03a7f2ddad7c7f9f6e3b5f972ba1db2df33 100644
|
| --- a/webrtc/modules/video_coding/jitter_buffer.cc
|
| +++ b/webrtc/modules/video_coding/jitter_buffer.cc
|
| @@ -93,7 +93,7 @@ int FrameList::RecycleFramesUntilKeyFrame(FrameList::iterator* key_frame_it,
|
| }
|
|
|
| void FrameList::CleanUpOldOrEmptyFrames(VCMDecodingState* decoding_state,
|
| - UnorderedFrameList* free_frames) {
|
| + UnorderedFrameList* free_frames) {
|
| while (!empty()) {
|
| VCMFrameBuffer* oldest_frame = Front();
|
| bool remove_frame = false;
|
| @@ -282,14 +282,15 @@ void VCMJitterBuffer::UpdateHistograms() {
|
| }
|
|
|
| RTC_HISTOGRAM_PERCENTAGE("WebRTC.Video.DiscardedPacketsInPercent",
|
| - num_discarded_packets_ * 100 / num_packets_);
|
| + num_discarded_packets_ * 100 / num_packets_);
|
| RTC_HISTOGRAM_PERCENTAGE("WebRTC.Video.DuplicatedPacketsInPercent",
|
| - num_duplicated_packets_ * 100 / num_packets_);
|
| + num_duplicated_packets_ * 100 / num_packets_);
|
|
|
| int total_frames =
|
| receive_statistics_.key_frames + receive_statistics_.delta_frames;
|
| if (total_frames > 0) {
|
| - RTC_HISTOGRAM_COUNTS_100("WebRTC.Video.CompleteFramesReceivedPerSecond",
|
| + RTC_HISTOGRAM_COUNTS_100(
|
| + "WebRTC.Video.CompleteFramesReceivedPerSecond",
|
| static_cast<int>((total_frames / elapsed_sec) + 0.5f));
|
| RTC_HISTOGRAM_COUNTS_1000(
|
| "WebRTC.Video.KeyFramesReceivedInPermille",
|
| @@ -430,8 +431,8 @@ void VCMJitterBuffer::IncomingRateStatistics(unsigned int* framerate,
|
| if (incoming_bit_count_ == 0) {
|
| *bitrate = 0;
|
| } else {
|
| - *bitrate = 10 * ((100 * incoming_bit_count_) /
|
| - static_cast<unsigned int>(diff));
|
| + *bitrate =
|
| + 10 * ((100 * incoming_bit_count_) / static_cast<unsigned int>(diff));
|
| }
|
| incoming_bit_rate_ = *bitrate;
|
|
|
| @@ -472,8 +473,8 @@ bool VCMJitterBuffer::CompleteSequenceWithNextFrame() {
|
|
|
| // Returns immediately or a |max_wait_time_ms| ms event hang waiting for a
|
| // complete frame, |max_wait_time_ms| decided by caller.
|
| -bool VCMJitterBuffer::NextCompleteTimestamp(
|
| - uint32_t max_wait_time_ms, uint32_t* timestamp) {
|
| +bool VCMJitterBuffer::NextCompleteTimestamp(uint32_t max_wait_time_ms,
|
| + uint32_t* timestamp) {
|
| crit_sect_->Enter();
|
| if (!running_) {
|
| crit_sect_->Leave();
|
| @@ -483,13 +484,13 @@ bool VCMJitterBuffer::NextCompleteTimestamp(
|
|
|
| if (decodable_frames_.empty() ||
|
| decodable_frames_.Front()->GetState() != kStateComplete) {
|
| - const int64_t end_wait_time_ms = clock_->TimeInMilliseconds() +
|
| - max_wait_time_ms;
|
| + const int64_t end_wait_time_ms =
|
| + clock_->TimeInMilliseconds() + max_wait_time_ms;
|
| int64_t wait_time_ms = max_wait_time_ms;
|
| while (wait_time_ms > 0) {
|
| crit_sect_->Leave();
|
| const EventTypeWrapper ret =
|
| - frame_event_->Wait(static_cast<uint32_t>(wait_time_ms));
|
| + frame_event_->Wait(static_cast<uint32_t>(wait_time_ms));
|
| crit_sect_->Enter();
|
| if (ret == kEventSignaled) {
|
| // Are we shutting down the jitter buffer?
|
| @@ -547,8 +548,8 @@ bool VCMJitterBuffer::NextMaybeIncompleteTimestamp(uint32_t* timestamp) {
|
| // If we have exactly one frame in the buffer, release it only if it is
|
| // complete. We know decodable_frames_ is not empty due to the previous
|
| // check.
|
| - if (decodable_frames_.size() == 1 && incomplete_frames_.empty()
|
| - && oldest_frame->GetState() != kStateComplete) {
|
| + if (decodable_frames_.size() == 1 && incomplete_frames_.empty() &&
|
| + oldest_frame->GetState() != kStateComplete) {
|
| return false;
|
| }
|
| }
|
| @@ -587,8 +588,7 @@ VCMEncodedFrame* VCMJitterBuffer::ExtractAndSetDecode(uint32_t timestamp) {
|
| } else {
|
| // Wait for this one to get complete.
|
| waiting_for_completion_.frame_size = frame->Length();
|
| - waiting_for_completion_.latest_packet_time =
|
| - frame->LatestPacketTimeMs();
|
| + waiting_for_completion_.latest_packet_time = frame->LatestPacketTimeMs();
|
| waiting_for_completion_.timestamp = frame->TimeStamp();
|
| }
|
| }
|
| @@ -741,8 +741,8 @@ VCMFrameBufferEnum VCMJitterBuffer::InsertPacket(const VCMPacket& packet,
|
| frame->InsertPacket(packet, now_ms, decode_error_mode_, frame_data);
|
|
|
| if (previous_state != kStateComplete) {
|
| - TRACE_EVENT_ASYNC_BEGIN1("webrtc", "Video", frame->TimeStamp(),
|
| - "timestamp", frame->TimeStamp());
|
| + TRACE_EVENT_ASYNC_BEGIN1("webrtc", "Video", frame->TimeStamp(), "timestamp",
|
| + frame->TimeStamp());
|
| }
|
|
|
| if (buffer_state > 0) {
|
| @@ -759,8 +759,8 @@ VCMFrameBufferEnum VCMJitterBuffer::InsertPacket(const VCMPacket& packet,
|
| buffer_state = kFlushIndicator;
|
| }
|
|
|
| - latest_received_sequence_number_ = LatestSequenceNumber(
|
| - latest_received_sequence_number_, packet.seqNum);
|
| + latest_received_sequence_number_ =
|
| + LatestSequenceNumber(latest_received_sequence_number_, packet.seqNum);
|
| }
|
| }
|
|
|
| @@ -793,8 +793,9 @@ VCMFrameBufferEnum VCMJitterBuffer::InsertPacket(const VCMPacket& packet,
|
| } else {
|
| incomplete_frames_.InsertFrame(frame);
|
| // If NACKs are enabled, keyframes are triggered by |GetNackList|.
|
| - if (nack_mode_ == kNoNack && NonContinuousOrIncompleteDuration() >
|
| - 90 * kMaxDiscontinuousFramesTime) {
|
| + if (nack_mode_ == kNoNack &&
|
| + NonContinuousOrIncompleteDuration() >
|
| + 90 * kMaxDiscontinuousFramesTime) {
|
| return kFlushIndicator;
|
| }
|
| }
|
| @@ -808,8 +809,9 @@ VCMFrameBufferEnum VCMJitterBuffer::InsertPacket(const VCMPacket& packet,
|
| } else {
|
| incomplete_frames_.InsertFrame(frame);
|
| // If NACKs are enabled, keyframes are triggered by |GetNackList|.
|
| - if (nack_mode_ == kNoNack && NonContinuousOrIncompleteDuration() >
|
| - 90 * kMaxDiscontinuousFramesTime) {
|
| + if (nack_mode_ == kNoNack &&
|
| + NonContinuousOrIncompleteDuration() >
|
| + 90 * kMaxDiscontinuousFramesTime) {
|
| return kFlushIndicator;
|
| }
|
| }
|
| @@ -830,12 +832,14 @@ VCMFrameBufferEnum VCMJitterBuffer::InsertPacket(const VCMPacket& packet,
|
| case kFlushIndicator:
|
| free_frames_.push_back(frame);
|
| return kFlushIndicator;
|
| - default: assert(false);
|
| + default:
|
| + assert(false);
|
| }
|
| return buffer_state;
|
| }
|
|
|
| -bool VCMJitterBuffer::IsContinuousInState(const VCMFrameBuffer& frame,
|
| +bool VCMJitterBuffer::IsContinuousInState(
|
| + const VCMFrameBuffer& frame,
|
| const VCMDecodingState& decoding_state) const {
|
| // Is this frame (complete or decodable) and continuous?
|
| // kStateDecodable will never be set when decode_error_mode_ is false
|
| @@ -853,7 +857,7 @@ bool VCMJitterBuffer::IsContinuous(const VCMFrameBuffer& frame) const {
|
| VCMDecodingState decoding_state;
|
| decoding_state.CopyFrom(last_decoded_state_);
|
| for (FrameList::const_iterator it = decodable_frames_.begin();
|
| - it != decodable_frames_.end(); ++it) {
|
| + it != decodable_frames_.end(); ++it) {
|
| VCMFrameBuffer* decodable_frame = it->second;
|
| if (IsNewerTimestamp(decodable_frame->TimeStamp(), frame.TimeStamp())) {
|
| break;
|
| @@ -886,7 +890,7 @@ void VCMJitterBuffer::FindAndInsertContinuousFramesWithState(
|
| // 1. Continuous base or sync layer.
|
| // 2. The end of the list was reached.
|
| for (FrameList::iterator it = incomplete_frames_.begin();
|
| - it != incomplete_frames_.end();) {
|
| + it != incomplete_frames_.end();) {
|
| VCMFrameBuffer* frame = it->second;
|
| if (IsNewerTimestamp(original_decoded_state.time_stamp(),
|
| frame->TimeStamp())) {
|
| @@ -996,16 +1000,18 @@ std::vector<uint16_t> VCMJitterBuffer::GetNackList(bool* request_key_frame) {
|
| if (last_decoded_state_.in_initial_state()) {
|
| VCMFrameBuffer* next_frame = NextFrame();
|
| const bool first_frame_is_key = next_frame &&
|
| - next_frame->FrameType() == kVideoFrameKey &&
|
| - next_frame->HaveFirstPacket();
|
| + next_frame->FrameType() == kVideoFrameKey &&
|
| + next_frame->HaveFirstPacket();
|
| if (!first_frame_is_key) {
|
| - bool have_non_empty_frame = decodable_frames_.end() != find_if(
|
| - decodable_frames_.begin(), decodable_frames_.end(),
|
| - HasNonEmptyState);
|
| + bool have_non_empty_frame =
|
| + decodable_frames_.end() != find_if(decodable_frames_.begin(),
|
| + decodable_frames_.end(),
|
| + HasNonEmptyState);
|
| if (!have_non_empty_frame) {
|
| - have_non_empty_frame = incomplete_frames_.end() != find_if(
|
| - incomplete_frames_.begin(), incomplete_frames_.end(),
|
| - HasNonEmptyState);
|
| + have_non_empty_frame =
|
| + incomplete_frames_.end() != find_if(incomplete_frames_.begin(),
|
| + incomplete_frames_.end(),
|
| + HasNonEmptyState);
|
| }
|
| bool found_key_frame = RecycleFramesUntilKeyFrame();
|
| if (!found_key_frame) {
|
| @@ -1024,8 +1030,8 @@ std::vector<uint16_t> VCMJitterBuffer::GetNackList(bool* request_key_frame) {
|
| LOG_F(LS_WARNING) << "Too long non-decodable duration: "
|
| << non_continuous_incomplete_duration << " > "
|
| << 90 * max_incomplete_time_ms_;
|
| - FrameList::reverse_iterator rit = find_if(incomplete_frames_.rbegin(),
|
| - incomplete_frames_.rend(), IsKeyFrame);
|
| + FrameList::reverse_iterator rit = find_if(
|
| + incomplete_frames_.rbegin(), incomplete_frames_.rend(), IsKeyFrame);
|
| if (rit == incomplete_frames_.rend()) {
|
| // Request a key frame if we don't have one already.
|
| *request_key_frame = true;
|
| @@ -1065,8 +1071,7 @@ bool VCMJitterBuffer::UpdateNackList(uint16_t sequence_number) {
|
| // Make sure we don't add packets which are already too old to be decoded.
|
| if (!last_decoded_state_.in_initial_state()) {
|
| latest_received_sequence_number_ = LatestSequenceNumber(
|
| - latest_received_sequence_number_,
|
| - last_decoded_state_.sequence_num());
|
| + latest_received_sequence_number_, last_decoded_state_.sequence_num());
|
| }
|
| if (IsNewerSequenceNumber(sequence_number,
|
| latest_received_sequence_number_)) {
|
| @@ -1116,8 +1121,8 @@ bool VCMJitterBuffer::MissingTooOldPacket(
|
| if (missing_sequence_numbers_.empty()) {
|
| return false;
|
| }
|
| - const uint16_t age_of_oldest_missing_packet = latest_sequence_number -
|
| - *missing_sequence_numbers_.begin();
|
| + const uint16_t age_of_oldest_missing_packet =
|
| + latest_sequence_number - *missing_sequence_numbers_.begin();
|
| // Recycle frames if the NACK list contains too old sequence numbers as
|
| // the packets may have already been dropped by the sender.
|
| return age_of_oldest_missing_packet > max_packet_age_to_nack_;
|
| @@ -1125,8 +1130,8 @@ bool VCMJitterBuffer::MissingTooOldPacket(
|
|
|
| bool VCMJitterBuffer::HandleTooOldPackets(uint16_t latest_sequence_number) {
|
| bool key_frame_found = false;
|
| - const uint16_t age_of_oldest_missing_packet = latest_sequence_number -
|
| - *missing_sequence_numbers_.begin();
|
| + const uint16_t age_of_oldest_missing_packet =
|
| + latest_sequence_number - *missing_sequence_numbers_.begin();
|
| LOG_F(LS_WARNING) << "NACK list contains too old sequence numbers: "
|
| << age_of_oldest_missing_packet << " > "
|
| << max_packet_age_to_nack_;
|
| @@ -1140,9 +1145,9 @@ void VCMJitterBuffer::DropPacketsFromNackList(
|
| uint16_t last_decoded_sequence_number) {
|
| // Erase all sequence numbers from the NACK list which we won't need any
|
| // longer.
|
| - missing_sequence_numbers_.erase(missing_sequence_numbers_.begin(),
|
| - missing_sequence_numbers_.upper_bound(
|
| - last_decoded_sequence_number));
|
| + missing_sequence_numbers_.erase(
|
| + missing_sequence_numbers_.begin(),
|
| + missing_sequence_numbers_.upper_bound(last_decoded_sequence_number));
|
| }
|
|
|
| int64_t VCMJitterBuffer::LastDecodedTimestamp() const {
|
| @@ -1226,11 +1231,11 @@ void VCMJitterBuffer::CountFrame(const VCMFrameBuffer& frame) {
|
| incoming_frame_count_++;
|
|
|
| if (frame.FrameType() == kVideoFrameKey) {
|
| - TRACE_EVENT_ASYNC_STEP0("webrtc", "Video",
|
| - frame.TimeStamp(), "KeyComplete");
|
| + TRACE_EVENT_ASYNC_STEP0("webrtc", "Video", frame.TimeStamp(),
|
| + "KeyComplete");
|
| } else {
|
| - TRACE_EVENT_ASYNC_STEP0("webrtc", "Video",
|
| - frame.TimeStamp(), "DeltaComplete");
|
| + TRACE_EVENT_ASYNC_STEP0("webrtc", "Video", frame.TimeStamp(),
|
| + "DeltaComplete");
|
| }
|
|
|
| // Update receive statistics. We count all layers, thus when you use layers
|
| @@ -1248,13 +1253,13 @@ void VCMJitterBuffer::CountFrame(const VCMFrameBuffer& frame) {
|
|
|
| void VCMJitterBuffer::UpdateAveragePacketsPerFrame(int current_number_packets) {
|
| if (frame_counter_ > kFastConvergeThreshold) {
|
| - average_packets_per_frame_ = average_packets_per_frame_
|
| - * (1 - kNormalConvergeMultiplier)
|
| - + current_number_packets * kNormalConvergeMultiplier;
|
| + average_packets_per_frame_ =
|
| + average_packets_per_frame_ * (1 - kNormalConvergeMultiplier) +
|
| + current_number_packets * kNormalConvergeMultiplier;
|
| } else if (frame_counter_ > 0) {
|
| - average_packets_per_frame_ = average_packets_per_frame_
|
| - * (1 - kFastConvergeMultiplier)
|
| - + current_number_packets * kFastConvergeMultiplier;
|
| + average_packets_per_frame_ =
|
| + average_packets_per_frame_ * (1 - kFastConvergeMultiplier) +
|
| + current_number_packets * kFastConvergeMultiplier;
|
| frame_counter_++;
|
| } else {
|
| average_packets_per_frame_ = current_number_packets;
|
| @@ -1276,7 +1281,7 @@ void VCMJitterBuffer::CleanUpOldOrEmptyFrames() {
|
| // Must be called from within |crit_sect_|.
|
| bool VCMJitterBuffer::IsPacketRetransmitted(const VCMPacket& packet) const {
|
| return missing_sequence_numbers_.find(packet.seqNum) !=
|
| - missing_sequence_numbers_.end();
|
| + missing_sequence_numbers_.end();
|
| }
|
|
|
| // Must be called under the critical section |crit_sect_|. Should never be
|
| @@ -1308,18 +1313,16 @@ void VCMJitterBuffer::UpdateJitterEstimate(const VCMFrameBuffer& frame,
|
| // Must be called under the critical section |crit_sect_|. Should never be
|
| // called with retransmitted frames, they must be filtered out before this
|
| // function is called.
|
| -void VCMJitterBuffer::UpdateJitterEstimate(
|
| - int64_t latest_packet_time_ms,
|
| - uint32_t timestamp,
|
| - unsigned int frame_size,
|
| - bool incomplete_frame) {
|
| +void VCMJitterBuffer::UpdateJitterEstimate(int64_t latest_packet_time_ms,
|
| + uint32_t timestamp,
|
| + unsigned int frame_size,
|
| + bool incomplete_frame) {
|
| if (latest_packet_time_ms == -1) {
|
| return;
|
| }
|
| int64_t frame_delay;
|
| - bool not_reordered = inter_frame_delay_.CalculateDelay(timestamp,
|
| - &frame_delay,
|
| - latest_packet_time_ms);
|
| + bool not_reordered = inter_frame_delay_.CalculateDelay(
|
| + timestamp, &frame_delay, latest_packet_time_ms);
|
| // Filter out frames which have been reordered in time by the network
|
| if (not_reordered) {
|
| // Update the jitter estimate with the new samples
|
|
|