| Index: webrtc/modules/video_coding/main/source/jitter_buffer.cc
|
| diff --git a/webrtc/modules/video_coding/main/source/jitter_buffer.cc b/webrtc/modules/video_coding/main/source/jitter_buffer.cc
|
| deleted file mode 100644
|
| index b5314de885b2fe741fb434c2d5976ca9e2cbdd47..0000000000000000000000000000000000000000
|
| --- a/webrtc/modules/video_coding/main/source/jitter_buffer.cc
|
| +++ /dev/null
|
| @@ -1,1322 +0,0 @@
|
| -/*
|
| - * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved.
|
| - *
|
| - * Use of this source code is governed by a BSD-style license
|
| - * that can be found in the LICENSE file in the root of the source
|
| - * tree. An additional intellectual property rights grant can be found
|
| - * in the file PATENTS. All contributing project authors may
|
| - * be found in the AUTHORS file in the root of the source tree.
|
| - */
|
| -#include "webrtc/modules/video_coding/main/source/jitter_buffer.h"
|
| -
|
| -#include <assert.h>
|
| -
|
| -#include <algorithm>
|
| -#include <utility>
|
| -
|
| -#include "webrtc/base/checks.h"
|
| -#include "webrtc/base/logging.h"
|
| -#include "webrtc/base/trace_event.h"
|
| -#include "webrtc/modules/rtp_rtcp/include/rtp_rtcp_defines.h"
|
| -#include "webrtc/modules/video_coding/main/interface/video_coding.h"
|
| -#include "webrtc/modules/video_coding/main/source/frame_buffer.h"
|
| -#include "webrtc/modules/video_coding/main/source/inter_frame_delay.h"
|
| -#include "webrtc/modules/video_coding/main/source/internal_defines.h"
|
| -#include "webrtc/modules/video_coding/main/source/jitter_buffer_common.h"
|
| -#include "webrtc/modules/video_coding/main/source/jitter_estimator.h"
|
| -#include "webrtc/modules/video_coding/main/source/packet.h"
|
| -#include "webrtc/system_wrappers/include/clock.h"
|
| -#include "webrtc/system_wrappers/include/critical_section_wrapper.h"
|
| -#include "webrtc/system_wrappers/include/event_wrapper.h"
|
| -#include "webrtc/system_wrappers/include/metrics.h"
|
| -
|
| -namespace webrtc {
|
| -
|
| -// Interval for updating SS data.
|
| -static const uint32_t kSsCleanupIntervalSec = 60;
|
| -
|
| -// Use this rtt if no value has been reported.
|
| -static const int64_t kDefaultRtt = 200;
|
| -
|
| -typedef std::pair<uint32_t, VCMFrameBuffer*> FrameListPair;
|
| -
|
| -bool IsKeyFrame(FrameListPair pair) {
|
| - return pair.second->FrameType() == kVideoFrameKey;
|
| -}
|
| -
|
| -bool HasNonEmptyState(FrameListPair pair) {
|
| - return pair.second->GetState() != kStateEmpty;
|
| -}
|
| -
|
| -void FrameList::InsertFrame(VCMFrameBuffer* frame) {
|
| - insert(rbegin().base(), FrameListPair(frame->TimeStamp(), frame));
|
| -}
|
| -
|
| -VCMFrameBuffer* FrameList::PopFrame(uint32_t timestamp) {
|
| - FrameList::iterator it = find(timestamp);
|
| - if (it == end())
|
| - return NULL;
|
| - VCMFrameBuffer* frame = it->second;
|
| - erase(it);
|
| - return frame;
|
| -}
|
| -
|
| -VCMFrameBuffer* FrameList::Front() const {
|
| - return begin()->second;
|
| -}
|
| -
|
| -VCMFrameBuffer* FrameList::Back() const {
|
| - return rbegin()->second;
|
| -}
|
| -
|
| -int FrameList::RecycleFramesUntilKeyFrame(FrameList::iterator* key_frame_it,
|
| - UnorderedFrameList* free_frames) {
|
| - int drop_count = 0;
|
| - FrameList::iterator it = begin();
|
| - while (!empty()) {
|
| - // Throw at least one frame.
|
| - it->second->Reset();
|
| - free_frames->push_back(it->second);
|
| - erase(it++);
|
| - ++drop_count;
|
| - if (it != end() && it->second->FrameType() == kVideoFrameKey) {
|
| - *key_frame_it = it;
|
| - return drop_count;
|
| - }
|
| - }
|
| - *key_frame_it = end();
|
| - return drop_count;
|
| -}
|
| -
|
| -void FrameList::CleanUpOldOrEmptyFrames(VCMDecodingState* decoding_state,
|
| - UnorderedFrameList* free_frames) {
|
| - while (!empty()) {
|
| - VCMFrameBuffer* oldest_frame = Front();
|
| - bool remove_frame = false;
|
| - if (oldest_frame->GetState() == kStateEmpty && size() > 1) {
|
| - // This frame is empty, try to update the last decoded state and drop it
|
| - // if successful.
|
| - remove_frame = decoding_state->UpdateEmptyFrame(oldest_frame);
|
| - } else {
|
| - remove_frame = decoding_state->IsOldFrame(oldest_frame);
|
| - }
|
| - if (!remove_frame) {
|
| - break;
|
| - }
|
| - free_frames->push_back(oldest_frame);
|
| - TRACE_EVENT_INSTANT1("webrtc", "JB::OldOrEmptyFrameDropped", "timestamp",
|
| - oldest_frame->TimeStamp());
|
| - erase(begin());
|
| - }
|
| -}
|
| -
|
| -void FrameList::Reset(UnorderedFrameList* free_frames) {
|
| - while (!empty()) {
|
| - begin()->second->Reset();
|
| - free_frames->push_back(begin()->second);
|
| - erase(begin());
|
| - }
|
| -}
|
| -
|
| -bool Vp9SsMap::Insert(const VCMPacket& packet) {
|
| - if (!packet.codecSpecificHeader.codecHeader.VP9.ss_data_available)
|
| - return false;
|
| -
|
| - ss_map_[packet.timestamp] = packet.codecSpecificHeader.codecHeader.VP9.gof;
|
| - return true;
|
| -}
|
| -
|
| -void Vp9SsMap::Reset() {
|
| - ss_map_.clear();
|
| -}
|
| -
|
| -bool Vp9SsMap::Find(uint32_t timestamp, SsMap::iterator* it_out) {
|
| - bool found = false;
|
| - for (SsMap::iterator it = ss_map_.begin(); it != ss_map_.end(); ++it) {
|
| - if (it->first == timestamp || IsNewerTimestamp(timestamp, it->first)) {
|
| - *it_out = it;
|
| - found = true;
|
| - }
|
| - }
|
| - return found;
|
| -}
|
| -
|
| -void Vp9SsMap::RemoveOld(uint32_t timestamp) {
|
| - if (!TimeForCleanup(timestamp))
|
| - return;
|
| -
|
| - SsMap::iterator it;
|
| - if (!Find(timestamp, &it))
|
| - return;
|
| -
|
| - ss_map_.erase(ss_map_.begin(), it);
|
| - AdvanceFront(timestamp);
|
| -}
|
| -
|
| -bool Vp9SsMap::TimeForCleanup(uint32_t timestamp) const {
|
| - if (ss_map_.empty() || !IsNewerTimestamp(timestamp, ss_map_.begin()->first))
|
| - return false;
|
| -
|
| - uint32_t diff = timestamp - ss_map_.begin()->first;
|
| - return diff / kVideoPayloadTypeFrequency >= kSsCleanupIntervalSec;
|
| -}
|
| -
|
| -void Vp9SsMap::AdvanceFront(uint32_t timestamp) {
|
| - RTC_DCHECK(!ss_map_.empty());
|
| - GofInfoVP9 gof = ss_map_.begin()->second;
|
| - ss_map_.erase(ss_map_.begin());
|
| - ss_map_[timestamp] = gof;
|
| -}
|
| -
|
| -// TODO(asapersson): Update according to updates in RTP payload profile.
|
| -bool Vp9SsMap::UpdatePacket(VCMPacket* packet) {
|
| - uint8_t gof_idx = packet->codecSpecificHeader.codecHeader.VP9.gof_idx;
|
| - if (gof_idx == kNoGofIdx)
|
| - return false; // No update needed.
|
| -
|
| - SsMap::iterator it;
|
| - if (!Find(packet->timestamp, &it))
|
| - return false; // Corresponding SS not yet received.
|
| -
|
| - if (gof_idx >= it->second.num_frames_in_gof)
|
| - return false; // Assume corresponding SS not yet received.
|
| -
|
| - RTPVideoHeaderVP9* vp9 = &packet->codecSpecificHeader.codecHeader.VP9;
|
| - vp9->temporal_idx = it->second.temporal_idx[gof_idx];
|
| - vp9->temporal_up_switch = it->second.temporal_up_switch[gof_idx];
|
| -
|
| - // TODO(asapersson): Set vp9.ref_picture_id[i] and add usage.
|
| - vp9->num_ref_pics = it->second.num_ref_pics[gof_idx];
|
| - for (uint8_t i = 0; i < it->second.num_ref_pics[gof_idx]; ++i) {
|
| - vp9->pid_diff[i] = it->second.pid_diff[gof_idx][i];
|
| - }
|
| - return true;
|
| -}
|
| -
|
| -void Vp9SsMap::UpdateFrames(FrameList* frames) {
|
| - for (const auto& frame_it : *frames) {
|
| - uint8_t gof_idx =
|
| - frame_it.second->CodecSpecific()->codecSpecific.VP9.gof_idx;
|
| - if (gof_idx == kNoGofIdx) {
|
| - continue;
|
| - }
|
| - SsMap::iterator ss_it;
|
| - if (Find(frame_it.second->TimeStamp(), &ss_it)) {
|
| - if (gof_idx >= ss_it->second.num_frames_in_gof) {
|
| - continue; // Assume corresponding SS not yet received.
|
| - }
|
| - frame_it.second->SetGofInfo(ss_it->second, gof_idx);
|
| - }
|
| - }
|
| -}
|
| -
|
| -VCMJitterBuffer::VCMJitterBuffer(Clock* clock,
|
| - rtc::scoped_ptr<EventWrapper> event)
|
| - : clock_(clock),
|
| - running_(false),
|
| - crit_sect_(CriticalSectionWrapper::CreateCriticalSection()),
|
| - frame_event_(event.Pass()),
|
| - max_number_of_frames_(kStartNumberOfFrames),
|
| - free_frames_(),
|
| - decodable_frames_(),
|
| - incomplete_frames_(),
|
| - last_decoded_state_(),
|
| - first_packet_since_reset_(true),
|
| - stats_callback_(NULL),
|
| - incoming_frame_rate_(0),
|
| - incoming_frame_count_(0),
|
| - time_last_incoming_frame_count_(0),
|
| - incoming_bit_count_(0),
|
| - incoming_bit_rate_(0),
|
| - num_consecutive_old_packets_(0),
|
| - num_packets_(0),
|
| - num_duplicated_packets_(0),
|
| - num_discarded_packets_(0),
|
| - time_first_packet_ms_(0),
|
| - jitter_estimate_(clock),
|
| - inter_frame_delay_(clock_->TimeInMilliseconds()),
|
| - rtt_ms_(kDefaultRtt),
|
| - nack_mode_(kNoNack),
|
| - low_rtt_nack_threshold_ms_(-1),
|
| - high_rtt_nack_threshold_ms_(-1),
|
| - missing_sequence_numbers_(SequenceNumberLessThan()),
|
| - max_nack_list_size_(0),
|
| - max_packet_age_to_nack_(0),
|
| - max_incomplete_time_ms_(0),
|
| - decode_error_mode_(kNoErrors),
|
| - average_packets_per_frame_(0.0f),
|
| - frame_counter_(0) {
|
| - for (int i = 0; i < kStartNumberOfFrames; i++)
|
| - free_frames_.push_back(new VCMFrameBuffer());
|
| -}
|
| -
|
| -VCMJitterBuffer::~VCMJitterBuffer() {
|
| - Stop();
|
| - for (UnorderedFrameList::iterator it = free_frames_.begin();
|
| - it != free_frames_.end(); ++it) {
|
| - delete *it;
|
| - }
|
| - for (FrameList::iterator it = incomplete_frames_.begin();
|
| - it != incomplete_frames_.end(); ++it) {
|
| - delete it->second;
|
| - }
|
| - for (FrameList::iterator it = decodable_frames_.begin();
|
| - it != decodable_frames_.end(); ++it) {
|
| - delete it->second;
|
| - }
|
| - delete crit_sect_;
|
| -}
|
| -
|
| -void VCMJitterBuffer::UpdateHistograms() {
|
| - if (num_packets_ <= 0 || !running_) {
|
| - return;
|
| - }
|
| - int64_t elapsed_sec =
|
| - (clock_->TimeInMilliseconds() - time_first_packet_ms_) / 1000;
|
| - if (elapsed_sec < metrics::kMinRunTimeInSeconds) {
|
| - return;
|
| - }
|
| -
|
| - RTC_HISTOGRAM_PERCENTAGE("WebRTC.Video.DiscardedPacketsInPercent",
|
| - num_discarded_packets_ * 100 / num_packets_);
|
| - RTC_HISTOGRAM_PERCENTAGE("WebRTC.Video.DuplicatedPacketsInPercent",
|
| - num_duplicated_packets_ * 100 / num_packets_);
|
| -
|
| - int total_frames =
|
| - receive_statistics_.key_frames + receive_statistics_.delta_frames;
|
| - if (total_frames > 0) {
|
| - RTC_HISTOGRAM_COUNTS_100("WebRTC.Video.CompleteFramesReceivedPerSecond",
|
| - static_cast<int>((total_frames / elapsed_sec) + 0.5f));
|
| - RTC_HISTOGRAM_COUNTS_1000(
|
| - "WebRTC.Video.KeyFramesReceivedInPermille",
|
| - static_cast<int>(
|
| - (receive_statistics_.key_frames * 1000.0f / total_frames) + 0.5f));
|
| - }
|
| -}
|
| -
|
| -void VCMJitterBuffer::Start() {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - running_ = true;
|
| - incoming_frame_count_ = 0;
|
| - incoming_frame_rate_ = 0;
|
| - incoming_bit_count_ = 0;
|
| - incoming_bit_rate_ = 0;
|
| - time_last_incoming_frame_count_ = clock_->TimeInMilliseconds();
|
| - receive_statistics_ = FrameCounts();
|
| -
|
| - num_consecutive_old_packets_ = 0;
|
| - num_packets_ = 0;
|
| - num_duplicated_packets_ = 0;
|
| - num_discarded_packets_ = 0;
|
| - time_first_packet_ms_ = 0;
|
| -
|
| - // Start in a non-signaled state.
|
| - waiting_for_completion_.frame_size = 0;
|
| - waiting_for_completion_.timestamp = 0;
|
| - waiting_for_completion_.latest_packet_time = -1;
|
| - first_packet_since_reset_ = true;
|
| - rtt_ms_ = kDefaultRtt;
|
| - last_decoded_state_.Reset();
|
| -}
|
| -
|
| -void VCMJitterBuffer::Stop() {
|
| - crit_sect_->Enter();
|
| - UpdateHistograms();
|
| - running_ = false;
|
| - last_decoded_state_.Reset();
|
| -
|
| - // Make sure all frames are free and reset.
|
| - for (FrameList::iterator it = decodable_frames_.begin();
|
| - it != decodable_frames_.end(); ++it) {
|
| - free_frames_.push_back(it->second);
|
| - }
|
| - for (FrameList::iterator it = incomplete_frames_.begin();
|
| - it != incomplete_frames_.end(); ++it) {
|
| - free_frames_.push_back(it->second);
|
| - }
|
| - for (UnorderedFrameList::iterator it = free_frames_.begin();
|
| - it != free_frames_.end(); ++it) {
|
| - (*it)->Reset();
|
| - }
|
| - decodable_frames_.clear();
|
| - incomplete_frames_.clear();
|
| - crit_sect_->Leave();
|
| - // Make sure we wake up any threads waiting on these events.
|
| - frame_event_->Set();
|
| -}
|
| -
|
| -bool VCMJitterBuffer::Running() const {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - return running_;
|
| -}
|
| -
|
| -void VCMJitterBuffer::Flush() {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - decodable_frames_.Reset(&free_frames_);
|
| - incomplete_frames_.Reset(&free_frames_);
|
| - last_decoded_state_.Reset(); // TODO(mikhal): sync reset.
|
| - num_consecutive_old_packets_ = 0;
|
| - // Also reset the jitter and delay estimates
|
| - jitter_estimate_.Reset();
|
| - inter_frame_delay_.Reset(clock_->TimeInMilliseconds());
|
| - waiting_for_completion_.frame_size = 0;
|
| - waiting_for_completion_.timestamp = 0;
|
| - waiting_for_completion_.latest_packet_time = -1;
|
| - first_packet_since_reset_ = true;
|
| - missing_sequence_numbers_.clear();
|
| -}
|
| -
|
| -// Get received key and delta frames
|
| -FrameCounts VCMJitterBuffer::FrameStatistics() const {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - return receive_statistics_;
|
| -}
|
| -
|
| -int VCMJitterBuffer::num_packets() const {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - return num_packets_;
|
| -}
|
| -
|
| -int VCMJitterBuffer::num_duplicated_packets() const {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - return num_duplicated_packets_;
|
| -}
|
| -
|
| -int VCMJitterBuffer::num_discarded_packets() const {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - return num_discarded_packets_;
|
| -}
|
| -
|
| -// Calculate framerate and bitrate.
|
| -void VCMJitterBuffer::IncomingRateStatistics(unsigned int* framerate,
|
| - unsigned int* bitrate) {
|
| - assert(framerate);
|
| - assert(bitrate);
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - const int64_t now = clock_->TimeInMilliseconds();
|
| - int64_t diff = now - time_last_incoming_frame_count_;
|
| - if (diff < 1000 && incoming_frame_rate_ > 0 && incoming_bit_rate_ > 0) {
|
| - // Make sure we report something even though less than
|
| - // 1 second has passed since last update.
|
| - *framerate = incoming_frame_rate_;
|
| - *bitrate = incoming_bit_rate_;
|
| - } else if (incoming_frame_count_ != 0) {
|
| - // We have received frame(s) since last call to this function
|
| -
|
| - // Prepare calculations
|
| - if (diff <= 0) {
|
| - diff = 1;
|
| - }
|
| - // we add 0.5f for rounding
|
| - float rate = 0.5f + ((incoming_frame_count_ * 1000.0f) / diff);
|
| - if (rate < 1.0f) {
|
| - rate = 1.0f;
|
| - }
|
| -
|
| - // Calculate frame rate
|
| - // Let r be rate.
|
| - // r(0) = 1000*framecount/delta_time.
|
| - // (I.e. frames per second since last calculation.)
|
| - // frame_rate = r(0)/2 + r(-1)/2
|
| - // (I.e. fr/s average this and the previous calculation.)
|
| - *framerate = (incoming_frame_rate_ + static_cast<unsigned int>(rate)) / 2;
|
| - incoming_frame_rate_ = static_cast<unsigned int>(rate);
|
| -
|
| - // Calculate bit rate
|
| - if (incoming_bit_count_ == 0) {
|
| - *bitrate = 0;
|
| - } else {
|
| - *bitrate = 10 * ((100 * incoming_bit_count_) /
|
| - static_cast<unsigned int>(diff));
|
| - }
|
| - incoming_bit_rate_ = *bitrate;
|
| -
|
| - // Reset count
|
| - incoming_frame_count_ = 0;
|
| - incoming_bit_count_ = 0;
|
| - time_last_incoming_frame_count_ = now;
|
| -
|
| - } else {
|
| - // No frames since last call
|
| - time_last_incoming_frame_count_ = clock_->TimeInMilliseconds();
|
| - *framerate = 0;
|
| - *bitrate = 0;
|
| - incoming_frame_rate_ = 0;
|
| - incoming_bit_rate_ = 0;
|
| - }
|
| -}
|
| -
|
| -// Answers the question:
|
| -// Will the packet sequence be complete if the next frame is grabbed for
|
| -// decoding right now? That is, have we lost a frame between the last decoded
|
| -// frame and the next, or is the next
|
| -// frame missing one or more packets?
|
| -bool VCMJitterBuffer::CompleteSequenceWithNextFrame() {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - // Finding oldest frame ready for decoder, check sequence number and size
|
| - CleanUpOldOrEmptyFrames();
|
| - if (!decodable_frames_.empty()) {
|
| - if (decodable_frames_.Front()->GetState() == kStateComplete) {
|
| - return true;
|
| - }
|
| - } else if (incomplete_frames_.size() <= 1) {
|
| - // Frame not ready to be decoded.
|
| - return true;
|
| - }
|
| - return false;
|
| -}
|
| -
|
| -// Returns immediately or a |max_wait_time_ms| ms event hang waiting for a
|
| -// complete frame, |max_wait_time_ms| decided by caller.
|
| -bool VCMJitterBuffer::NextCompleteTimestamp(
|
| - uint32_t max_wait_time_ms, uint32_t* timestamp) {
|
| - crit_sect_->Enter();
|
| - if (!running_) {
|
| - crit_sect_->Leave();
|
| - return false;
|
| - }
|
| - CleanUpOldOrEmptyFrames();
|
| -
|
| - if (decodable_frames_.empty() ||
|
| - decodable_frames_.Front()->GetState() != kStateComplete) {
|
| - const int64_t end_wait_time_ms = clock_->TimeInMilliseconds() +
|
| - max_wait_time_ms;
|
| - int64_t wait_time_ms = max_wait_time_ms;
|
| - while (wait_time_ms > 0) {
|
| - crit_sect_->Leave();
|
| - const EventTypeWrapper ret =
|
| - frame_event_->Wait(static_cast<uint32_t>(wait_time_ms));
|
| - crit_sect_->Enter();
|
| - if (ret == kEventSignaled) {
|
| - // Are we shutting down the jitter buffer?
|
| - if (!running_) {
|
| - crit_sect_->Leave();
|
| - return false;
|
| - }
|
| - // Finding oldest frame ready for decoder.
|
| - CleanUpOldOrEmptyFrames();
|
| - if (decodable_frames_.empty() ||
|
| - decodable_frames_.Front()->GetState() != kStateComplete) {
|
| - wait_time_ms = end_wait_time_ms - clock_->TimeInMilliseconds();
|
| - } else {
|
| - break;
|
| - }
|
| - } else {
|
| - break;
|
| - }
|
| - }
|
| - }
|
| - if (decodable_frames_.empty() ||
|
| - decodable_frames_.Front()->GetState() != kStateComplete) {
|
| - crit_sect_->Leave();
|
| - return false;
|
| - }
|
| - *timestamp = decodable_frames_.Front()->TimeStamp();
|
| - crit_sect_->Leave();
|
| - return true;
|
| -}
|
| -
|
| -bool VCMJitterBuffer::NextMaybeIncompleteTimestamp(uint32_t* timestamp) {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - if (!running_) {
|
| - return false;
|
| - }
|
| - if (decode_error_mode_ == kNoErrors) {
|
| - // No point to continue, as we are not decoding with errors.
|
| - return false;
|
| - }
|
| -
|
| - CleanUpOldOrEmptyFrames();
|
| -
|
| - if (decodable_frames_.empty()) {
|
| - return false;
|
| - }
|
| - VCMFrameBuffer* oldest_frame = decodable_frames_.Front();
|
| - // If we have exactly one frame in the buffer, release it only if it is
|
| - // complete. We know decodable_frames_ is not empty due to the previous
|
| - // check.
|
| - if (decodable_frames_.size() == 1 && incomplete_frames_.empty()
|
| - && oldest_frame->GetState() != kStateComplete) {
|
| - return false;
|
| - }
|
| -
|
| - *timestamp = oldest_frame->TimeStamp();
|
| - return true;
|
| -}
|
| -
|
| -VCMEncodedFrame* VCMJitterBuffer::ExtractAndSetDecode(uint32_t timestamp) {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - if (!running_) {
|
| - return NULL;
|
| - }
|
| - // Extract the frame with the desired timestamp.
|
| - VCMFrameBuffer* frame = decodable_frames_.PopFrame(timestamp);
|
| - bool continuous = true;
|
| - if (!frame) {
|
| - frame = incomplete_frames_.PopFrame(timestamp);
|
| - if (frame)
|
| - continuous = last_decoded_state_.ContinuousFrame(frame);
|
| - else
|
| - return NULL;
|
| - }
|
| - TRACE_EVENT_ASYNC_STEP0("webrtc", "Video", timestamp, "Extract");
|
| - // Frame pulled out from jitter buffer, update the jitter estimate.
|
| - const bool retransmitted = (frame->GetNackCount() > 0);
|
| - if (retransmitted) {
|
| - jitter_estimate_.FrameNacked();
|
| - } else if (frame->Length() > 0) {
|
| - // Ignore retransmitted and empty frames.
|
| - if (waiting_for_completion_.latest_packet_time >= 0) {
|
| - UpdateJitterEstimate(waiting_for_completion_, true);
|
| - }
|
| - if (frame->GetState() == kStateComplete) {
|
| - UpdateJitterEstimate(*frame, false);
|
| - } else {
|
| - // Wait for this one to get complete.
|
| - waiting_for_completion_.frame_size = frame->Length();
|
| - waiting_for_completion_.latest_packet_time =
|
| - frame->LatestPacketTimeMs();
|
| - waiting_for_completion_.timestamp = frame->TimeStamp();
|
| - }
|
| - }
|
| -
|
| - // The state must be changed to decoding before cleaning up zero sized
|
| - // frames to avoid empty frames being cleaned up and then given to the
|
| - // decoder. Propagates the missing_frame bit.
|
| - frame->PrepareForDecode(continuous);
|
| -
|
| - // We have a frame - update the last decoded state and nack list.
|
| - last_decoded_state_.SetState(frame);
|
| - DropPacketsFromNackList(last_decoded_state_.sequence_num());
|
| -
|
| - if ((*frame).IsSessionComplete())
|
| - UpdateAveragePacketsPerFrame(frame->NumPackets());
|
| -
|
| - return frame;
|
| -}
|
| -
|
| -// Release frame when done with decoding. Should never be used to release
|
| -// frames from within the jitter buffer.
|
| -void VCMJitterBuffer::ReleaseFrame(VCMEncodedFrame* frame) {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - VCMFrameBuffer* frame_buffer = static_cast<VCMFrameBuffer*>(frame);
|
| - if (frame_buffer) {
|
| - free_frames_.push_back(frame_buffer);
|
| - }
|
| -}
|
| -
|
| -// Gets frame to use for this timestamp. If no match, get empty frame.
|
| -VCMFrameBufferEnum VCMJitterBuffer::GetFrame(const VCMPacket& packet,
|
| - VCMFrameBuffer** frame,
|
| - FrameList** frame_list) {
|
| - *frame = incomplete_frames_.PopFrame(packet.timestamp);
|
| - if (*frame != NULL) {
|
| - *frame_list = &incomplete_frames_;
|
| - return kNoError;
|
| - }
|
| - *frame = decodable_frames_.PopFrame(packet.timestamp);
|
| - if (*frame != NULL) {
|
| - *frame_list = &decodable_frames_;
|
| - return kNoError;
|
| - }
|
| -
|
| - *frame_list = NULL;
|
| - // No match, return empty frame.
|
| - *frame = GetEmptyFrame();
|
| - if (*frame == NULL) {
|
| - // No free frame! Try to reclaim some...
|
| - LOG(LS_WARNING) << "Unable to get empty frame; Recycling.";
|
| - bool found_key_frame = RecycleFramesUntilKeyFrame();
|
| - *frame = GetEmptyFrame();
|
| - assert(*frame);
|
| - if (!found_key_frame) {
|
| - free_frames_.push_back(*frame);
|
| - return kFlushIndicator;
|
| - }
|
| - }
|
| - (*frame)->Reset();
|
| - return kNoError;
|
| -}
|
| -
|
| -int64_t VCMJitterBuffer::LastPacketTime(const VCMEncodedFrame* frame,
|
| - bool* retransmitted) const {
|
| - assert(retransmitted);
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - const VCMFrameBuffer* frame_buffer =
|
| - static_cast<const VCMFrameBuffer*>(frame);
|
| - *retransmitted = (frame_buffer->GetNackCount() > 0);
|
| - return frame_buffer->LatestPacketTimeMs();
|
| -}
|
| -
|
| -VCMFrameBufferEnum VCMJitterBuffer::InsertPacket(const VCMPacket& packet,
|
| - bool* retransmitted) {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| -
|
| - ++num_packets_;
|
| - if (num_packets_ == 1) {
|
| - time_first_packet_ms_ = clock_->TimeInMilliseconds();
|
| - }
|
| - // Does this packet belong to an old frame?
|
| - if (last_decoded_state_.IsOldPacket(&packet)) {
|
| - // Account only for media packets.
|
| - if (packet.sizeBytes > 0) {
|
| - num_discarded_packets_++;
|
| - num_consecutive_old_packets_++;
|
| - if (stats_callback_ != NULL)
|
| - stats_callback_->OnDiscardedPacketsUpdated(num_discarded_packets_);
|
| - }
|
| - // Update last decoded sequence number if the packet arrived late and
|
| - // belongs to a frame with a timestamp equal to the last decoded
|
| - // timestamp.
|
| - last_decoded_state_.UpdateOldPacket(&packet);
|
| - DropPacketsFromNackList(last_decoded_state_.sequence_num());
|
| -
|
| - // Also see if this old packet made more incomplete frames continuous.
|
| - FindAndInsertContinuousFramesWithState(last_decoded_state_);
|
| -
|
| - if (num_consecutive_old_packets_ > kMaxConsecutiveOldPackets) {
|
| - LOG(LS_WARNING)
|
| - << num_consecutive_old_packets_
|
| - << " consecutive old packets received. Flushing the jitter buffer.";
|
| - Flush();
|
| - return kFlushIndicator;
|
| - }
|
| - return kOldPacket;
|
| - }
|
| -
|
| - num_consecutive_old_packets_ = 0;
|
| -
|
| - VCMFrameBuffer* frame;
|
| - FrameList* frame_list;
|
| - const VCMFrameBufferEnum error = GetFrame(packet, &frame, &frame_list);
|
| - if (error != kNoError)
|
| - return error;
|
| -
|
| - int64_t now_ms = clock_->TimeInMilliseconds();
|
| - // We are keeping track of the first and latest seq numbers, and
|
| - // the number of wraps to be able to calculate how many packets we expect.
|
| - if (first_packet_since_reset_) {
|
| - // Now it's time to start estimating jitter
|
| - // reset the delay estimate.
|
| - inter_frame_delay_.Reset(now_ms);
|
| - }
|
| -
|
| - // Empty packets may bias the jitter estimate (lacking size component),
|
| - // therefore don't let empty packet trigger the following updates:
|
| - if (packet.frameType != kEmptyFrame) {
|
| - if (waiting_for_completion_.timestamp == packet.timestamp) {
|
| - // This can get bad if we have a lot of duplicate packets,
|
| - // we will then count some packet multiple times.
|
| - waiting_for_completion_.frame_size += packet.sizeBytes;
|
| - waiting_for_completion_.latest_packet_time = now_ms;
|
| - } else if (waiting_for_completion_.latest_packet_time >= 0 &&
|
| - waiting_for_completion_.latest_packet_time + 2000 <= now_ms) {
|
| - // A packet should never be more than two seconds late
|
| - UpdateJitterEstimate(waiting_for_completion_, true);
|
| - waiting_for_completion_.latest_packet_time = -1;
|
| - waiting_for_completion_.frame_size = 0;
|
| - waiting_for_completion_.timestamp = 0;
|
| - }
|
| - }
|
| -
|
| - VCMFrameBufferStateEnum previous_state = frame->GetState();
|
| - // Insert packet.
|
| - FrameData frame_data;
|
| - frame_data.rtt_ms = rtt_ms_;
|
| - frame_data.rolling_average_packets_per_frame = average_packets_per_frame_;
|
| - VCMFrameBufferEnum buffer_state =
|
| - frame->InsertPacket(packet, now_ms, decode_error_mode_, frame_data);
|
| -
|
| - if (previous_state != kStateComplete) {
|
| - TRACE_EVENT_ASYNC_BEGIN1("webrtc", "Video", frame->TimeStamp(),
|
| - "timestamp", frame->TimeStamp());
|
| - }
|
| -
|
| - if (buffer_state > 0) {
|
| - incoming_bit_count_ += packet.sizeBytes << 3;
|
| - if (first_packet_since_reset_) {
|
| - latest_received_sequence_number_ = packet.seqNum;
|
| - first_packet_since_reset_ = false;
|
| - } else {
|
| - if (IsPacketRetransmitted(packet)) {
|
| - frame->IncrementNackCount();
|
| - }
|
| - if (!UpdateNackList(packet.seqNum) &&
|
| - packet.frameType != kVideoFrameKey) {
|
| - buffer_state = kFlushIndicator;
|
| - }
|
| -
|
| - latest_received_sequence_number_ = LatestSequenceNumber(
|
| - latest_received_sequence_number_, packet.seqNum);
|
| - }
|
| - }
|
| -
|
| - // Is the frame already in the decodable list?
|
| - bool continuous = IsContinuous(*frame);
|
| - switch (buffer_state) {
|
| - case kGeneralError:
|
| - case kTimeStampError:
|
| - case kSizeError: {
|
| - free_frames_.push_back(frame);
|
| - break;
|
| - }
|
| - case kCompleteSession: {
|
| - if (previous_state != kStateDecodable &&
|
| - previous_state != kStateComplete) {
|
| - CountFrame(*frame);
|
| - if (continuous) {
|
| - // Signal that we have a complete session.
|
| - frame_event_->Set();
|
| - }
|
| - }
|
| - FALLTHROUGH();
|
| - }
|
| - // Note: There is no break here - continuing to kDecodableSession.
|
| - case kDecodableSession: {
|
| - *retransmitted = (frame->GetNackCount() > 0);
|
| - if (continuous) {
|
| - decodable_frames_.InsertFrame(frame);
|
| - FindAndInsertContinuousFrames(*frame);
|
| - } else {
|
| - incomplete_frames_.InsertFrame(frame);
|
| - }
|
| - break;
|
| - }
|
| - case kIncomplete: {
|
| - if (frame->GetState() == kStateEmpty &&
|
| - last_decoded_state_.UpdateEmptyFrame(frame)) {
|
| - free_frames_.push_back(frame);
|
| - return kNoError;
|
| - } else {
|
| - incomplete_frames_.InsertFrame(frame);
|
| - }
|
| - break;
|
| - }
|
| - case kNoError:
|
| - case kOutOfBoundsPacket:
|
| - case kDuplicatePacket: {
|
| - // Put back the frame where it came from.
|
| - if (frame_list != NULL) {
|
| - frame_list->InsertFrame(frame);
|
| - } else {
|
| - free_frames_.push_back(frame);
|
| - }
|
| - ++num_duplicated_packets_;
|
| - break;
|
| - }
|
| - case kFlushIndicator:
|
| - free_frames_.push_back(frame);
|
| - return kFlushIndicator;
|
| - default: assert(false);
|
| - }
|
| - return buffer_state;
|
| -}
|
| -
|
| -bool VCMJitterBuffer::IsContinuousInState(const VCMFrameBuffer& frame,
|
| - const VCMDecodingState& decoding_state) const {
|
| - if (decode_error_mode_ == kWithErrors)
|
| - return true;
|
| - // Is this frame (complete or decodable) and continuous?
|
| - // kStateDecodable will never be set when decode_error_mode_ is false
|
| - // as SessionInfo determines this state based on the error mode (and frame
|
| - // completeness).
|
| - return (frame.GetState() == kStateComplete ||
|
| - frame.GetState() == kStateDecodable) &&
|
| - decoding_state.ContinuousFrame(&frame);
|
| -}
|
| -
|
| -bool VCMJitterBuffer::IsContinuous(const VCMFrameBuffer& frame) const {
|
| - if (IsContinuousInState(frame, last_decoded_state_)) {
|
| - return true;
|
| - }
|
| - VCMDecodingState decoding_state;
|
| - decoding_state.CopyFrom(last_decoded_state_);
|
| - for (FrameList::const_iterator it = decodable_frames_.begin();
|
| - it != decodable_frames_.end(); ++it) {
|
| - VCMFrameBuffer* decodable_frame = it->second;
|
| - if (IsNewerTimestamp(decodable_frame->TimeStamp(), frame.TimeStamp())) {
|
| - break;
|
| - }
|
| - decoding_state.SetState(decodable_frame);
|
| - if (IsContinuousInState(frame, decoding_state)) {
|
| - return true;
|
| - }
|
| - }
|
| - return false;
|
| -}
|
| -
|
| -void VCMJitterBuffer::FindAndInsertContinuousFrames(
|
| - const VCMFrameBuffer& new_frame) {
|
| - VCMDecodingState decoding_state;
|
| - decoding_state.CopyFrom(last_decoded_state_);
|
| - decoding_state.SetState(&new_frame);
|
| - FindAndInsertContinuousFramesWithState(decoding_state);
|
| -}
|
| -
|
| -void VCMJitterBuffer::FindAndInsertContinuousFramesWithState(
|
| - const VCMDecodingState& original_decoded_state) {
|
| - // Copy original_decoded_state so we can move the state forward with each
|
| - // decodable frame we find.
|
| - VCMDecodingState decoding_state;
|
| - decoding_state.CopyFrom(original_decoded_state);
|
| -
|
| - // When temporal layers are available, we search for a complete or decodable
|
| - // frame until we hit one of the following:
|
| - // 1. Continuous base or sync layer.
|
| - // 2. The end of the list was reached.
|
| - for (FrameList::iterator it = incomplete_frames_.begin();
|
| - it != incomplete_frames_.end();) {
|
| - VCMFrameBuffer* frame = it->second;
|
| - if (IsNewerTimestamp(original_decoded_state.time_stamp(),
|
| - frame->TimeStamp())) {
|
| - ++it;
|
| - continue;
|
| - }
|
| - if (IsContinuousInState(*frame, decoding_state)) {
|
| - decodable_frames_.InsertFrame(frame);
|
| - incomplete_frames_.erase(it++);
|
| - decoding_state.SetState(frame);
|
| - } else if (frame->TemporalId() <= 0) {
|
| - break;
|
| - } else {
|
| - ++it;
|
| - }
|
| - }
|
| -}
|
| -
|
| -uint32_t VCMJitterBuffer::EstimatedJitterMs() {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - // Compute RTT multiplier for estimation.
|
| - // low_rtt_nackThresholdMs_ == -1 means no FEC.
|
| - double rtt_mult = 1.0f;
|
| - if (low_rtt_nack_threshold_ms_ >= 0 &&
|
| - rtt_ms_ >= low_rtt_nack_threshold_ms_) {
|
| - // For RTTs above low_rtt_nack_threshold_ms_ we don't apply extra delay
|
| - // when waiting for retransmissions.
|
| - rtt_mult = 0.0f;
|
| - }
|
| - return jitter_estimate_.GetJitterEstimate(rtt_mult);
|
| -}
|
| -
|
| -void VCMJitterBuffer::UpdateRtt(int64_t rtt_ms) {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - rtt_ms_ = rtt_ms;
|
| - jitter_estimate_.UpdateRtt(rtt_ms);
|
| -}
|
| -
|
| -void VCMJitterBuffer::SetNackMode(VCMNackMode mode,
|
| - int64_t low_rtt_nack_threshold_ms,
|
| - int64_t high_rtt_nack_threshold_ms) {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - nack_mode_ = mode;
|
| - if (mode == kNoNack) {
|
| - missing_sequence_numbers_.clear();
|
| - }
|
| - assert(low_rtt_nack_threshold_ms >= -1 && high_rtt_nack_threshold_ms >= -1);
|
| - assert(high_rtt_nack_threshold_ms == -1 ||
|
| - low_rtt_nack_threshold_ms <= high_rtt_nack_threshold_ms);
|
| - assert(low_rtt_nack_threshold_ms > -1 || high_rtt_nack_threshold_ms == -1);
|
| - low_rtt_nack_threshold_ms_ = low_rtt_nack_threshold_ms;
|
| - high_rtt_nack_threshold_ms_ = high_rtt_nack_threshold_ms;
|
| - // Don't set a high start rtt if high_rtt_nack_threshold_ms_ is used, to not
|
| - // disable NACK in |kNack| mode.
|
| - if (rtt_ms_ == kDefaultRtt && high_rtt_nack_threshold_ms_ != -1) {
|
| - rtt_ms_ = 0;
|
| - }
|
| - if (!WaitForRetransmissions()) {
|
| - jitter_estimate_.ResetNackCount();
|
| - }
|
| -}
|
| -
|
| -void VCMJitterBuffer::SetNackSettings(size_t max_nack_list_size,
|
| - int max_packet_age_to_nack,
|
| - int max_incomplete_time_ms) {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - assert(max_packet_age_to_nack >= 0);
|
| - assert(max_incomplete_time_ms_ >= 0);
|
| - max_nack_list_size_ = max_nack_list_size;
|
| - max_packet_age_to_nack_ = max_packet_age_to_nack;
|
| - max_incomplete_time_ms_ = max_incomplete_time_ms;
|
| -}
|
| -
|
| -VCMNackMode VCMJitterBuffer::nack_mode() const {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - return nack_mode_;
|
| -}
|
| -
|
| -int VCMJitterBuffer::NonContinuousOrIncompleteDuration() {
|
| - if (incomplete_frames_.empty()) {
|
| - return 0;
|
| - }
|
| - uint32_t start_timestamp = incomplete_frames_.Front()->TimeStamp();
|
| - if (!decodable_frames_.empty()) {
|
| - start_timestamp = decodable_frames_.Back()->TimeStamp();
|
| - }
|
| - return incomplete_frames_.Back()->TimeStamp() - start_timestamp;
|
| -}
|
| -
|
| -uint16_t VCMJitterBuffer::EstimatedLowSequenceNumber(
|
| - const VCMFrameBuffer& frame) const {
|
| - assert(frame.GetLowSeqNum() >= 0);
|
| - if (frame.HaveFirstPacket())
|
| - return frame.GetLowSeqNum();
|
| -
|
| - // This estimate is not accurate if more than one packet with lower sequence
|
| - // number is lost.
|
| - return frame.GetLowSeqNum() - 1;
|
| -}
|
| -
|
| -std::vector<uint16_t> VCMJitterBuffer::GetNackList(bool* request_key_frame) {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - *request_key_frame = false;
|
| - if (nack_mode_ == kNoNack) {
|
| - return std::vector<uint16_t>();
|
| - }
|
| - if (last_decoded_state_.in_initial_state()) {
|
| - VCMFrameBuffer* next_frame = NextFrame();
|
| - const bool first_frame_is_key = next_frame &&
|
| - next_frame->FrameType() == kVideoFrameKey &&
|
| - next_frame->HaveFirstPacket();
|
| - if (!first_frame_is_key) {
|
| - bool have_non_empty_frame = decodable_frames_.end() != find_if(
|
| - decodable_frames_.begin(), decodable_frames_.end(),
|
| - HasNonEmptyState);
|
| - if (!have_non_empty_frame) {
|
| - have_non_empty_frame = incomplete_frames_.end() != find_if(
|
| - incomplete_frames_.begin(), incomplete_frames_.end(),
|
| - HasNonEmptyState);
|
| - }
|
| - bool found_key_frame = RecycleFramesUntilKeyFrame();
|
| - if (!found_key_frame) {
|
| - *request_key_frame = have_non_empty_frame;
|
| - return std::vector<uint16_t>();
|
| - }
|
| - }
|
| - }
|
| - if (TooLargeNackList()) {
|
| - *request_key_frame = !HandleTooLargeNackList();
|
| - }
|
| - if (max_incomplete_time_ms_ > 0) {
|
| - int non_continuous_incomplete_duration =
|
| - NonContinuousOrIncompleteDuration();
|
| - if (non_continuous_incomplete_duration > 90 * max_incomplete_time_ms_) {
|
| - LOG_F(LS_WARNING) << "Too long non-decodable duration: "
|
| - << non_continuous_incomplete_duration << " > "
|
| - << 90 * max_incomplete_time_ms_;
|
| - FrameList::reverse_iterator rit = find_if(incomplete_frames_.rbegin(),
|
| - incomplete_frames_.rend(), IsKeyFrame);
|
| - if (rit == incomplete_frames_.rend()) {
|
| - // Request a key frame if we don't have one already.
|
| - *request_key_frame = true;
|
| - return std::vector<uint16_t>();
|
| - } else {
|
| - // Skip to the last key frame. If it's incomplete we will start
|
| - // NACKing it.
|
| - // Note that the estimated low sequence number is correct for VP8
|
| - // streams because only the first packet of a key frame is marked.
|
| - last_decoded_state_.Reset();
|
| - DropPacketsFromNackList(EstimatedLowSequenceNumber(*rit->second));
|
| - }
|
| - }
|
| - }
|
| - std::vector<uint16_t> nack_list(missing_sequence_numbers_.begin(),
|
| - missing_sequence_numbers_.end());
|
| - return nack_list;
|
| -}
|
| -
|
| -void VCMJitterBuffer::SetDecodeErrorMode(VCMDecodeErrorMode error_mode) {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - decode_error_mode_ = error_mode;
|
| -}
|
| -
|
| -VCMFrameBuffer* VCMJitterBuffer::NextFrame() const {
|
| - if (!decodable_frames_.empty())
|
| - return decodable_frames_.Front();
|
| - if (!incomplete_frames_.empty())
|
| - return incomplete_frames_.Front();
|
| - return NULL;
|
| -}
|
| -
|
| -bool VCMJitterBuffer::UpdateNackList(uint16_t sequence_number) {
|
| - if (nack_mode_ == kNoNack) {
|
| - return true;
|
| - }
|
| - // Make sure we don't add packets which are already too old to be decoded.
|
| - if (!last_decoded_state_.in_initial_state()) {
|
| - latest_received_sequence_number_ = LatestSequenceNumber(
|
| - latest_received_sequence_number_,
|
| - last_decoded_state_.sequence_num());
|
| - }
|
| - if (IsNewerSequenceNumber(sequence_number,
|
| - latest_received_sequence_number_)) {
|
| - // Push any missing sequence numbers to the NACK list.
|
| - for (uint16_t i = latest_received_sequence_number_ + 1;
|
| - IsNewerSequenceNumber(sequence_number, i); ++i) {
|
| - missing_sequence_numbers_.insert(missing_sequence_numbers_.end(), i);
|
| - TRACE_EVENT_INSTANT1(TRACE_DISABLED_BY_DEFAULT("webrtc_rtp"), "AddNack",
|
| - "seqnum", i);
|
| - }
|
| - if (TooLargeNackList() && !HandleTooLargeNackList()) {
|
| - LOG(LS_WARNING) << "Requesting key frame due to too large NACK list.";
|
| - return false;
|
| - }
|
| - if (MissingTooOldPacket(sequence_number) &&
|
| - !HandleTooOldPackets(sequence_number)) {
|
| - LOG(LS_WARNING) << "Requesting key frame due to missing too old packets";
|
| - return false;
|
| - }
|
| - } else {
|
| - missing_sequence_numbers_.erase(sequence_number);
|
| - TRACE_EVENT_INSTANT1(TRACE_DISABLED_BY_DEFAULT("webrtc_rtp"), "RemoveNack",
|
| - "seqnum", sequence_number);
|
| - }
|
| - return true;
|
| -}
|
| -
|
| -bool VCMJitterBuffer::TooLargeNackList() const {
|
| - return missing_sequence_numbers_.size() > max_nack_list_size_;
|
| -}
|
| -
|
| -bool VCMJitterBuffer::HandleTooLargeNackList() {
|
| - // Recycle frames until the NACK list is small enough. It is likely cheaper to
|
| - // request a key frame than to retransmit this many missing packets.
|
| - LOG_F(LS_WARNING) << "NACK list has grown too large: "
|
| - << missing_sequence_numbers_.size() << " > "
|
| - << max_nack_list_size_;
|
| - bool key_frame_found = false;
|
| - while (TooLargeNackList()) {
|
| - key_frame_found = RecycleFramesUntilKeyFrame();
|
| - }
|
| - return key_frame_found;
|
| -}
|
| -
|
| -bool VCMJitterBuffer::MissingTooOldPacket(
|
| - uint16_t latest_sequence_number) const {
|
| - if (missing_sequence_numbers_.empty()) {
|
| - return false;
|
| - }
|
| - const uint16_t age_of_oldest_missing_packet = latest_sequence_number -
|
| - *missing_sequence_numbers_.begin();
|
| - // Recycle frames if the NACK list contains too old sequence numbers as
|
| - // the packets may have already been dropped by the sender.
|
| - return age_of_oldest_missing_packet > max_packet_age_to_nack_;
|
| -}
|
| -
|
| -bool VCMJitterBuffer::HandleTooOldPackets(uint16_t latest_sequence_number) {
|
| - bool key_frame_found = false;
|
| - const uint16_t age_of_oldest_missing_packet = latest_sequence_number -
|
| - *missing_sequence_numbers_.begin();
|
| - LOG_F(LS_WARNING) << "NACK list contains too old sequence numbers: "
|
| - << age_of_oldest_missing_packet << " > "
|
| - << max_packet_age_to_nack_;
|
| - while (MissingTooOldPacket(latest_sequence_number)) {
|
| - key_frame_found = RecycleFramesUntilKeyFrame();
|
| - }
|
| - return key_frame_found;
|
| -}
|
| -
|
| -void VCMJitterBuffer::DropPacketsFromNackList(
|
| - uint16_t last_decoded_sequence_number) {
|
| - // Erase all sequence numbers from the NACK list which we won't need any
|
| - // longer.
|
| - missing_sequence_numbers_.erase(missing_sequence_numbers_.begin(),
|
| - missing_sequence_numbers_.upper_bound(
|
| - last_decoded_sequence_number));
|
| -}
|
| -
|
| -int64_t VCMJitterBuffer::LastDecodedTimestamp() const {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - return last_decoded_state_.time_stamp();
|
| -}
|
| -
|
| -void VCMJitterBuffer::RenderBufferSize(uint32_t* timestamp_start,
|
| - uint32_t* timestamp_end) {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - CleanUpOldOrEmptyFrames();
|
| - *timestamp_start = 0;
|
| - *timestamp_end = 0;
|
| - if (decodable_frames_.empty()) {
|
| - return;
|
| - }
|
| - *timestamp_start = decodable_frames_.Front()->TimeStamp();
|
| - *timestamp_end = decodable_frames_.Back()->TimeStamp();
|
| -}
|
| -
|
| -void VCMJitterBuffer::RegisterStatsCallback(
|
| - VCMReceiveStatisticsCallback* callback) {
|
| - CriticalSectionScoped cs(crit_sect_);
|
| - stats_callback_ = callback;
|
| -}
|
| -
|
| -VCMFrameBuffer* VCMJitterBuffer::GetEmptyFrame() {
|
| - if (free_frames_.empty()) {
|
| - if (!TryToIncreaseJitterBufferSize()) {
|
| - return NULL;
|
| - }
|
| - }
|
| - VCMFrameBuffer* frame = free_frames_.front();
|
| - free_frames_.pop_front();
|
| - return frame;
|
| -}
|
| -
|
| -bool VCMJitterBuffer::TryToIncreaseJitterBufferSize() {
|
| - if (max_number_of_frames_ >= kMaxNumberOfFrames)
|
| - return false;
|
| - free_frames_.push_back(new VCMFrameBuffer());
|
| - ++max_number_of_frames_;
|
| - TRACE_COUNTER1("webrtc", "JBMaxFrames", max_number_of_frames_);
|
| - return true;
|
| -}
|
| -
|
| -// Recycle oldest frames up to a key frame, used if jitter buffer is completely
|
| -// full.
|
| -bool VCMJitterBuffer::RecycleFramesUntilKeyFrame() {
|
| - // First release incomplete frames, and only release decodable frames if there
|
| - // are no incomplete ones.
|
| - FrameList::iterator key_frame_it;
|
| - bool key_frame_found = false;
|
| - int dropped_frames = 0;
|
| - dropped_frames += incomplete_frames_.RecycleFramesUntilKeyFrame(
|
| - &key_frame_it, &free_frames_);
|
| - key_frame_found = key_frame_it != incomplete_frames_.end();
|
| - if (dropped_frames == 0) {
|
| - dropped_frames += decodable_frames_.RecycleFramesUntilKeyFrame(
|
| - &key_frame_it, &free_frames_);
|
| - key_frame_found = key_frame_it != decodable_frames_.end();
|
| - }
|
| - TRACE_EVENT_INSTANT0("webrtc", "JB::RecycleFramesUntilKeyFrame");
|
| - if (key_frame_found) {
|
| - LOG(LS_INFO) << "Found key frame while dropping frames.";
|
| - // Reset last decoded state to make sure the next frame decoded is a key
|
| - // frame, and start NACKing from here.
|
| - last_decoded_state_.Reset();
|
| - DropPacketsFromNackList(EstimatedLowSequenceNumber(*key_frame_it->second));
|
| - } else if (decodable_frames_.empty()) {
|
| - // All frames dropped. Reset the decoding state and clear missing sequence
|
| - // numbers as we're starting fresh.
|
| - last_decoded_state_.Reset();
|
| - missing_sequence_numbers_.clear();
|
| - }
|
| - return key_frame_found;
|
| -}
|
| -
|
| -// Must be called under the critical section |crit_sect_|.
|
| -void VCMJitterBuffer::CountFrame(const VCMFrameBuffer& frame) {
|
| - incoming_frame_count_++;
|
| -
|
| - if (frame.FrameType() == kVideoFrameKey) {
|
| - TRACE_EVENT_ASYNC_STEP0("webrtc", "Video",
|
| - frame.TimeStamp(), "KeyComplete");
|
| - } else {
|
| - TRACE_EVENT_ASYNC_STEP0("webrtc", "Video",
|
| - frame.TimeStamp(), "DeltaComplete");
|
| - }
|
| -
|
| - // Update receive statistics. We count all layers, thus when you use layers
|
| - // adding all key and delta frames might differ from frame count.
|
| - if (frame.IsSessionComplete()) {
|
| - if (frame.FrameType() == kVideoFrameKey) {
|
| - ++receive_statistics_.key_frames;
|
| - } else {
|
| - ++receive_statistics_.delta_frames;
|
| - }
|
| - if (stats_callback_ != NULL)
|
| - stats_callback_->OnFrameCountsUpdated(receive_statistics_);
|
| - }
|
| -}
|
| -
|
| -void VCMJitterBuffer::UpdateAveragePacketsPerFrame(int current_number_packets) {
|
| - if (frame_counter_ > kFastConvergeThreshold) {
|
| - average_packets_per_frame_ = average_packets_per_frame_
|
| - * (1 - kNormalConvergeMultiplier)
|
| - + current_number_packets * kNormalConvergeMultiplier;
|
| - } else if (frame_counter_ > 0) {
|
| - average_packets_per_frame_ = average_packets_per_frame_
|
| - * (1 - kFastConvergeMultiplier)
|
| - + current_number_packets * kFastConvergeMultiplier;
|
| - frame_counter_++;
|
| - } else {
|
| - average_packets_per_frame_ = current_number_packets;
|
| - frame_counter_++;
|
| - }
|
| -}
|
| -
|
| -// Must be called under the critical section |crit_sect_|.
|
| -void VCMJitterBuffer::CleanUpOldOrEmptyFrames() {
|
| - decodable_frames_.CleanUpOldOrEmptyFrames(&last_decoded_state_,
|
| - &free_frames_);
|
| - incomplete_frames_.CleanUpOldOrEmptyFrames(&last_decoded_state_,
|
| - &free_frames_);
|
| - if (!last_decoded_state_.in_initial_state()) {
|
| - DropPacketsFromNackList(last_decoded_state_.sequence_num());
|
| - }
|
| -}
|
| -
|
| -// Must be called from within |crit_sect_|.
|
| -bool VCMJitterBuffer::IsPacketRetransmitted(const VCMPacket& packet) const {
|
| - return missing_sequence_numbers_.find(packet.seqNum) !=
|
| - missing_sequence_numbers_.end();
|
| -}
|
| -
|
| -// Must be called under the critical section |crit_sect_|. Should never be
|
| -// called with retransmitted frames, they must be filtered out before this
|
| -// function is called.
|
| -void VCMJitterBuffer::UpdateJitterEstimate(const VCMJitterSample& sample,
|
| - bool incomplete_frame) {
|
| - if (sample.latest_packet_time == -1) {
|
| - return;
|
| - }
|
| - UpdateJitterEstimate(sample.latest_packet_time, sample.timestamp,
|
| - sample.frame_size, incomplete_frame);
|
| -}
|
| -
|
| -// Must be called under the critical section crit_sect_. Should never be
|
| -// called with retransmitted frames, they must be filtered out before this
|
| -// function is called.
|
| -void VCMJitterBuffer::UpdateJitterEstimate(const VCMFrameBuffer& frame,
|
| - bool incomplete_frame) {
|
| - if (frame.LatestPacketTimeMs() == -1) {
|
| - return;
|
| - }
|
| - // No retransmitted frames should be a part of the jitter
|
| - // estimate.
|
| - UpdateJitterEstimate(frame.LatestPacketTimeMs(), frame.TimeStamp(),
|
| - frame.Length(), incomplete_frame);
|
| -}
|
| -
|
| -// Must be called under the critical section |crit_sect_|. Should never be
|
| -// called with retransmitted frames, they must be filtered out before this
|
| -// function is called.
|
| -void VCMJitterBuffer::UpdateJitterEstimate(
|
| - int64_t latest_packet_time_ms,
|
| - uint32_t timestamp,
|
| - unsigned int frame_size,
|
| - bool incomplete_frame) {
|
| - if (latest_packet_time_ms == -1) {
|
| - return;
|
| - }
|
| - int64_t frame_delay;
|
| - bool not_reordered = inter_frame_delay_.CalculateDelay(timestamp,
|
| - &frame_delay,
|
| - latest_packet_time_ms);
|
| - // Filter out frames which have been reordered in time by the network
|
| - if (not_reordered) {
|
| - // Update the jitter estimate with the new samples
|
| - jitter_estimate_.UpdateEstimate(frame_delay, frame_size, incomplete_frame);
|
| - }
|
| -}
|
| -
|
| -bool VCMJitterBuffer::WaitForRetransmissions() {
|
| - if (nack_mode_ == kNoNack) {
|
| - // NACK disabled -> don't wait for retransmissions.
|
| - return false;
|
| - }
|
| - // Evaluate if the RTT is higher than |high_rtt_nack_threshold_ms_|, and in
|
| - // that case we don't wait for retransmissions.
|
| - if (high_rtt_nack_threshold_ms_ >= 0 &&
|
| - rtt_ms_ >= high_rtt_nack_threshold_ms_) {
|
| - return false;
|
| - }
|
| - return true;
|
| -}
|
| -} // namespace webrtc
|
|
|