| OLD | NEW |
| 1 /* | 1 /* |
| 2 * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved. | 2 * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved. |
| 3 * | 3 * |
| 4 * Use of this source code is governed by a BSD-style license | 4 * Use of this source code is governed by a BSD-style license |
| 5 * that can be found in the LICENSE file in the root of the source | 5 * that can be found in the LICENSE file in the root of the source |
| 6 * tree. An additional intellectual property rights grant can be found | 6 * tree. An additional intellectual property rights grant can be found |
| 7 * in the file PATENTS. All contributing project authors may | 7 * in the file PATENTS. All contributing project authors may |
| 8 * be found in the AUTHORS file in the root of the source tree. | 8 * be found in the AUTHORS file in the root of the source tree. |
| 9 */ | 9 */ |
| 10 #include "webrtc/modules/video_coding/jitter_buffer.h" | 10 #include "webrtc/modules/video_coding/jitter_buffer.h" |
| (...skipping 75 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 86 if (it != end() && it->second->FrameType() == kVideoFrameKey) { | 86 if (it != end() && it->second->FrameType() == kVideoFrameKey) { |
| 87 *key_frame_it = it; | 87 *key_frame_it = it; |
| 88 return drop_count; | 88 return drop_count; |
| 89 } | 89 } |
| 90 } | 90 } |
| 91 *key_frame_it = end(); | 91 *key_frame_it = end(); |
| 92 return drop_count; | 92 return drop_count; |
| 93 } | 93 } |
| 94 | 94 |
| 95 void FrameList::CleanUpOldOrEmptyFrames(VCMDecodingState* decoding_state, | 95 void FrameList::CleanUpOldOrEmptyFrames(VCMDecodingState* decoding_state, |
| 96 UnorderedFrameList* free_frames) { | 96 UnorderedFrameList* free_frames) { |
| 97 while (!empty()) { | 97 while (!empty()) { |
| 98 VCMFrameBuffer* oldest_frame = Front(); | 98 VCMFrameBuffer* oldest_frame = Front(); |
| 99 bool remove_frame = false; | 99 bool remove_frame = false; |
| 100 if (oldest_frame->GetState() == kStateEmpty && size() > 1) { | 100 if (oldest_frame->GetState() == kStateEmpty && size() > 1) { |
| 101 // This frame is empty, try to update the last decoded state and drop it | 101 // This frame is empty, try to update the last decoded state and drop it |
| 102 // if successful. | 102 // if successful. |
| 103 remove_frame = decoding_state->UpdateEmptyFrame(oldest_frame); | 103 remove_frame = decoding_state->UpdateEmptyFrame(oldest_frame); |
| 104 } else { | 104 } else { |
| 105 remove_frame = decoding_state->IsOldFrame(oldest_frame); | 105 remove_frame = decoding_state->IsOldFrame(oldest_frame); |
| 106 } | 106 } |
| (...skipping 168 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 275 if (num_packets_ <= 0 || !running_) { | 275 if (num_packets_ <= 0 || !running_) { |
| 276 return; | 276 return; |
| 277 } | 277 } |
| 278 int64_t elapsed_sec = | 278 int64_t elapsed_sec = |
| 279 (clock_->TimeInMilliseconds() - time_first_packet_ms_) / 1000; | 279 (clock_->TimeInMilliseconds() - time_first_packet_ms_) / 1000; |
| 280 if (elapsed_sec < metrics::kMinRunTimeInSeconds) { | 280 if (elapsed_sec < metrics::kMinRunTimeInSeconds) { |
| 281 return; | 281 return; |
| 282 } | 282 } |
| 283 | 283 |
| 284 RTC_HISTOGRAM_PERCENTAGE("WebRTC.Video.DiscardedPacketsInPercent", | 284 RTC_HISTOGRAM_PERCENTAGE("WebRTC.Video.DiscardedPacketsInPercent", |
| 285 num_discarded_packets_ * 100 / num_packets_); | 285 num_discarded_packets_ * 100 / num_packets_); |
| 286 RTC_HISTOGRAM_PERCENTAGE("WebRTC.Video.DuplicatedPacketsInPercent", | 286 RTC_HISTOGRAM_PERCENTAGE("WebRTC.Video.DuplicatedPacketsInPercent", |
| 287 num_duplicated_packets_ * 100 / num_packets_); | 287 num_duplicated_packets_ * 100 / num_packets_); |
| 288 | 288 |
| 289 int total_frames = | 289 int total_frames = |
| 290 receive_statistics_.key_frames + receive_statistics_.delta_frames; | 290 receive_statistics_.key_frames + receive_statistics_.delta_frames; |
| 291 if (total_frames > 0) { | 291 if (total_frames > 0) { |
| 292 RTC_HISTOGRAM_COUNTS_100("WebRTC.Video.CompleteFramesReceivedPerSecond", | 292 RTC_HISTOGRAM_COUNTS_100( |
| 293 "WebRTC.Video.CompleteFramesReceivedPerSecond", |
| 293 static_cast<int>((total_frames / elapsed_sec) + 0.5f)); | 294 static_cast<int>((total_frames / elapsed_sec) + 0.5f)); |
| 294 RTC_HISTOGRAM_COUNTS_1000( | 295 RTC_HISTOGRAM_COUNTS_1000( |
| 295 "WebRTC.Video.KeyFramesReceivedInPermille", | 296 "WebRTC.Video.KeyFramesReceivedInPermille", |
| 296 static_cast<int>( | 297 static_cast<int>( |
| 297 (receive_statistics_.key_frames * 1000.0f / total_frames) + 0.5f)); | 298 (receive_statistics_.key_frames * 1000.0f / total_frames) + 0.5f)); |
| 298 } | 299 } |
| 299 } | 300 } |
| 300 | 301 |
| 301 void VCMJitterBuffer::Start() { | 302 void VCMJitterBuffer::Start() { |
| 302 CriticalSectionScoped cs(crit_sect_); | 303 CriticalSectionScoped cs(crit_sect_); |
| (...skipping 120 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 423 // (I.e. frames per second since last calculation.) | 424 // (I.e. frames per second since last calculation.) |
| 424 // frame_rate = r(0)/2 + r(-1)/2 | 425 // frame_rate = r(0)/2 + r(-1)/2 |
| 425 // (I.e. fr/s average this and the previous calculation.) | 426 // (I.e. fr/s average this and the previous calculation.) |
| 426 *framerate = (incoming_frame_rate_ + static_cast<unsigned int>(rate)) / 2; | 427 *framerate = (incoming_frame_rate_ + static_cast<unsigned int>(rate)) / 2; |
| 427 incoming_frame_rate_ = static_cast<unsigned int>(rate); | 428 incoming_frame_rate_ = static_cast<unsigned int>(rate); |
| 428 | 429 |
| 429 // Calculate bit rate | 430 // Calculate bit rate |
| 430 if (incoming_bit_count_ == 0) { | 431 if (incoming_bit_count_ == 0) { |
| 431 *bitrate = 0; | 432 *bitrate = 0; |
| 432 } else { | 433 } else { |
| 433 *bitrate = 10 * ((100 * incoming_bit_count_) / | 434 *bitrate = |
| 434 static_cast<unsigned int>(diff)); | 435 10 * ((100 * incoming_bit_count_) / static_cast<unsigned int>(diff)); |
| 435 } | 436 } |
| 436 incoming_bit_rate_ = *bitrate; | 437 incoming_bit_rate_ = *bitrate; |
| 437 | 438 |
| 438 // Reset count | 439 // Reset count |
| 439 incoming_frame_count_ = 0; | 440 incoming_frame_count_ = 0; |
| 440 incoming_bit_count_ = 0; | 441 incoming_bit_count_ = 0; |
| 441 time_last_incoming_frame_count_ = now; | 442 time_last_incoming_frame_count_ = now; |
| 442 | 443 |
| 443 } else { | 444 } else { |
| 444 // No frames since last call | 445 // No frames since last call |
| (...skipping 20 matching lines...) Expand all Loading... |
| 465 } | 466 } |
| 466 } else if (incomplete_frames_.size() <= 1) { | 467 } else if (incomplete_frames_.size() <= 1) { |
| 467 // Frame not ready to be decoded. | 468 // Frame not ready to be decoded. |
| 468 return true; | 469 return true; |
| 469 } | 470 } |
| 470 return false; | 471 return false; |
| 471 } | 472 } |
| 472 | 473 |
| 473 // Returns immediately or a |max_wait_time_ms| ms event hang waiting for a | 474 // Returns immediately or a |max_wait_time_ms| ms event hang waiting for a |
| 474 // complete frame, |max_wait_time_ms| decided by caller. | 475 // complete frame, |max_wait_time_ms| decided by caller. |
| 475 bool VCMJitterBuffer::NextCompleteTimestamp( | 476 bool VCMJitterBuffer::NextCompleteTimestamp(uint32_t max_wait_time_ms, |
| 476 uint32_t max_wait_time_ms, uint32_t* timestamp) { | 477 uint32_t* timestamp) { |
| 477 crit_sect_->Enter(); | 478 crit_sect_->Enter(); |
| 478 if (!running_) { | 479 if (!running_) { |
| 479 crit_sect_->Leave(); | 480 crit_sect_->Leave(); |
| 480 return false; | 481 return false; |
| 481 } | 482 } |
| 482 CleanUpOldOrEmptyFrames(); | 483 CleanUpOldOrEmptyFrames(); |
| 483 | 484 |
| 484 if (decodable_frames_.empty() || | 485 if (decodable_frames_.empty() || |
| 485 decodable_frames_.Front()->GetState() != kStateComplete) { | 486 decodable_frames_.Front()->GetState() != kStateComplete) { |
| 486 const int64_t end_wait_time_ms = clock_->TimeInMilliseconds() + | 487 const int64_t end_wait_time_ms = |
| 487 max_wait_time_ms; | 488 clock_->TimeInMilliseconds() + max_wait_time_ms; |
| 488 int64_t wait_time_ms = max_wait_time_ms; | 489 int64_t wait_time_ms = max_wait_time_ms; |
| 489 while (wait_time_ms > 0) { | 490 while (wait_time_ms > 0) { |
| 490 crit_sect_->Leave(); | 491 crit_sect_->Leave(); |
| 491 const EventTypeWrapper ret = | 492 const EventTypeWrapper ret = |
| 492 frame_event_->Wait(static_cast<uint32_t>(wait_time_ms)); | 493 frame_event_->Wait(static_cast<uint32_t>(wait_time_ms)); |
| 493 crit_sect_->Enter(); | 494 crit_sect_->Enter(); |
| 494 if (ret == kEventSignaled) { | 495 if (ret == kEventSignaled) { |
| 495 // Are we shutting down the jitter buffer? | 496 // Are we shutting down the jitter buffer? |
| 496 if (!running_) { | 497 if (!running_) { |
| 497 crit_sect_->Leave(); | 498 crit_sect_->Leave(); |
| 498 return false; | 499 return false; |
| 499 } | 500 } |
| 500 // Finding oldest frame ready for decoder. | 501 // Finding oldest frame ready for decoder. |
| 501 CleanUpOldOrEmptyFrames(); | 502 CleanUpOldOrEmptyFrames(); |
| 502 if (decodable_frames_.empty() || | 503 if (decodable_frames_.empty() || |
| (...skipping 37 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 540 oldest_frame = incomplete_frames_.Front(); | 541 oldest_frame = incomplete_frames_.Front(); |
| 541 // Frame will only be removed from buffer if it is complete (or decodable). | 542 // Frame will only be removed from buffer if it is complete (or decodable). |
| 542 if (oldest_frame->GetState() < kStateComplete) { | 543 if (oldest_frame->GetState() < kStateComplete) { |
| 543 return false; | 544 return false; |
| 544 } | 545 } |
| 545 } else { | 546 } else { |
| 546 oldest_frame = decodable_frames_.Front(); | 547 oldest_frame = decodable_frames_.Front(); |
| 547 // If we have exactly one frame in the buffer, release it only if it is | 548 // If we have exactly one frame in the buffer, release it only if it is |
| 548 // complete. We know decodable_frames_ is not empty due to the previous | 549 // complete. We know decodable_frames_ is not empty due to the previous |
| 549 // check. | 550 // check. |
| 550 if (decodable_frames_.size() == 1 && incomplete_frames_.empty() | 551 if (decodable_frames_.size() == 1 && incomplete_frames_.empty() && |
| 551 && oldest_frame->GetState() != kStateComplete) { | 552 oldest_frame->GetState() != kStateComplete) { |
| 552 return false; | 553 return false; |
| 553 } | 554 } |
| 554 } | 555 } |
| 555 | 556 |
| 556 *timestamp = oldest_frame->TimeStamp(); | 557 *timestamp = oldest_frame->TimeStamp(); |
| 557 return true; | 558 return true; |
| 558 } | 559 } |
| 559 | 560 |
| 560 VCMEncodedFrame* VCMJitterBuffer::ExtractAndSetDecode(uint32_t timestamp) { | 561 VCMEncodedFrame* VCMJitterBuffer::ExtractAndSetDecode(uint32_t timestamp) { |
| 561 CriticalSectionScoped cs(crit_sect_); | 562 CriticalSectionScoped cs(crit_sect_); |
| (...skipping 18 matching lines...) Expand all Loading... |
| 580 } else if (frame->Length() > 0) { | 581 } else if (frame->Length() > 0) { |
| 581 // Ignore retransmitted and empty frames. | 582 // Ignore retransmitted and empty frames. |
| 582 if (waiting_for_completion_.latest_packet_time >= 0) { | 583 if (waiting_for_completion_.latest_packet_time >= 0) { |
| 583 UpdateJitterEstimate(waiting_for_completion_, true); | 584 UpdateJitterEstimate(waiting_for_completion_, true); |
| 584 } | 585 } |
| 585 if (frame->GetState() == kStateComplete) { | 586 if (frame->GetState() == kStateComplete) { |
| 586 UpdateJitterEstimate(*frame, false); | 587 UpdateJitterEstimate(*frame, false); |
| 587 } else { | 588 } else { |
| 588 // Wait for this one to get complete. | 589 // Wait for this one to get complete. |
| 589 waiting_for_completion_.frame_size = frame->Length(); | 590 waiting_for_completion_.frame_size = frame->Length(); |
| 590 waiting_for_completion_.latest_packet_time = | 591 waiting_for_completion_.latest_packet_time = frame->LatestPacketTimeMs(); |
| 591 frame->LatestPacketTimeMs(); | |
| 592 waiting_for_completion_.timestamp = frame->TimeStamp(); | 592 waiting_for_completion_.timestamp = frame->TimeStamp(); |
| 593 } | 593 } |
| 594 } | 594 } |
| 595 | 595 |
| 596 // The state must be changed to decoding before cleaning up zero sized | 596 // The state must be changed to decoding before cleaning up zero sized |
| 597 // frames to avoid empty frames being cleaned up and then given to the | 597 // frames to avoid empty frames being cleaned up and then given to the |
| 598 // decoder. Propagates the missing_frame bit. | 598 // decoder. Propagates the missing_frame bit. |
| 599 frame->PrepareForDecode(continuous); | 599 frame->PrepareForDecode(continuous); |
| 600 | 600 |
| 601 // We have a frame - update the last decoded state and nack list. | 601 // We have a frame - update the last decoded state and nack list. |
| (...skipping 132 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 734 | 734 |
| 735 VCMFrameBufferStateEnum previous_state = frame->GetState(); | 735 VCMFrameBufferStateEnum previous_state = frame->GetState(); |
| 736 // Insert packet. | 736 // Insert packet. |
| 737 FrameData frame_data; | 737 FrameData frame_data; |
| 738 frame_data.rtt_ms = rtt_ms_; | 738 frame_data.rtt_ms = rtt_ms_; |
| 739 frame_data.rolling_average_packets_per_frame = average_packets_per_frame_; | 739 frame_data.rolling_average_packets_per_frame = average_packets_per_frame_; |
| 740 VCMFrameBufferEnum buffer_state = | 740 VCMFrameBufferEnum buffer_state = |
| 741 frame->InsertPacket(packet, now_ms, decode_error_mode_, frame_data); | 741 frame->InsertPacket(packet, now_ms, decode_error_mode_, frame_data); |
| 742 | 742 |
| 743 if (previous_state != kStateComplete) { | 743 if (previous_state != kStateComplete) { |
| 744 TRACE_EVENT_ASYNC_BEGIN1("webrtc", "Video", frame->TimeStamp(), | 744 TRACE_EVENT_ASYNC_BEGIN1("webrtc", "Video", frame->TimeStamp(), "timestamp", |
| 745 "timestamp", frame->TimeStamp()); | 745 frame->TimeStamp()); |
| 746 } | 746 } |
| 747 | 747 |
| 748 if (buffer_state > 0) { | 748 if (buffer_state > 0) { |
| 749 incoming_bit_count_ += packet.sizeBytes << 3; | 749 incoming_bit_count_ += packet.sizeBytes << 3; |
| 750 if (first_packet_since_reset_) { | 750 if (first_packet_since_reset_) { |
| 751 latest_received_sequence_number_ = packet.seqNum; | 751 latest_received_sequence_number_ = packet.seqNum; |
| 752 first_packet_since_reset_ = false; | 752 first_packet_since_reset_ = false; |
| 753 } else { | 753 } else { |
| 754 if (IsPacketRetransmitted(packet)) { | 754 if (IsPacketRetransmitted(packet)) { |
| 755 frame->IncrementNackCount(); | 755 frame->IncrementNackCount(); |
| 756 } | 756 } |
| 757 if (!UpdateNackList(packet.seqNum) && | 757 if (!UpdateNackList(packet.seqNum) && |
| 758 packet.frameType != kVideoFrameKey) { | 758 packet.frameType != kVideoFrameKey) { |
| 759 buffer_state = kFlushIndicator; | 759 buffer_state = kFlushIndicator; |
| 760 } | 760 } |
| 761 | 761 |
| 762 latest_received_sequence_number_ = LatestSequenceNumber( | 762 latest_received_sequence_number_ = |
| 763 latest_received_sequence_number_, packet.seqNum); | 763 LatestSequenceNumber(latest_received_sequence_number_, packet.seqNum); |
| 764 } | 764 } |
| 765 } | 765 } |
| 766 | 766 |
| 767 // Is the frame already in the decodable list? | 767 // Is the frame already in the decodable list? |
| 768 bool continuous = IsContinuous(*frame); | 768 bool continuous = IsContinuous(*frame); |
| 769 switch (buffer_state) { | 769 switch (buffer_state) { |
| 770 case kGeneralError: | 770 case kGeneralError: |
| 771 case kTimeStampError: | 771 case kTimeStampError: |
| 772 case kSizeError: { | 772 case kSizeError: { |
| 773 free_frames_.push_back(frame); | 773 free_frames_.push_back(frame); |
| (...skipping 12 matching lines...) Expand all Loading... |
| 786 } | 786 } |
| 787 // Note: There is no break here - continuing to kDecodableSession. | 787 // Note: There is no break here - continuing to kDecodableSession. |
| 788 case kDecodableSession: { | 788 case kDecodableSession: { |
| 789 *retransmitted = (frame->GetNackCount() > 0); | 789 *retransmitted = (frame->GetNackCount() > 0); |
| 790 if (continuous) { | 790 if (continuous) { |
| 791 decodable_frames_.InsertFrame(frame); | 791 decodable_frames_.InsertFrame(frame); |
| 792 FindAndInsertContinuousFrames(*frame); | 792 FindAndInsertContinuousFrames(*frame); |
| 793 } else { | 793 } else { |
| 794 incomplete_frames_.InsertFrame(frame); | 794 incomplete_frames_.InsertFrame(frame); |
| 795 // If NACKs are enabled, keyframes are triggered by |GetNackList|. | 795 // If NACKs are enabled, keyframes are triggered by |GetNackList|. |
| 796 if (nack_mode_ == kNoNack && NonContinuousOrIncompleteDuration() > | 796 if (nack_mode_ == kNoNack && |
| 797 90 * kMaxDiscontinuousFramesTime) { | 797 NonContinuousOrIncompleteDuration() > |
| 798 90 * kMaxDiscontinuousFramesTime) { |
| 798 return kFlushIndicator; | 799 return kFlushIndicator; |
| 799 } | 800 } |
| 800 } | 801 } |
| 801 break; | 802 break; |
| 802 } | 803 } |
| 803 case kIncomplete: { | 804 case kIncomplete: { |
| 804 if (frame->GetState() == kStateEmpty && | 805 if (frame->GetState() == kStateEmpty && |
| 805 last_decoded_state_.UpdateEmptyFrame(frame)) { | 806 last_decoded_state_.UpdateEmptyFrame(frame)) { |
| 806 free_frames_.push_back(frame); | 807 free_frames_.push_back(frame); |
| 807 return kNoError; | 808 return kNoError; |
| 808 } else { | 809 } else { |
| 809 incomplete_frames_.InsertFrame(frame); | 810 incomplete_frames_.InsertFrame(frame); |
| 810 // If NACKs are enabled, keyframes are triggered by |GetNackList|. | 811 // If NACKs are enabled, keyframes are triggered by |GetNackList|. |
| 811 if (nack_mode_ == kNoNack && NonContinuousOrIncompleteDuration() > | 812 if (nack_mode_ == kNoNack && |
| 812 90 * kMaxDiscontinuousFramesTime) { | 813 NonContinuousOrIncompleteDuration() > |
| 814 90 * kMaxDiscontinuousFramesTime) { |
| 813 return kFlushIndicator; | 815 return kFlushIndicator; |
| 814 } | 816 } |
| 815 } | 817 } |
| 816 break; | 818 break; |
| 817 } | 819 } |
| 818 case kNoError: | 820 case kNoError: |
| 819 case kOutOfBoundsPacket: | 821 case kOutOfBoundsPacket: |
| 820 case kDuplicatePacket: { | 822 case kDuplicatePacket: { |
| 821 // Put back the frame where it came from. | 823 // Put back the frame where it came from. |
| 822 if (frame_list != NULL) { | 824 if (frame_list != NULL) { |
| 823 frame_list->InsertFrame(frame); | 825 frame_list->InsertFrame(frame); |
| 824 } else { | 826 } else { |
| 825 free_frames_.push_back(frame); | 827 free_frames_.push_back(frame); |
| 826 } | 828 } |
| 827 ++num_duplicated_packets_; | 829 ++num_duplicated_packets_; |
| 828 break; | 830 break; |
| 829 } | 831 } |
| 830 case kFlushIndicator: | 832 case kFlushIndicator: |
| 831 free_frames_.push_back(frame); | 833 free_frames_.push_back(frame); |
| 832 return kFlushIndicator; | 834 return kFlushIndicator; |
| 833 default: assert(false); | 835 default: |
| 836 assert(false); |
| 834 } | 837 } |
| 835 return buffer_state; | 838 return buffer_state; |
| 836 } | 839 } |
| 837 | 840 |
| 838 bool VCMJitterBuffer::IsContinuousInState(const VCMFrameBuffer& frame, | 841 bool VCMJitterBuffer::IsContinuousInState( |
| 842 const VCMFrameBuffer& frame, |
| 839 const VCMDecodingState& decoding_state) const { | 843 const VCMDecodingState& decoding_state) const { |
| 840 // Is this frame (complete or decodable) and continuous? | 844 // Is this frame (complete or decodable) and continuous? |
| 841 // kStateDecodable will never be set when decode_error_mode_ is false | 845 // kStateDecodable will never be set when decode_error_mode_ is false |
| 842 // as SessionInfo determines this state based on the error mode (and frame | 846 // as SessionInfo determines this state based on the error mode (and frame |
| 843 // completeness). | 847 // completeness). |
| 844 return (frame.GetState() == kStateComplete || | 848 return (frame.GetState() == kStateComplete || |
| 845 frame.GetState() == kStateDecodable) && | 849 frame.GetState() == kStateDecodable) && |
| 846 decoding_state.ContinuousFrame(&frame); | 850 decoding_state.ContinuousFrame(&frame); |
| 847 } | 851 } |
| 848 | 852 |
| 849 bool VCMJitterBuffer::IsContinuous(const VCMFrameBuffer& frame) const { | 853 bool VCMJitterBuffer::IsContinuous(const VCMFrameBuffer& frame) const { |
| 850 if (IsContinuousInState(frame, last_decoded_state_)) { | 854 if (IsContinuousInState(frame, last_decoded_state_)) { |
| 851 return true; | 855 return true; |
| 852 } | 856 } |
| 853 VCMDecodingState decoding_state; | 857 VCMDecodingState decoding_state; |
| 854 decoding_state.CopyFrom(last_decoded_state_); | 858 decoding_state.CopyFrom(last_decoded_state_); |
| 855 for (FrameList::const_iterator it = decodable_frames_.begin(); | 859 for (FrameList::const_iterator it = decodable_frames_.begin(); |
| 856 it != decodable_frames_.end(); ++it) { | 860 it != decodable_frames_.end(); ++it) { |
| 857 VCMFrameBuffer* decodable_frame = it->second; | 861 VCMFrameBuffer* decodable_frame = it->second; |
| 858 if (IsNewerTimestamp(decodable_frame->TimeStamp(), frame.TimeStamp())) { | 862 if (IsNewerTimestamp(decodable_frame->TimeStamp(), frame.TimeStamp())) { |
| 859 break; | 863 break; |
| 860 } | 864 } |
| 861 decoding_state.SetState(decodable_frame); | 865 decoding_state.SetState(decodable_frame); |
| 862 if (IsContinuousInState(frame, decoding_state)) { | 866 if (IsContinuousInState(frame, decoding_state)) { |
| 863 return true; | 867 return true; |
| 864 } | 868 } |
| 865 } | 869 } |
| 866 return false; | 870 return false; |
| (...skipping 12 matching lines...) Expand all Loading... |
| 879 // Copy original_decoded_state so we can move the state forward with each | 883 // Copy original_decoded_state so we can move the state forward with each |
| 880 // decodable frame we find. | 884 // decodable frame we find. |
| 881 VCMDecodingState decoding_state; | 885 VCMDecodingState decoding_state; |
| 882 decoding_state.CopyFrom(original_decoded_state); | 886 decoding_state.CopyFrom(original_decoded_state); |
| 883 | 887 |
| 884 // When temporal layers are available, we search for a complete or decodable | 888 // When temporal layers are available, we search for a complete or decodable |
| 885 // frame until we hit one of the following: | 889 // frame until we hit one of the following: |
| 886 // 1. Continuous base or sync layer. | 890 // 1. Continuous base or sync layer. |
| 887 // 2. The end of the list was reached. | 891 // 2. The end of the list was reached. |
| 888 for (FrameList::iterator it = incomplete_frames_.begin(); | 892 for (FrameList::iterator it = incomplete_frames_.begin(); |
| 889 it != incomplete_frames_.end();) { | 893 it != incomplete_frames_.end();) { |
| 890 VCMFrameBuffer* frame = it->second; | 894 VCMFrameBuffer* frame = it->second; |
| 891 if (IsNewerTimestamp(original_decoded_state.time_stamp(), | 895 if (IsNewerTimestamp(original_decoded_state.time_stamp(), |
| 892 frame->TimeStamp())) { | 896 frame->TimeStamp())) { |
| 893 ++it; | 897 ++it; |
| 894 continue; | 898 continue; |
| 895 } | 899 } |
| 896 if (IsContinuousInState(*frame, decoding_state)) { | 900 if (IsContinuousInState(*frame, decoding_state)) { |
| 897 decodable_frames_.InsertFrame(frame); | 901 decodable_frames_.InsertFrame(frame); |
| 898 incomplete_frames_.erase(it++); | 902 incomplete_frames_.erase(it++); |
| 899 decoding_state.SetState(frame); | 903 decoding_state.SetState(frame); |
| (...skipping 89 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 989 | 993 |
| 990 std::vector<uint16_t> VCMJitterBuffer::GetNackList(bool* request_key_frame) { | 994 std::vector<uint16_t> VCMJitterBuffer::GetNackList(bool* request_key_frame) { |
| 991 CriticalSectionScoped cs(crit_sect_); | 995 CriticalSectionScoped cs(crit_sect_); |
| 992 *request_key_frame = false; | 996 *request_key_frame = false; |
| 993 if (nack_mode_ == kNoNack) { | 997 if (nack_mode_ == kNoNack) { |
| 994 return std::vector<uint16_t>(); | 998 return std::vector<uint16_t>(); |
| 995 } | 999 } |
| 996 if (last_decoded_state_.in_initial_state()) { | 1000 if (last_decoded_state_.in_initial_state()) { |
| 997 VCMFrameBuffer* next_frame = NextFrame(); | 1001 VCMFrameBuffer* next_frame = NextFrame(); |
| 998 const bool first_frame_is_key = next_frame && | 1002 const bool first_frame_is_key = next_frame && |
| 999 next_frame->FrameType() == kVideoFrameKey && | 1003 next_frame->FrameType() == kVideoFrameKey && |
| 1000 next_frame->HaveFirstPacket(); | 1004 next_frame->HaveFirstPacket(); |
| 1001 if (!first_frame_is_key) { | 1005 if (!first_frame_is_key) { |
| 1002 bool have_non_empty_frame = decodable_frames_.end() != find_if( | 1006 bool have_non_empty_frame = |
| 1003 decodable_frames_.begin(), decodable_frames_.end(), | 1007 decodable_frames_.end() != find_if(decodable_frames_.begin(), |
| 1004 HasNonEmptyState); | 1008 decodable_frames_.end(), |
| 1009 HasNonEmptyState); |
| 1005 if (!have_non_empty_frame) { | 1010 if (!have_non_empty_frame) { |
| 1006 have_non_empty_frame = incomplete_frames_.end() != find_if( | 1011 have_non_empty_frame = |
| 1007 incomplete_frames_.begin(), incomplete_frames_.end(), | 1012 incomplete_frames_.end() != find_if(incomplete_frames_.begin(), |
| 1008 HasNonEmptyState); | 1013 incomplete_frames_.end(), |
| 1014 HasNonEmptyState); |
| 1009 } | 1015 } |
| 1010 bool found_key_frame = RecycleFramesUntilKeyFrame(); | 1016 bool found_key_frame = RecycleFramesUntilKeyFrame(); |
| 1011 if (!found_key_frame) { | 1017 if (!found_key_frame) { |
| 1012 *request_key_frame = have_non_empty_frame; | 1018 *request_key_frame = have_non_empty_frame; |
| 1013 return std::vector<uint16_t>(); | 1019 return std::vector<uint16_t>(); |
| 1014 } | 1020 } |
| 1015 } | 1021 } |
| 1016 } | 1022 } |
| 1017 if (TooLargeNackList()) { | 1023 if (TooLargeNackList()) { |
| 1018 *request_key_frame = !HandleTooLargeNackList(); | 1024 *request_key_frame = !HandleTooLargeNackList(); |
| 1019 } | 1025 } |
| 1020 if (max_incomplete_time_ms_ > 0) { | 1026 if (max_incomplete_time_ms_ > 0) { |
| 1021 int non_continuous_incomplete_duration = | 1027 int non_continuous_incomplete_duration = |
| 1022 NonContinuousOrIncompleteDuration(); | 1028 NonContinuousOrIncompleteDuration(); |
| 1023 if (non_continuous_incomplete_duration > 90 * max_incomplete_time_ms_) { | 1029 if (non_continuous_incomplete_duration > 90 * max_incomplete_time_ms_) { |
| 1024 LOG_F(LS_WARNING) << "Too long non-decodable duration: " | 1030 LOG_F(LS_WARNING) << "Too long non-decodable duration: " |
| 1025 << non_continuous_incomplete_duration << " > " | 1031 << non_continuous_incomplete_duration << " > " |
| 1026 << 90 * max_incomplete_time_ms_; | 1032 << 90 * max_incomplete_time_ms_; |
| 1027 FrameList::reverse_iterator rit = find_if(incomplete_frames_.rbegin(), | 1033 FrameList::reverse_iterator rit = find_if( |
| 1028 incomplete_frames_.rend(), IsKeyFrame); | 1034 incomplete_frames_.rbegin(), incomplete_frames_.rend(), IsKeyFrame); |
| 1029 if (rit == incomplete_frames_.rend()) { | 1035 if (rit == incomplete_frames_.rend()) { |
| 1030 // Request a key frame if we don't have one already. | 1036 // Request a key frame if we don't have one already. |
| 1031 *request_key_frame = true; | 1037 *request_key_frame = true; |
| 1032 return std::vector<uint16_t>(); | 1038 return std::vector<uint16_t>(); |
| 1033 } else { | 1039 } else { |
| 1034 // Skip to the last key frame. If it's incomplete we will start | 1040 // Skip to the last key frame. If it's incomplete we will start |
| 1035 // NACKing it. | 1041 // NACKing it. |
| 1036 // Note that the estimated low sequence number is correct for VP8 | 1042 // Note that the estimated low sequence number is correct for VP8 |
| 1037 // streams because only the first packet of a key frame is marked. | 1043 // streams because only the first packet of a key frame is marked. |
| 1038 last_decoded_state_.Reset(); | 1044 last_decoded_state_.Reset(); |
| (...skipping 19 matching lines...) Expand all Loading... |
| 1058 return NULL; | 1064 return NULL; |
| 1059 } | 1065 } |
| 1060 | 1066 |
| 1061 bool VCMJitterBuffer::UpdateNackList(uint16_t sequence_number) { | 1067 bool VCMJitterBuffer::UpdateNackList(uint16_t sequence_number) { |
| 1062 if (nack_mode_ == kNoNack) { | 1068 if (nack_mode_ == kNoNack) { |
| 1063 return true; | 1069 return true; |
| 1064 } | 1070 } |
| 1065 // Make sure we don't add packets which are already too old to be decoded. | 1071 // Make sure we don't add packets which are already too old to be decoded. |
| 1066 if (!last_decoded_state_.in_initial_state()) { | 1072 if (!last_decoded_state_.in_initial_state()) { |
| 1067 latest_received_sequence_number_ = LatestSequenceNumber( | 1073 latest_received_sequence_number_ = LatestSequenceNumber( |
| 1068 latest_received_sequence_number_, | 1074 latest_received_sequence_number_, last_decoded_state_.sequence_num()); |
| 1069 last_decoded_state_.sequence_num()); | |
| 1070 } | 1075 } |
| 1071 if (IsNewerSequenceNumber(sequence_number, | 1076 if (IsNewerSequenceNumber(sequence_number, |
| 1072 latest_received_sequence_number_)) { | 1077 latest_received_sequence_number_)) { |
| 1073 // Push any missing sequence numbers to the NACK list. | 1078 // Push any missing sequence numbers to the NACK list. |
| 1074 for (uint16_t i = latest_received_sequence_number_ + 1; | 1079 for (uint16_t i = latest_received_sequence_number_ + 1; |
| 1075 IsNewerSequenceNumber(sequence_number, i); ++i) { | 1080 IsNewerSequenceNumber(sequence_number, i); ++i) { |
| 1076 missing_sequence_numbers_.insert(missing_sequence_numbers_.end(), i); | 1081 missing_sequence_numbers_.insert(missing_sequence_numbers_.end(), i); |
| 1077 TRACE_EVENT_INSTANT1(TRACE_DISABLED_BY_DEFAULT("webrtc_rtp"), "AddNack", | 1082 TRACE_EVENT_INSTANT1(TRACE_DISABLED_BY_DEFAULT("webrtc_rtp"), "AddNack", |
| 1078 "seqnum", i); | 1083 "seqnum", i); |
| 1079 } | 1084 } |
| (...skipping 29 matching lines...) Expand all Loading... |
| 1109 key_frame_found = RecycleFramesUntilKeyFrame(); | 1114 key_frame_found = RecycleFramesUntilKeyFrame(); |
| 1110 } | 1115 } |
| 1111 return key_frame_found; | 1116 return key_frame_found; |
| 1112 } | 1117 } |
| 1113 | 1118 |
| 1114 bool VCMJitterBuffer::MissingTooOldPacket( | 1119 bool VCMJitterBuffer::MissingTooOldPacket( |
| 1115 uint16_t latest_sequence_number) const { | 1120 uint16_t latest_sequence_number) const { |
| 1116 if (missing_sequence_numbers_.empty()) { | 1121 if (missing_sequence_numbers_.empty()) { |
| 1117 return false; | 1122 return false; |
| 1118 } | 1123 } |
| 1119 const uint16_t age_of_oldest_missing_packet = latest_sequence_number - | 1124 const uint16_t age_of_oldest_missing_packet = |
| 1120 *missing_sequence_numbers_.begin(); | 1125 latest_sequence_number - *missing_sequence_numbers_.begin(); |
| 1121 // Recycle frames if the NACK list contains too old sequence numbers as | 1126 // Recycle frames if the NACK list contains too old sequence numbers as |
| 1122 // the packets may have already been dropped by the sender. | 1127 // the packets may have already been dropped by the sender. |
| 1123 return age_of_oldest_missing_packet > max_packet_age_to_nack_; | 1128 return age_of_oldest_missing_packet > max_packet_age_to_nack_; |
| 1124 } | 1129 } |
| 1125 | 1130 |
| 1126 bool VCMJitterBuffer::HandleTooOldPackets(uint16_t latest_sequence_number) { | 1131 bool VCMJitterBuffer::HandleTooOldPackets(uint16_t latest_sequence_number) { |
| 1127 bool key_frame_found = false; | 1132 bool key_frame_found = false; |
| 1128 const uint16_t age_of_oldest_missing_packet = latest_sequence_number - | 1133 const uint16_t age_of_oldest_missing_packet = |
| 1129 *missing_sequence_numbers_.begin(); | 1134 latest_sequence_number - *missing_sequence_numbers_.begin(); |
| 1130 LOG_F(LS_WARNING) << "NACK list contains too old sequence numbers: " | 1135 LOG_F(LS_WARNING) << "NACK list contains too old sequence numbers: " |
| 1131 << age_of_oldest_missing_packet << " > " | 1136 << age_of_oldest_missing_packet << " > " |
| 1132 << max_packet_age_to_nack_; | 1137 << max_packet_age_to_nack_; |
| 1133 while (MissingTooOldPacket(latest_sequence_number)) { | 1138 while (MissingTooOldPacket(latest_sequence_number)) { |
| 1134 key_frame_found = RecycleFramesUntilKeyFrame(); | 1139 key_frame_found = RecycleFramesUntilKeyFrame(); |
| 1135 } | 1140 } |
| 1136 return key_frame_found; | 1141 return key_frame_found; |
| 1137 } | 1142 } |
| 1138 | 1143 |
| 1139 void VCMJitterBuffer::DropPacketsFromNackList( | 1144 void VCMJitterBuffer::DropPacketsFromNackList( |
| 1140 uint16_t last_decoded_sequence_number) { | 1145 uint16_t last_decoded_sequence_number) { |
| 1141 // Erase all sequence numbers from the NACK list which we won't need any | 1146 // Erase all sequence numbers from the NACK list which we won't need any |
| 1142 // longer. | 1147 // longer. |
| 1143 missing_sequence_numbers_.erase(missing_sequence_numbers_.begin(), | 1148 missing_sequence_numbers_.erase( |
| 1144 missing_sequence_numbers_.upper_bound( | 1149 missing_sequence_numbers_.begin(), |
| 1145 last_decoded_sequence_number)); | 1150 missing_sequence_numbers_.upper_bound(last_decoded_sequence_number)); |
| 1146 } | 1151 } |
| 1147 | 1152 |
| 1148 int64_t VCMJitterBuffer::LastDecodedTimestamp() const { | 1153 int64_t VCMJitterBuffer::LastDecodedTimestamp() const { |
| 1149 CriticalSectionScoped cs(crit_sect_); | 1154 CriticalSectionScoped cs(crit_sect_); |
| 1150 return last_decoded_state_.time_stamp(); | 1155 return last_decoded_state_.time_stamp(); |
| 1151 } | 1156 } |
| 1152 | 1157 |
| 1153 void VCMJitterBuffer::RenderBufferSize(uint32_t* timestamp_start, | 1158 void VCMJitterBuffer::RenderBufferSize(uint32_t* timestamp_start, |
| 1154 uint32_t* timestamp_end) { | 1159 uint32_t* timestamp_end) { |
| 1155 CriticalSectionScoped cs(crit_sect_); | 1160 CriticalSectionScoped cs(crit_sect_); |
| (...skipping 63 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1219 missing_sequence_numbers_.clear(); | 1224 missing_sequence_numbers_.clear(); |
| 1220 } | 1225 } |
| 1221 return key_frame_found; | 1226 return key_frame_found; |
| 1222 } | 1227 } |
| 1223 | 1228 |
| 1224 // Must be called under the critical section |crit_sect_|. | 1229 // Must be called under the critical section |crit_sect_|. |
| 1225 void VCMJitterBuffer::CountFrame(const VCMFrameBuffer& frame) { | 1230 void VCMJitterBuffer::CountFrame(const VCMFrameBuffer& frame) { |
| 1226 incoming_frame_count_++; | 1231 incoming_frame_count_++; |
| 1227 | 1232 |
| 1228 if (frame.FrameType() == kVideoFrameKey) { | 1233 if (frame.FrameType() == kVideoFrameKey) { |
| 1229 TRACE_EVENT_ASYNC_STEP0("webrtc", "Video", | 1234 TRACE_EVENT_ASYNC_STEP0("webrtc", "Video", frame.TimeStamp(), |
| 1230 frame.TimeStamp(), "KeyComplete"); | 1235 "KeyComplete"); |
| 1231 } else { | 1236 } else { |
| 1232 TRACE_EVENT_ASYNC_STEP0("webrtc", "Video", | 1237 TRACE_EVENT_ASYNC_STEP0("webrtc", "Video", frame.TimeStamp(), |
| 1233 frame.TimeStamp(), "DeltaComplete"); | 1238 "DeltaComplete"); |
| 1234 } | 1239 } |
| 1235 | 1240 |
| 1236 // Update receive statistics. We count all layers, thus when you use layers | 1241 // Update receive statistics. We count all layers, thus when you use layers |
| 1237 // adding all key and delta frames might differ from frame count. | 1242 // adding all key and delta frames might differ from frame count. |
| 1238 if (frame.IsSessionComplete()) { | 1243 if (frame.IsSessionComplete()) { |
| 1239 if (frame.FrameType() == kVideoFrameKey) { | 1244 if (frame.FrameType() == kVideoFrameKey) { |
| 1240 ++receive_statistics_.key_frames; | 1245 ++receive_statistics_.key_frames; |
| 1241 } else { | 1246 } else { |
| 1242 ++receive_statistics_.delta_frames; | 1247 ++receive_statistics_.delta_frames; |
| 1243 } | 1248 } |
| 1244 if (stats_callback_ != NULL) | 1249 if (stats_callback_ != NULL) |
| 1245 stats_callback_->OnFrameCountsUpdated(receive_statistics_); | 1250 stats_callback_->OnFrameCountsUpdated(receive_statistics_); |
| 1246 } | 1251 } |
| 1247 } | 1252 } |
| 1248 | 1253 |
| 1249 void VCMJitterBuffer::UpdateAveragePacketsPerFrame(int current_number_packets) { | 1254 void VCMJitterBuffer::UpdateAveragePacketsPerFrame(int current_number_packets) { |
| 1250 if (frame_counter_ > kFastConvergeThreshold) { | 1255 if (frame_counter_ > kFastConvergeThreshold) { |
| 1251 average_packets_per_frame_ = average_packets_per_frame_ | 1256 average_packets_per_frame_ = |
| 1252 * (1 - kNormalConvergeMultiplier) | 1257 average_packets_per_frame_ * (1 - kNormalConvergeMultiplier) + |
| 1253 + current_number_packets * kNormalConvergeMultiplier; | 1258 current_number_packets * kNormalConvergeMultiplier; |
| 1254 } else if (frame_counter_ > 0) { | 1259 } else if (frame_counter_ > 0) { |
| 1255 average_packets_per_frame_ = average_packets_per_frame_ | 1260 average_packets_per_frame_ = |
| 1256 * (1 - kFastConvergeMultiplier) | 1261 average_packets_per_frame_ * (1 - kFastConvergeMultiplier) + |
| 1257 + current_number_packets * kFastConvergeMultiplier; | 1262 current_number_packets * kFastConvergeMultiplier; |
| 1258 frame_counter_++; | 1263 frame_counter_++; |
| 1259 } else { | 1264 } else { |
| 1260 average_packets_per_frame_ = current_number_packets; | 1265 average_packets_per_frame_ = current_number_packets; |
| 1261 frame_counter_++; | 1266 frame_counter_++; |
| 1262 } | 1267 } |
| 1263 } | 1268 } |
| 1264 | 1269 |
| 1265 // Must be called under the critical section |crit_sect_|. | 1270 // Must be called under the critical section |crit_sect_|. |
| 1266 void VCMJitterBuffer::CleanUpOldOrEmptyFrames() { | 1271 void VCMJitterBuffer::CleanUpOldOrEmptyFrames() { |
| 1267 decodable_frames_.CleanUpOldOrEmptyFrames(&last_decoded_state_, | 1272 decodable_frames_.CleanUpOldOrEmptyFrames(&last_decoded_state_, |
| 1268 &free_frames_); | 1273 &free_frames_); |
| 1269 incomplete_frames_.CleanUpOldOrEmptyFrames(&last_decoded_state_, | 1274 incomplete_frames_.CleanUpOldOrEmptyFrames(&last_decoded_state_, |
| 1270 &free_frames_); | 1275 &free_frames_); |
| 1271 if (!last_decoded_state_.in_initial_state()) { | 1276 if (!last_decoded_state_.in_initial_state()) { |
| 1272 DropPacketsFromNackList(last_decoded_state_.sequence_num()); | 1277 DropPacketsFromNackList(last_decoded_state_.sequence_num()); |
| 1273 } | 1278 } |
| 1274 } | 1279 } |
| 1275 | 1280 |
| 1276 // Must be called from within |crit_sect_|. | 1281 // Must be called from within |crit_sect_|. |
| 1277 bool VCMJitterBuffer::IsPacketRetransmitted(const VCMPacket& packet) const { | 1282 bool VCMJitterBuffer::IsPacketRetransmitted(const VCMPacket& packet) const { |
| 1278 return missing_sequence_numbers_.find(packet.seqNum) != | 1283 return missing_sequence_numbers_.find(packet.seqNum) != |
| 1279 missing_sequence_numbers_.end(); | 1284 missing_sequence_numbers_.end(); |
| 1280 } | 1285 } |
| 1281 | 1286 |
| 1282 // Must be called under the critical section |crit_sect_|. Should never be | 1287 // Must be called under the critical section |crit_sect_|. Should never be |
| 1283 // called with retransmitted frames, they must be filtered out before this | 1288 // called with retransmitted frames, they must be filtered out before this |
| 1284 // function is called. | 1289 // function is called. |
| 1285 void VCMJitterBuffer::UpdateJitterEstimate(const VCMJitterSample& sample, | 1290 void VCMJitterBuffer::UpdateJitterEstimate(const VCMJitterSample& sample, |
| 1286 bool incomplete_frame) { | 1291 bool incomplete_frame) { |
| 1287 if (sample.latest_packet_time == -1) { | 1292 if (sample.latest_packet_time == -1) { |
| 1288 return; | 1293 return; |
| 1289 } | 1294 } |
| (...skipping 11 matching lines...) Expand all Loading... |
| 1301 } | 1306 } |
| 1302 // No retransmitted frames should be a part of the jitter | 1307 // No retransmitted frames should be a part of the jitter |
| 1303 // estimate. | 1308 // estimate. |
| 1304 UpdateJitterEstimate(frame.LatestPacketTimeMs(), frame.TimeStamp(), | 1309 UpdateJitterEstimate(frame.LatestPacketTimeMs(), frame.TimeStamp(), |
| 1305 frame.Length(), incomplete_frame); | 1310 frame.Length(), incomplete_frame); |
| 1306 } | 1311 } |
| 1307 | 1312 |
| 1308 // Must be called under the critical section |crit_sect_|. Should never be | 1313 // Must be called under the critical section |crit_sect_|. Should never be |
| 1309 // called with retransmitted frames, they must be filtered out before this | 1314 // called with retransmitted frames, they must be filtered out before this |
| 1310 // function is called. | 1315 // function is called. |
| 1311 void VCMJitterBuffer::UpdateJitterEstimate( | 1316 void VCMJitterBuffer::UpdateJitterEstimate(int64_t latest_packet_time_ms, |
| 1312 int64_t latest_packet_time_ms, | 1317 uint32_t timestamp, |
| 1313 uint32_t timestamp, | 1318 unsigned int frame_size, |
| 1314 unsigned int frame_size, | 1319 bool incomplete_frame) { |
| 1315 bool incomplete_frame) { | |
| 1316 if (latest_packet_time_ms == -1) { | 1320 if (latest_packet_time_ms == -1) { |
| 1317 return; | 1321 return; |
| 1318 } | 1322 } |
| 1319 int64_t frame_delay; | 1323 int64_t frame_delay; |
| 1320 bool not_reordered = inter_frame_delay_.CalculateDelay(timestamp, | 1324 bool not_reordered = inter_frame_delay_.CalculateDelay( |
| 1321 &frame_delay, | 1325 timestamp, &frame_delay, latest_packet_time_ms); |
| 1322 latest_packet_time_ms); | |
| 1323 // Filter out frames which have been reordered in time by the network | 1326 // Filter out frames which have been reordered in time by the network |
| 1324 if (not_reordered) { | 1327 if (not_reordered) { |
| 1325 // Update the jitter estimate with the new samples | 1328 // Update the jitter estimate with the new samples |
| 1326 jitter_estimate_.UpdateEstimate(frame_delay, frame_size, incomplete_frame); | 1329 jitter_estimate_.UpdateEstimate(frame_delay, frame_size, incomplete_frame); |
| 1327 } | 1330 } |
| 1328 } | 1331 } |
| 1329 | 1332 |
| 1330 bool VCMJitterBuffer::WaitForRetransmissions() { | 1333 bool VCMJitterBuffer::WaitForRetransmissions() { |
| 1331 if (nack_mode_ == kNoNack) { | 1334 if (nack_mode_ == kNoNack) { |
| 1332 // NACK disabled -> don't wait for retransmissions. | 1335 // NACK disabled -> don't wait for retransmissions. |
| 1333 return false; | 1336 return false; |
| 1334 } | 1337 } |
| 1335 // Evaluate if the RTT is higher than |high_rtt_nack_threshold_ms_|, and in | 1338 // Evaluate if the RTT is higher than |high_rtt_nack_threshold_ms_|, and in |
| 1336 // that case we don't wait for retransmissions. | 1339 // that case we don't wait for retransmissions. |
| 1337 if (high_rtt_nack_threshold_ms_ >= 0 && | 1340 if (high_rtt_nack_threshold_ms_ >= 0 && |
| 1338 rtt_ms_ >= high_rtt_nack_threshold_ms_) { | 1341 rtt_ms_ >= high_rtt_nack_threshold_ms_) { |
| 1339 return false; | 1342 return false; |
| 1340 } | 1343 } |
| 1341 return true; | 1344 return true; |
| 1342 } | 1345 } |
| 1343 } // namespace webrtc | 1346 } // namespace webrtc |
| OLD | NEW |