| Index: webrtc/video/video_send_stream_tests.cc
|
| diff --git a/webrtc/video/video_send_stream_tests.cc b/webrtc/video/video_send_stream_tests.cc
|
| index eefcfebc1866a94e6799b09787a3bae2ad63120b..1550d6c99c763d0b0d2e7f1b3a40d51065e17a8e 100644
|
| --- a/webrtc/video/video_send_stream_tests.cc
|
| +++ b/webrtc/video/video_send_stream_tests.cc
|
| @@ -58,6 +58,9 @@ class VideoSendStreamTest : public test::CallTest {
|
| void TestNackRetransmission(uint32_t retransmit_ssrc,
|
| uint8_t retransmit_payload_type);
|
| void TestPacketFragmentationSize(VideoFormat format, bool with_fec);
|
| +
|
| + void TestVp9NonFlexMode(uint8_t num_temporal_layers,
|
| + uint8_t num_spatial_layers);
|
| };
|
|
|
| TEST_F(VideoSendStreamTest, CanStartStartedStream) {
|
| @@ -1791,22 +1794,21 @@ TEST_F(VideoSendStreamTest, ReportsSentResolution) {
|
| RunBaseTest(&test, FakeNetworkPipe::Config());
|
| }
|
|
|
| -class VP9HeaderObeserver : public test::SendTest {
|
| +class Vp9HeaderObserver : public test::SendTest {
|
| public:
|
| - VP9HeaderObeserver()
|
| + Vp9HeaderObserver()
|
| : SendTest(VideoSendStreamTest::kDefaultTimeoutMs),
|
| vp9_encoder_(VP9Encoder::Create()),
|
| - vp9_settings_(VideoEncoder::GetDefaultVp9Settings()) {
|
| - vp9_settings_.numberOfTemporalLayers = 1;
|
| - vp9_settings_.numberOfSpatialLayers = 2;
|
| - }
|
| + vp9_settings_(VideoEncoder::GetDefaultVp9Settings()),
|
| + packets_sent_(0),
|
| + frames_sent_(0) {}
|
|
|
| virtual void ModifyConfigsHook(
|
| VideoSendStream::Config* send_config,
|
| std::vector<VideoReceiveStream::Config>* receive_configs,
|
| VideoEncoderConfig* encoder_config) {}
|
|
|
| - virtual void InspectHeader(RTPVideoHeaderVP9* vp9videoHeader) = 0;
|
| + virtual void InspectHeader(const RTPVideoHeaderVP9& vp9) = 0;
|
|
|
| private:
|
| const int kVp9PayloadType = 105;
|
| @@ -1815,11 +1817,14 @@ class VP9HeaderObeserver : public test::SendTest {
|
| std::vector<VideoReceiveStream::Config>* receive_configs,
|
| VideoEncoderConfig* encoder_config) override {
|
| encoder_config->encoder_specific_settings = &vp9_settings_;
|
| - encoder_config->content_type = VideoEncoderConfig::ContentType::kScreen;
|
| send_config->encoder_settings.encoder = vp9_encoder_.get();
|
| send_config->encoder_settings.payload_name = "VP9";
|
| send_config->encoder_settings.payload_type = kVp9PayloadType;
|
| ModifyConfigsHook(send_config, receive_configs, encoder_config);
|
| + EXPECT_EQ(1u, encoder_config->streams.size());
|
| + encoder_config->streams[0].temporal_layer_thresholds_bps.resize(
|
| + vp9_settings_.numberOfTemporalLayers - 1);
|
| + encoder_config_ = *encoder_config;
|
| }
|
|
|
| void PerformTest() override {
|
| @@ -1831,113 +1836,340 @@ class VP9HeaderObeserver : public test::SendTest {
|
| RTPHeader header;
|
| EXPECT_TRUE(parser_->Parse(packet, length, &header));
|
|
|
| - if (header.payloadType == kVp9PayloadType) {
|
| - RtpDepacketizerVp9 vp9depacketizer;
|
| - RtpDepacketizer::ParsedPayload vp9payload;
|
| - const uint8_t* vp9_packet = packet + header.headerLength;
|
| - size_t payload_length =
|
| - length - header.headerLength - header.paddingLength;
|
| -
|
| - if (payload_length > 0) {
|
| - bool parse_vp9header_successful =
|
| - vp9depacketizer.Parse(&vp9payload, vp9_packet, payload_length);
|
| - bool is_vp9_codec_type =
|
| - vp9payload.type.Video.codec == RtpVideoCodecTypes::kRtpVideoVp9;
|
| - EXPECT_TRUE(parse_vp9header_successful);
|
| - EXPECT_TRUE(is_vp9_codec_type);
|
| -
|
| - RTPVideoHeaderVP9* vp9videoHeader =
|
| - &vp9payload.type.Video.codecHeader.VP9;
|
| - if (parse_vp9header_successful && is_vp9_codec_type) {
|
| - InspectHeader(vp9videoHeader);
|
| - } else {
|
| - observation_complete_->Set();
|
| - }
|
| + EXPECT_EQ(kVp9PayloadType, header.payloadType);
|
| + const uint8_t* payload = packet + header.headerLength;
|
| + size_t payload_length = length - header.headerLength - header.paddingLength;
|
| +
|
| + bool new_packet = packets_sent_ == 0 ||
|
| + IsNewerSequenceNumber(header.sequenceNumber,
|
| + last_header_.sequenceNumber);
|
| + if (payload_length > 0 && new_packet) {
|
| + RtpDepacketizer::ParsedPayload parsed;
|
| + RtpDepacketizerVp9 depacketizer;
|
| + EXPECT_TRUE(depacketizer.Parse(&parsed, payload, payload_length));
|
| + EXPECT_EQ(RtpVideoCodecTypes::kRtpVideoVp9, parsed.type.Video.codec);
|
| + // Verify common fields for all configurations.
|
| + VerifyCommonHeader(parsed.type.Video.codecHeader.VP9);
|
| + CompareConsecutiveFrames(header, parsed.type.Video);
|
| + // Verify configuration specific settings.
|
| + InspectHeader(parsed.type.Video.codecHeader.VP9);
|
| +
|
| + ++packets_sent_;
|
| + if (header.markerBit) {
|
| + ++frames_sent_;
|
| }
|
| + last_header_ = header;
|
| + last_vp9_ = parsed.type.Video.codecHeader.VP9;
|
| }
|
| -
|
| return SEND_PACKET;
|
| }
|
|
|
| protected:
|
| - rtc::scoped_ptr<VP9Encoder> vp9_encoder_;
|
| - VideoCodecVP9 vp9_settings_;
|
| -};
|
| + bool ContinuousPictureId(const RTPVideoHeaderVP9& vp9) const {
|
| + if (last_vp9_.picture_id > vp9.picture_id) {
|
| + return vp9.picture_id == 0; // Wrap.
|
| + } else {
|
| + return vp9.picture_id == last_vp9_.picture_id + 1;
|
| + }
|
| + }
|
|
|
| -TEST_F(VideoSendStreamTest, DISABLED_VP9FlexMode) {
|
| - class FlexibleMode : public VP9HeaderObeserver {
|
| - void ModifyConfigsHook(
|
| - VideoSendStream::Config* send_config,
|
| - std::vector<VideoReceiveStream::Config>* receive_configs,
|
| - VideoEncoderConfig* encoder_config) override {
|
| - vp9_settings_.flexibleMode = true;
|
| + void VerifySpatialIdxWithinFrame(const RTPVideoHeaderVP9& vp9) const {
|
| + if (frames_sent_ == 0)
|
| + return;
|
| +
|
| + bool new_layer = vp9.spatial_idx != last_vp9_.spatial_idx;
|
| + EXPECT_EQ(new_layer, vp9.beginning_of_frame);
|
| + EXPECT_EQ(new_layer, last_vp9_.end_of_frame);
|
| + EXPECT_EQ(new_layer ? last_vp9_.spatial_idx + 1 : last_vp9_.spatial_idx,
|
| + vp9.spatial_idx);
|
| + }
|
| +
|
| + void VerifyFixedTemporalLayerStructure(const RTPVideoHeaderVP9& vp9,
|
| + uint8_t num_layers) const {
|
| + switch (num_layers) {
|
| + case 0:
|
| + VerifyTemporalLayerStructure0(vp9);
|
| + break;
|
| + case 1:
|
| + VerifyTemporalLayerStructure1(vp9);
|
| + break;
|
| + case 2:
|
| + VerifyTemporalLayerStructure2(vp9);
|
| + break;
|
| + case 3:
|
| + VerifyTemporalLayerStructure3(vp9);
|
| + break;
|
| + default:
|
| + RTC_NOTREACHED();
|
| }
|
| + }
|
|
|
| - void InspectHeader(RTPVideoHeaderVP9* vp9videoHeader) override {
|
| - EXPECT_TRUE(vp9videoHeader->flexible_mode);
|
| - observation_complete_->Set();
|
| + void VerifyTemporalLayerStructure0(const RTPVideoHeaderVP9& vp9) const {
|
| + EXPECT_EQ(kNoTl0PicIdx, vp9.tl0_pic_idx);
|
| + EXPECT_EQ(kNoTemporalIdx, vp9.temporal_idx); // no tid
|
| + EXPECT_FALSE(vp9.temporal_up_switch);
|
| + }
|
| +
|
| + void VerifyTemporalLayerStructure1(const RTPVideoHeaderVP9& vp9) const {
|
| + EXPECT_NE(kNoTl0PicIdx, vp9.tl0_pic_idx);
|
| + EXPECT_EQ(0, vp9.temporal_idx); // 0,0,0,...
|
| + EXPECT_FALSE(vp9.temporal_up_switch);
|
| + }
|
| +
|
| + void VerifyTemporalLayerStructure2(const RTPVideoHeaderVP9& vp9) const {
|
| + EXPECT_NE(kNoTl0PicIdx, vp9.tl0_pic_idx);
|
| + EXPECT_GE(vp9.temporal_idx, 0); // 0,1,0,1,... (tid reset on I-frames).
|
| + EXPECT_LE(vp9.temporal_idx, 1);
|
| + EXPECT_EQ(vp9.temporal_idx > 0, vp9.temporal_up_switch);
|
| + if (IsNewPictureId(vp9)) {
|
| + uint8_t expected_tid =
|
| + (!vp9.inter_pic_predicted || last_vp9_.temporal_idx == 1) ? 0 : 1;
|
| + EXPECT_EQ(expected_tid, vp9.temporal_idx);
|
| }
|
| - } test;
|
| + }
|
|
|
| - RunBaseTest(&test, FakeNetworkPipe::Config());
|
| -}
|
| + void VerifyTemporalLayerStructure3(const RTPVideoHeaderVP9& vp9) const {
|
| + EXPECT_NE(kNoTl0PicIdx, vp9.tl0_pic_idx);
|
| + EXPECT_GE(vp9.temporal_idx, 0); // 0,2,1,2,... (tid reset on I-frames).
|
| + EXPECT_LE(vp9.temporal_idx, 2);
|
| + if (IsNewPictureId(vp9) && vp9.inter_pic_predicted) {
|
| + EXPECT_NE(vp9.temporal_idx, last_vp9_.temporal_idx);
|
| + switch (vp9.temporal_idx) {
|
| + case 0:
|
| + EXPECT_EQ(2, last_vp9_.temporal_idx);
|
| + EXPECT_FALSE(vp9.temporal_up_switch);
|
| + break;
|
| + case 1:
|
| + EXPECT_EQ(2, last_vp9_.temporal_idx);
|
| + EXPECT_TRUE(vp9.temporal_up_switch);
|
| + break;
|
| + case 2:
|
| + EXPECT_EQ(last_vp9_.temporal_idx == 0, vp9.temporal_up_switch);
|
| + break;
|
| + }
|
| + }
|
| + }
|
|
|
| -TEST_F(VideoSendStreamTest, VP9FlexModeHasPictureId) {
|
| - class FlexibleMode : public VP9HeaderObeserver {
|
| - void ModifyConfigsHook(
|
| - VideoSendStream::Config* send_config,
|
| - std::vector<VideoReceiveStream::Config>* receive_configs,
|
| - VideoEncoderConfig* encoder_config) override {
|
| - vp9_settings_.flexibleMode = true;
|
| + void VerifyTl0Idx(const RTPVideoHeaderVP9& vp9) const {
|
| + if (vp9.tl0_pic_idx == kNoTl0PicIdx)
|
| + return;
|
| +
|
| + uint8_t expected_tl0_idx = last_vp9_.tl0_pic_idx;
|
| + if (vp9.temporal_idx == 0)
|
| + ++expected_tl0_idx;
|
| + EXPECT_EQ(expected_tl0_idx, vp9.tl0_pic_idx);
|
| + }
|
| +
|
| + bool IsNewPictureId(const RTPVideoHeaderVP9& vp9) const {
|
| + return frames_sent_ > 0 && (vp9.picture_id != last_vp9_.picture_id);
|
| + }
|
| +
|
| + // Flexible mode (F=1): Non-flexible mode (F=0):
|
| + //
|
| + // +-+-+-+-+-+-+-+-+ +-+-+-+-+-+-+-+-+
|
| + // |I|P|L|F|B|E|V|-| |I|P|L|F|B|E|V|-|
|
| + // +-+-+-+-+-+-+-+-+ +-+-+-+-+-+-+-+-+
|
| + // I: |M| PICTURE ID | I: |M| PICTURE ID |
|
| + // +-+-+-+-+-+-+-+-+ +-+-+-+-+-+-+-+-+
|
| + // M: | EXTENDED PID | M: | EXTENDED PID |
|
| + // +-+-+-+-+-+-+-+-+ +-+-+-+-+-+-+-+-+
|
| + // L: | T |U| S |D| L: | T |U| S |D|
|
| + // +-+-+-+-+-+-+-+-+ +-+-+-+-+-+-+-+-+
|
| + // P,F: | P_DIFF |X|N| | TL0PICIDX |
|
| + // +-+-+-+-+-+-+-+-+ +-+-+-+-+-+-+-+-+
|
| + // X: |EXTENDED P_DIFF| V: | SS .. |
|
| + // +-+-+-+-+-+-+-+-+ +-+-+-+-+-+-+-+-+
|
| + // V: | SS .. |
|
| + // +-+-+-+-+-+-+-+-+
|
| + void VerifyCommonHeader(const RTPVideoHeaderVP9& vp9) const {
|
| + EXPECT_EQ(kMaxTwoBytePictureId, vp9.max_picture_id); // M:1
|
| + EXPECT_NE(kNoPictureId, vp9.picture_id); // I:1
|
| + EXPECT_EQ(vp9_settings_.flexibleMode, vp9.flexible_mode); // F
|
| + EXPECT_GE(vp9.spatial_idx, 0); // S
|
| + EXPECT_LT(vp9.spatial_idx, vp9_settings_.numberOfSpatialLayers);
|
| + if (vp9.ss_data_available) // V
|
| + VerifySsData(vp9);
|
| +
|
| + if (frames_sent_ == 0)
|
| + EXPECT_FALSE(vp9.inter_pic_predicted); // P
|
| +
|
| + if (!vp9.inter_pic_predicted) {
|
| + EXPECT_TRUE(vp9.temporal_idx == 0 || vp9.temporal_idx == kNoTemporalIdx);
|
| + EXPECT_FALSE(vp9.temporal_up_switch);
|
| }
|
| + }
|
|
|
| - void InspectHeader(RTPVideoHeaderVP9* vp9videoHeader) override {
|
| - EXPECT_NE(vp9videoHeader->picture_id, kNoPictureId);
|
| - observation_complete_->Set();
|
| + // Scalability structure (SS).
|
| + //
|
| + // +-+-+-+-+-+-+-+-+
|
| + // V: | N_S |Y|G|-|-|-|
|
| + // +-+-+-+-+-+-+-+-+
|
| + // Y: | WIDTH | N_S + 1 times
|
| + // +-+-+-+-+-+-+-+-+
|
| + // | HEIGHT |
|
| + // +-+-+-+-+-+-+-+-+
|
| + // G: | N_G |
|
| + // +-+-+-+-+-+-+-+-+
|
| + // N_G: | T |U| R |-|-| N_G times
|
| + // +-+-+-+-+-+-+-+-+
|
| + // | P_DIFF | R times
|
| + // +-+-+-+-+-+-+-+-+
|
| + void VerifySsData(const RTPVideoHeaderVP9& vp9) const {
|
| + EXPECT_TRUE(vp9.ss_data_available); // V
|
| + EXPECT_EQ(vp9_settings_.numberOfSpatialLayers, // N_S + 1
|
| + vp9.num_spatial_layers);
|
| + EXPECT_TRUE(vp9.spatial_layer_resolution_present); // Y:1
|
| + size_t expected_width = encoder_config_.streams[0].width;
|
| + size_t expected_height = encoder_config_.streams[0].height;
|
| + for (int i = vp9.num_spatial_layers - 1; i >= 0; --i) {
|
| + EXPECT_EQ(expected_width, vp9.width[i]); // WIDTH
|
| + EXPECT_EQ(expected_height, vp9.height[i]); // HEIGHT
|
| + expected_width /= 2;
|
| + expected_height /= 2;
|
| }
|
| - } test;
|
| + }
|
|
|
| - RunBaseTest(&test, FakeNetworkPipe::Config());
|
| + void CompareConsecutiveFrames(const RTPHeader& header,
|
| + const RTPVideoHeader& video) const {
|
| + const RTPVideoHeaderVP9& vp9 = video.codecHeader.VP9;
|
| +
|
| + bool new_frame = packets_sent_ == 0 ||
|
| + IsNewerTimestamp(header.timestamp, last_header_.timestamp);
|
| + EXPECT_EQ(new_frame, video.isFirstPacket);
|
| + if (!new_frame) {
|
| + EXPECT_FALSE(last_header_.markerBit);
|
| + EXPECT_EQ(last_header_.timestamp, header.timestamp);
|
| + EXPECT_EQ(last_vp9_.picture_id, vp9.picture_id);
|
| + EXPECT_EQ(last_vp9_.temporal_idx, vp9.temporal_idx);
|
| + EXPECT_EQ(last_vp9_.tl0_pic_idx, vp9.tl0_pic_idx);
|
| + VerifySpatialIdxWithinFrame(vp9);
|
| + return;
|
| + }
|
| + // New frame.
|
| + EXPECT_TRUE(vp9.beginning_of_frame);
|
| +
|
| + // Compare with last packet in previous frame.
|
| + if (frames_sent_ == 0)
|
| + return;
|
| + EXPECT_TRUE(last_vp9_.end_of_frame);
|
| + EXPECT_TRUE(last_header_.markerBit);
|
| + EXPECT_TRUE(ContinuousPictureId(vp9));
|
| + VerifyTl0Idx(vp9);
|
| + }
|
| +
|
| + rtc::scoped_ptr<VP9Encoder> vp9_encoder_;
|
| + VideoCodecVP9 vp9_settings_;
|
| + webrtc::VideoEncoderConfig encoder_config_;
|
| + RTPHeader last_header_;
|
| + RTPVideoHeaderVP9 last_vp9_;
|
| + size_t packets_sent_;
|
| + size_t frames_sent_;
|
| +};
|
| +
|
| +TEST_F(VideoSendStreamTest, Vp9NonFlexMode_1Tl1SLayers) {
|
| + const uint8_t kNumTemporalLayers = 1;
|
| + const uint8_t kNumSpatialLayers = 1;
|
| + TestVp9NonFlexMode(kNumTemporalLayers, kNumSpatialLayers);
|
| }
|
|
|
| -TEST_F(VideoSendStreamTest, VP9FlexModeRefCount) {
|
| - class FlexibleMode : public VP9HeaderObeserver {
|
| +TEST_F(VideoSendStreamTest, Vp9NonFlexMode_2Tl1SLayers) {
|
| + const uint8_t kNumTemporalLayers = 2;
|
| + const uint8_t kNumSpatialLayers = 1;
|
| + TestVp9NonFlexMode(kNumTemporalLayers, kNumSpatialLayers);
|
| +}
|
| +
|
| +TEST_F(VideoSendStreamTest, Vp9NonFlexMode_3Tl1SLayers) {
|
| + const uint8_t kNumTemporalLayers = 3;
|
| + const uint8_t kNumSpatialLayers = 1;
|
| + TestVp9NonFlexMode(kNumTemporalLayers, kNumSpatialLayers);
|
| +}
|
| +
|
| +TEST_F(VideoSendStreamTest, Vp9NonFlexMode_1Tl2SLayers) {
|
| + const uint8_t kNumTemporalLayers = 1;
|
| + const uint8_t kNumSpatialLayers = 2;
|
| + TestVp9NonFlexMode(kNumTemporalLayers, kNumSpatialLayers);
|
| +}
|
| +
|
| +TEST_F(VideoSendStreamTest, Vp9NonFlexMode_2Tl2SLayers) {
|
| + const uint8_t kNumTemporalLayers = 2;
|
| + const uint8_t kNumSpatialLayers = 2;
|
| + TestVp9NonFlexMode(kNumTemporalLayers, kNumSpatialLayers);
|
| +}
|
| +
|
| +TEST_F(VideoSendStreamTest, Vp9NonFlexMode_3Tl2SLayers) {
|
| + const uint8_t kNumTemporalLayers = 3;
|
| + const uint8_t kNumSpatialLayers = 2;
|
| + TestVp9NonFlexMode(kNumTemporalLayers, kNumSpatialLayers);
|
| +}
|
| +
|
| +void VideoSendStreamTest::TestVp9NonFlexMode(uint8_t num_temporal_layers,
|
| + uint8_t num_spatial_layers) {
|
| + static const size_t kNumFramesToSend = 100;
|
| + // Set to < kNumFramesToSend and coprime to length of temporal layer
|
| + // structures to verify temporal id reset on key frame.
|
| + static const int kKeyFrameInterval = 31;
|
| + class NonFlexibleMode : public Vp9HeaderObserver {
|
| + public:
|
| + NonFlexibleMode(uint8_t num_temporal_layers, uint8_t num_spatial_layers)
|
| + : num_temporal_layers_(num_temporal_layers),
|
| + num_spatial_layers_(num_spatial_layers),
|
| + l_field_(num_temporal_layers > 1 || num_spatial_layers > 1) {}
|
| void ModifyConfigsHook(
|
| VideoSendStream::Config* send_config,
|
| std::vector<VideoReceiveStream::Config>* receive_configs,
|
| VideoEncoderConfig* encoder_config) override {
|
| - vp9_settings_.flexibleMode = true;
|
| - }
|
| + vp9_settings_.flexibleMode = false;
|
| + vp9_settings_.frameDroppingOn = false;
|
| + vp9_settings_.keyFrameInterval = kKeyFrameInterval;
|
| + vp9_settings_.numberOfTemporalLayers = num_temporal_layers_;
|
| + vp9_settings_.numberOfSpatialLayers = num_spatial_layers_;
|
| + }
|
| +
|
| + void InspectHeader(const RTPVideoHeaderVP9& vp9) override {
|
| + bool ss_data_expected = !vp9.inter_pic_predicted &&
|
| + vp9.beginning_of_frame && vp9.spatial_idx == 0;
|
| + EXPECT_EQ(ss_data_expected, vp9.ss_data_available);
|
| + EXPECT_EQ(vp9.spatial_idx > 0, vp9.inter_layer_predicted); // D
|
| +
|
| + if (IsNewPictureId(vp9)) {
|
| + EXPECT_EQ(0, vp9.spatial_idx);
|
| + EXPECT_EQ(num_spatial_layers_ - 1, last_vp9_.spatial_idx);
|
| + }
|
| +
|
| + VerifyFixedTemporalLayerStructure(vp9,
|
| + l_field_ ? num_temporal_layers_ : 0);
|
|
|
| - void InspectHeader(RTPVideoHeaderVP9* vp9videoHeader) override {
|
| - EXPECT_TRUE(vp9videoHeader->flexible_mode);
|
| - if (vp9videoHeader->inter_pic_predicted) {
|
| - EXPECT_GT(vp9videoHeader->num_ref_pics, 0u);
|
| + if (frames_sent_ > kNumFramesToSend)
|
| observation_complete_->Set();
|
| - }
|
| }
|
| - } test;
|
| + const uint8_t num_temporal_layers_;
|
| + const uint8_t num_spatial_layers_;
|
| + const bool l_field_;
|
| + } test(num_temporal_layers, num_spatial_layers);
|
|
|
| RunBaseTest(&test, FakeNetworkPipe::Config());
|
| }
|
|
|
| -TEST_F(VideoSendStreamTest, VP9FlexModeRefs) {
|
| - class FlexibleMode : public VP9HeaderObeserver {
|
| +TEST_F(VideoSendStreamTest, Vp9FlexModeRefCount) {
|
| + class FlexibleMode : public Vp9HeaderObserver {
|
| void ModifyConfigsHook(
|
| VideoSendStream::Config* send_config,
|
| std::vector<VideoReceiveStream::Config>* receive_configs,
|
| VideoEncoderConfig* encoder_config) override {
|
| + encoder_config->content_type = VideoEncoderConfig::ContentType::kScreen;
|
| vp9_settings_.flexibleMode = true;
|
| + vp9_settings_.numberOfTemporalLayers = 1;
|
| + vp9_settings_.numberOfSpatialLayers = 2;
|
| }
|
|
|
| - void InspectHeader(RTPVideoHeaderVP9* vp9videoHeader) override {
|
| - EXPECT_TRUE(vp9videoHeader->flexible_mode);
|
| - if (vp9videoHeader->inter_pic_predicted) {
|
| - EXPECT_GT(vp9videoHeader->num_ref_pics, 0u);
|
| + void InspectHeader(const RTPVideoHeaderVP9& vp9_header) override {
|
| + EXPECT_TRUE(vp9_header.flexible_mode);
|
| + EXPECT_EQ(kNoTl0PicIdx, vp9_header.tl0_pic_idx);
|
| + if (vp9_header.inter_pic_predicted) {
|
| + EXPECT_GT(vp9_header.num_ref_pics, 0u);
|
| observation_complete_->Set();
|
| }
|
| }
|
| -
|
| } test;
|
|
|
| RunBaseTest(&test, FakeNetworkPipe::Config());
|
|
|