Index: webrtc/modules/video_coding/codecs/test/videoprocessor.cc |
diff --git a/webrtc/modules/video_coding/codecs/test/videoprocessor.cc b/webrtc/modules/video_coding/codecs/test/videoprocessor.cc |
index 99a34fc60ee6446781a4e9924b40b17f110c9bf0..c7c152d16137559291d1bf9cdedc284d6d55ddf9 100644 |
--- a/webrtc/modules/video_coding/codecs/test/videoprocessor.cc |
+++ b/webrtc/modules/video_coding/codecs/test/videoprocessor.cc |
@@ -29,6 +29,10 @@ |
namespace webrtc { |
namespace test { |
+namespace { |
+const int k90khzTimestampFrameDiff = 3000; // Assuming 30 fps. |
+} // namespace |
+ |
const char* ExcludeFrameTypesToStr(ExcludeFrameTypes e) { |
switch (e) { |
case kExcludeOnlyFirstKeyFrame: |
@@ -60,18 +64,24 @@ TestConfig::~TestConfig() {} |
VideoProcessorImpl::VideoProcessorImpl(webrtc::VideoEncoder* encoder, |
webrtc::VideoDecoder* decoder, |
- FrameReader* frame_reader, |
- FrameWriter* frame_writer, |
+ FrameReader* analysis_frame_reader, |
+ FrameWriter* analysis_frame_writer, |
PacketManipulator* packet_manipulator, |
const TestConfig& config, |
- Stats* stats) |
+ Stats* stats, |
+ FrameWriter* source_frame_writer, |
+ IvfFileWriter* encoded_frame_writer, |
+ FrameWriter* decoded_frame_writer) |
: encoder_(encoder), |
decoder_(decoder), |
- frame_reader_(frame_reader), |
- frame_writer_(frame_writer), |
+ analysis_frame_reader_(analysis_frame_reader), |
+ analysis_frame_writer_(analysis_frame_writer), |
packet_manipulator_(packet_manipulator), |
config_(config), |
stats_(stats), |
+ source_frame_writer_(source_frame_writer), |
+ encoded_frame_writer_(encoded_frame_writer), |
+ decoded_frame_writer_(decoded_frame_writer), |
first_key_frame_has_been_excluded_(false), |
last_frame_missing_(false), |
initialized_(false), |
@@ -94,8 +104,8 @@ VideoProcessorImpl::VideoProcessorImpl(webrtc::VideoEncoder* encoder, |
*config.codec_settings, std::move(tl_factory)); |
RTC_DCHECK(encoder); |
RTC_DCHECK(decoder); |
- RTC_DCHECK(frame_reader); |
- RTC_DCHECK(frame_writer); |
+ RTC_DCHECK(analysis_frame_reader); |
+ RTC_DCHECK(analysis_frame_writer); |
RTC_DCHECK(packet_manipulator); |
RTC_DCHECK(stats); |
} |
@@ -105,7 +115,7 @@ bool VideoProcessorImpl::Init() { |
bit_rate_factor_ = config_.codec_settings->maxFramerate * 0.001 * 8; // bits |
// Initialize data structures used by the encoder/decoder APIs. |
- size_t frame_length_in_bytes = frame_reader_->FrameLength(); |
+ size_t frame_length_in_bytes = analysis_frame_reader_->FrameLength(); |
last_successful_frame_buffer_.reset(new uint8_t[frame_length_in_bytes]); |
// Set fixed properties common for all frames. |
@@ -139,7 +149,8 @@ bool VideoProcessorImpl::Init() { |
if (config_.verbose) { |
printf("Video Processor:\n"); |
printf(" #CPU cores used : %d\n", num_cores); |
- printf(" Total # of frames: %d\n", frame_reader_->NumberOfFrames()); |
+ printf(" Total # of frames: %d\n", |
+ analysis_frame_reader_->NumberOfFrames()); |
printf(" Codec settings:\n"); |
printf(" Start bitrate : %d kbps\n", |
config_.codec_settings->startBitrate); |
@@ -202,18 +213,29 @@ int VideoProcessorImpl::NumberSpatialResizes() { |
} |
bool VideoProcessorImpl::ProcessFrame(int frame_number) { |
- RTC_DCHECK_GE(frame_number, 0); |
+ RTC_CHECK_GE(frame_number, 0); |
RTC_CHECK(initialized_) << "Attempting to use uninitialized VideoProcessor"; |
- // |prev_time_stamp_| is used for getting number of dropped frames. |
- if (frame_number == 0) { |
- prev_time_stamp_ = -1; |
- } |
- |
- rtc::scoped_refptr<VideoFrameBuffer> buffer(frame_reader_->ReadFrame()); |
+ rtc::scoped_refptr<VideoFrameBuffer> buffer( |
+ analysis_frame_reader_->ReadFrame()); |
if (buffer) { |
- // Use the frame number as "timestamp" to identify frames. |
- VideoFrame source_frame(buffer, frame_number, 0, webrtc::kVideoRotation_0); |
+ if (source_frame_writer_) { |
+ // TODO(brandtr): Introduce temp buffer as data member, to avoid |
+ // allocating for every frame. |
+ size_t length = CalcBufferSize(kI420, buffer->width(), buffer->height()); |
+ std::unique_ptr<uint8_t[]> extracted_buffer(new uint8_t[length]); |
+ int extracted_length = |
+ ExtractBuffer(buffer, length, extracted_buffer.get()); |
+ RTC_CHECK_EQ(extracted_length, source_frame_writer_->FrameLength()); |
+ source_frame_writer_->WriteFrame(extracted_buffer.get()); |
+ } |
+ |
+ // Use the frame number as basis for timestamp to identify frames. Let the |
+ // first timestamp be non-zero, to not make the IvfFileWriter believe that |
+ // we want to use capture timestamps in the IVF files. |
+ VideoFrame source_frame(buffer, |
+ (frame_number + 1) * k90khzTimestampFrameDiff, 0, |
+ webrtc::kVideoRotation_0); |
// Ensure we have a new statistics data object we can fill. |
FrameStatistic& stat = stats_->NewFrame(frame_number); |
@@ -256,30 +278,41 @@ void VideoProcessorImpl::FrameEncoded( |
// time recordings should wrap the Encode call as tightly as possible. |
int64_t encode_stop_ns = rtc::TimeNanos(); |
- // Timestamp is frame number, so this gives us #dropped frames. |
+ if (encoded_frame_writer_) { |
+ RTC_CHECK(encoded_frame_writer_->WriteFrame(encoded_image, codec)); |
+ } |
+ |
+ // Timestamp is proportional to frame number, so this gives us number of |
+ // dropped frames. |
int num_dropped_from_prev_encode = |
- encoded_image._timeStamp - prev_time_stamp_ - 1; |
+ (encoded_image._timeStamp - prev_time_stamp_) / k90khzTimestampFrameDiff - |
+ 1; |
num_dropped_frames_ += num_dropped_from_prev_encode; |
prev_time_stamp_ = encoded_image._timeStamp; |
if (num_dropped_from_prev_encode > 0) { |
// For dropped frames, we write out the last decoded frame to avoid getting |
// out of sync for the computation of PSNR and SSIM. |
for (int i = 0; i < num_dropped_from_prev_encode; i++) { |
- frame_writer_->WriteFrame(last_successful_frame_buffer_.get()); |
+ RTC_CHECK(analysis_frame_writer_->WriteFrame( |
+ last_successful_frame_buffer_.get())); |
+ if (decoded_frame_writer_) { |
+ RTC_CHECK(decoded_frame_writer_->WriteFrame( |
+ last_successful_frame_buffer_.get())); |
+ } |
} |
} |
+ |
// Frame is not dropped, so update the encoded frame size |
// (encoder callback is only called for non-zero length frames). |
encoded_frame_size_ = encoded_image._length; |
encoded_frame_type_ = encoded_image._frameType; |
- int frame_number = encoded_image._timeStamp; |
- |
+ int frame_number = encoded_image._timeStamp / k90khzTimestampFrameDiff - 1; |
FrameStatistic& stat = stats_->stats_[frame_number]; |
stat.encode_time_in_us = |
GetElapsedTimeMicroseconds(encode_start_ns_, encode_stop_ns); |
stat.encoding_successful = true; |
stat.encoded_frame_length_in_bytes = encoded_image._length; |
- stat.frame_number = encoded_image._timeStamp; |
+ stat.frame_number = frame_number; |
stat.frame_type = encoded_image._frameType; |
stat.bit_rate_in_kbps = encoded_image._length * bit_rate_factor_; |
stat.total_packets = |
@@ -336,7 +369,12 @@ void VideoProcessorImpl::FrameEncoded( |
if (decode_result != WEBRTC_VIDEO_CODEC_OK) { |
// Write the last successful frame the output file to avoid getting it out |
// of sync with the source file for SSIM and PSNR comparisons. |
- frame_writer_->WriteFrame(last_successful_frame_buffer_.get()); |
+ RTC_CHECK(analysis_frame_writer_->WriteFrame( |
+ last_successful_frame_buffer_.get())); |
+ if (decoded_frame_writer_) { |
+ RTC_CHECK(decoded_frame_writer_->WriteFrame( |
+ last_successful_frame_buffer_.get())); |
+ } |
} |
// Save status for losses so we can inform the decoder for the next frame. |
@@ -349,7 +387,7 @@ void VideoProcessorImpl::FrameDecoded(const VideoFrame& image) { |
int64_t decode_stop_ns = rtc::TimeNanos(); |
// Report stats. |
- int frame_number = image.timestamp(); |
+ int frame_number = image.timestamp() / k90khzTimestampFrameDiff - 1; |
FrameStatistic& stat = stats_->stats_[frame_number]; |
stat.decode_time_in_us = |
GetElapsedTimeMicroseconds(decode_start_ns_, decode_stop_ns); |
@@ -383,14 +421,14 @@ void VideoProcessorImpl::FrameDecoded(const VideoFrame& image) { |
CalcBufferSize(kI420, up_image->width(), up_image->height()); |
std::unique_ptr<uint8_t[]> image_buffer(new uint8_t[length]); |
int extracted_length = ExtractBuffer(up_image, length, image_buffer.get()); |
- RTC_DCHECK_GT(extracted_length, 0); |
+ RTC_CHECK_GT(extracted_length, 0); |
// Update our copy of the last successful frame. |
memcpy(last_successful_frame_buffer_.get(), image_buffer.get(), |
extracted_length); |
- bool write_success = frame_writer_->WriteFrame(image_buffer.get()); |
- RTC_DCHECK(write_success); |
- if (!write_success) { |
- fprintf(stderr, "Failed to write frame %d to disk!", frame_number); |
+ |
+ RTC_CHECK(analysis_frame_writer_->WriteFrame(image_buffer.get())); |
+ if (decoded_frame_writer_) { |
+ RTC_CHECK(decoded_frame_writer_->WriteFrame(image_buffer.get())); |
} |
} else { // No resize. |
// Update our copy of the last successful frame. |
@@ -406,14 +444,13 @@ void VideoProcessorImpl::FrameDecoded(const VideoFrame& image) { |
extracted_length = |
ExtractBuffer(image.video_frame_buffer(), length, image_buffer.get()); |
} |
- RTC_DCHECK_GT(extracted_length, 0); |
+ RTC_CHECK_GT(extracted_length, 0); |
memcpy(last_successful_frame_buffer_.get(), image_buffer.get(), |
extracted_length); |
- bool write_success = frame_writer_->WriteFrame(image_buffer.get()); |
- RTC_DCHECK(write_success); |
- if (!write_success) { |
- fprintf(stderr, "Failed to write frame %d to disk!", frame_number); |
+ RTC_CHECK(analysis_frame_writer_->WriteFrame(image_buffer.get())); |
+ if (decoded_frame_writer_) { |
+ RTC_CHECK(decoded_frame_writer_->WriteFrame(image_buffer.get())); |
} |
} |
} |