Index: webrtc/sdk/objc/Framework/Classes/VideoToolbox/RTCVideoDecoderH264.mm |
diff --git a/webrtc/sdk/objc/Framework/Classes/VideoToolbox/decoder.mm b/webrtc/sdk/objc/Framework/Classes/VideoToolbox/RTCVideoDecoderH264.mm |
similarity index 38% |
rename from webrtc/sdk/objc/Framework/Classes/VideoToolbox/decoder.mm |
rename to webrtc/sdk/objc/Framework/Classes/VideoToolbox/RTCVideoDecoderH264.mm |
index a8da8091d9f9a67cd90852c4df077870e0d51781..f657f9c2f12b68b1d1e95adf6370f418fb2655de 100644 |
--- a/webrtc/sdk/objc/Framework/Classes/VideoToolbox/decoder.mm |
+++ b/webrtc/sdk/objc/Framework/Classes/VideoToolbox/RTCVideoDecoderH264.mm |
@@ -9,93 +9,80 @@ |
* |
*/ |
-#include "webrtc/sdk/objc/Framework/Classes/VideoToolbox/decoder.h" |
+#import "WebRTC/RTCVideoCodecH264.h" |
-#include <memory> |
+#import <VideoToolbox/VideoToolbox.h> |
-#include "libyuv/convert.h" |
-#include "webrtc/api/video/video_frame.h" |
-#include "webrtc/common_video/include/video_frame.h" |
+#include "webrtc/modules/video_coding/include/video_error_codes.h" |
#include "webrtc/rtc_base/checks.h" |
#include "webrtc/rtc_base/logging.h" |
-#include "webrtc/sdk/objc/Framework/Classes/Video/objc_frame_buffer.h" |
+#include "webrtc/rtc_base/timeutils.h" |
#include "webrtc/sdk/objc/Framework/Classes/VideoToolbox/nalu_rewriter.h" |
+#import "WebRTC/RTCVideoFrame.h" |
#import "WebRTC/RTCVideoFrameBuffer.h" |
+#import "helpers.h" |
#if defined(WEBRTC_IOS) |
#import "Common/RTCUIApplicationStatusObserver.h" |
#endif |
-namespace webrtc { |
-namespace { |
- |
-static const int64_t kMsPerSec = 1000; |
- |
-// Convenience function for creating a dictionary. |
-inline CFDictionaryRef CreateCFDictionary(CFTypeRef* keys, |
- CFTypeRef* values, |
- size_t size) { |
- return CFDictionaryCreate(nullptr, keys, values, size, |
- &kCFTypeDictionaryKeyCallBacks, |
- &kCFTypeDictionaryValueCallBacks); |
-} |
- |
// Struct that we pass to the decoder per frame to decode. We receive it again |
// in the decoder callback. |
-struct FrameDecodeParams { |
- FrameDecodeParams(DecodedImageCallback* cb, int64_t ts) |
- : callback(cb), timestamp(ts) {} |
- DecodedImageCallback* callback; |
+struct RTCFrameDecodeParams { |
+ RTCFrameDecodeParams(RTCVideoDecoderCallback cb, int64_t ts) : callback(cb), timestamp(ts) {} |
+ RTCVideoDecoderCallback callback; |
int64_t timestamp; |
}; |
// This is the callback function that VideoToolbox calls when decode is |
// complete. |
-void VTDecompressionOutputCallback(void* decoder, |
- void* params, |
- OSStatus status, |
- VTDecodeInfoFlags info_flags, |
- CVImageBufferRef image_buffer, |
- CMTime timestamp, |
- CMTime duration) { |
- std::unique_ptr<FrameDecodeParams> decode_params( |
- reinterpret_cast<FrameDecodeParams*>(params)); |
+void decompressionOutputCallback(void *decoder, |
+ void *params, |
+ OSStatus status, |
+ VTDecodeInfoFlags infoFlags, |
+ CVImageBufferRef imageBuffer, |
+ CMTime timestamp, |
+ CMTime duration) { |
+ std::unique_ptr<RTCFrameDecodeParams> decodeParams( |
+ reinterpret_cast<RTCFrameDecodeParams *>(params)); |
if (status != noErr) { |
LOG(LS_ERROR) << "Failed to decode frame. Status: " << status; |
return; |
} |
// TODO(tkchin): Handle CVO properly. |
- rtc::scoped_refptr<VideoFrameBuffer> buffer = new rtc::RefCountedObject<ObjCFrameBuffer>( |
- [[RTCCVPixelBuffer alloc] initWithPixelBuffer:image_buffer]); |
- VideoFrame decoded_frame(buffer, decode_params->timestamp, |
- CMTimeGetSeconds(timestamp) * kMsPerSec, |
- kVideoRotation_0); |
- decode_params->callback->Decoded(decoded_frame); |
+ RTCCVPixelBuffer *frameBuffer = [[RTCCVPixelBuffer alloc] initWithPixelBuffer:imageBuffer]; |
+ RTCVideoFrame *decodedFrame = |
+ [[RTCVideoFrame alloc] initWithBuffer:frameBuffer |
+ rotation:RTCVideoRotation_0 |
+ timeStampNs:CMTimeGetSeconds(timestamp) * rtc::kNumNanosecsPerSec]; |
+ decodedFrame.timeStamp = decodeParams->timestamp; |
+ decodeParams->callback(decodedFrame); |
} |
-} // namespace |
- |
-H264VideoToolboxDecoder::H264VideoToolboxDecoder() |
- : callback_(nullptr), video_format_(nullptr), decompression_session_(nullptr) {} |
+// Decoder. |
+@implementation RTCVideoDecoderH264 { |
+ CMVideoFormatDescriptionRef _videoFormat; |
+ VTDecompressionSessionRef _decompressionSession; |
+ RTCVideoDecoderCallback _callback; |
+} |
-H264VideoToolboxDecoder::~H264VideoToolboxDecoder() { |
- DestroyDecompressionSession(); |
- SetVideoFormat(nullptr); |
+- (void)dealloc { |
+ [self destroyDecompressionSession]; |
+ [self setVideoFormat:nullptr]; |
} |
-int H264VideoToolboxDecoder::InitDecode(const VideoCodec* video_codec, |
- int number_of_cores) { |
+- (NSInteger)startDecodeWithSettings:(RTCVideoEncoderSettings *)settings |
+ numberOfCores:(int)numberOfCores { |
return WEBRTC_VIDEO_CODEC_OK; |
} |
-int H264VideoToolboxDecoder::Decode( |
- const EncodedImage& input_image, |
- bool missing_frames, |
- const RTPFragmentationHeader* fragmentation, |
- const CodecSpecificInfo* codec_specific_info, |
- int64_t render_time_ms) { |
- RTC_DCHECK(input_image._buffer); |
+- (NSInteger)decode:(RTCEncodedImage *)inputImage |
+ missingFrames:(BOOL)missingFrames |
+ fragmentationHeader:(RTCRtpFragmentationHeader *)fragmentationHeader |
+ codecSpecificInfo:(__nullable id<RTCCodecSpecificInfo>)info |
+ renderTimeMs:(int64_t)renderTimeMs { |
+ RTC_DCHECK(inputImage.buffer); |
#if defined(WEBRTC_IOS) |
if (![[RTCUIApplicationStatusObserver sharedInstance] isApplicationActive]) { |
@@ -103,26 +90,26 @@ int H264VideoToolboxDecoder::Decode( |
// hardware decoder has been invalidated by the OS. |
// Reset video format so that we won't process frames until the next |
// keyframe. |
- SetVideoFormat(nullptr); |
+ [self setVideoFormat:nullptr]; |
return WEBRTC_VIDEO_CODEC_NO_OUTPUT; |
} |
#endif |
- CMVideoFormatDescriptionRef input_format = nullptr; |
- if (H264AnnexBBufferHasVideoFormatDescription(input_image._buffer, |
- input_image._length)) { |
- input_format = CreateVideoFormatDescription(input_image._buffer, |
- input_image._length); |
- if (input_format) { |
+ CMVideoFormatDescriptionRef inputFormat = nullptr; |
+ if (webrtc::H264AnnexBBufferHasVideoFormatDescription((uint8_t *)inputImage.buffer.bytes, |
+ inputImage.buffer.length)) { |
+ inputFormat = webrtc::CreateVideoFormatDescription((uint8_t *)inputImage.buffer.bytes, |
+ inputImage.buffer.length); |
+ if (inputFormat) { |
// Check if the video format has changed, and reinitialize decoder if |
// needed. |
- if (!CMFormatDescriptionEqual(input_format, video_format_)) { |
- SetVideoFormat(input_format); |
- ResetDecompressionSession(); |
+ if (!CMFormatDescriptionEqual(inputFormat, _videoFormat)) { |
+ [self setVideoFormat:inputFormat]; |
+ [self resetDecompressionSession]; |
} |
- CFRelease(input_format); |
+ CFRelease(inputFormat); |
} |
} |
- if (!video_format_) { |
+ if (!_videoFormat) { |
// We received a frame but we don't have format information so we can't |
// decode it. |
// This can happen after backgrounding. We need to wait for the next |
@@ -131,34 +118,29 @@ int H264VideoToolboxDecoder::Decode( |
LOG(LS_WARNING) << "Missing video format. Frame with sps/pps required."; |
return WEBRTC_VIDEO_CODEC_ERROR; |
} |
- CMSampleBufferRef sample_buffer = nullptr; |
- if (!H264AnnexBBufferToCMSampleBuffer(input_image._buffer, |
- input_image._length, video_format_, |
- &sample_buffer)) { |
+ CMSampleBufferRef sampleBuffer = nullptr; |
+ if (!webrtc::H264AnnexBBufferToCMSampleBuffer((uint8_t *)inputImage.buffer.bytes, |
+ inputImage.buffer.length, |
+ _videoFormat, |
+ &sampleBuffer)) { |
return WEBRTC_VIDEO_CODEC_ERROR; |
} |
- RTC_DCHECK(sample_buffer); |
- VTDecodeFrameFlags decode_flags = |
- kVTDecodeFrame_EnableAsynchronousDecompression; |
- std::unique_ptr<FrameDecodeParams> frame_decode_params; |
- frame_decode_params.reset( |
- new FrameDecodeParams(callback_, input_image._timeStamp)); |
+ RTC_DCHECK(sampleBuffer); |
+ VTDecodeFrameFlags decodeFlags = kVTDecodeFrame_EnableAsynchronousDecompression; |
+ std::unique_ptr<RTCFrameDecodeParams> frameDecodeParams; |
+ frameDecodeParams.reset(new RTCFrameDecodeParams(_callback, inputImage.timeStamp)); |
OSStatus status = VTDecompressionSessionDecodeFrame( |
- decompression_session_, sample_buffer, decode_flags, |
- frame_decode_params.release(), nullptr); |
+ _decompressionSession, sampleBuffer, decodeFlags, frameDecodeParams.release(), nullptr); |
#if defined(WEBRTC_IOS) |
// Re-initialize the decoder if we have an invalid session while the app is |
// active and retry the decode request. |
- if (status == kVTInvalidSessionErr && |
- ResetDecompressionSession() == WEBRTC_VIDEO_CODEC_OK) { |
- frame_decode_params.reset( |
- new FrameDecodeParams(callback_, input_image._timeStamp)); |
+ if (status == kVTInvalidSessionErr && [self resetDecompressionSession] == WEBRTC_VIDEO_CODEC_OK) { |
+ frameDecodeParams.reset(new RTCFrameDecodeParams(_callback, inputImage.timeStamp)); |
status = VTDecompressionSessionDecodeFrame( |
- decompression_session_, sample_buffer, decode_flags, |
- frame_decode_params.release(), nullptr); |
+ _decompressionSession, sampleBuffer, decodeFlags, frameDecodeParams.release(), nullptr); |
} |
#endif |
- CFRelease(sample_buffer); |
+ CFRelease(sampleBuffer); |
if (status != noErr) { |
LOG(LS_ERROR) << "Failed to decode frame with code: " << status; |
return WEBRTC_VIDEO_CODEC_ERROR; |
@@ -166,27 +148,26 @@ int H264VideoToolboxDecoder::Decode( |
return WEBRTC_VIDEO_CODEC_OK; |
} |
-int H264VideoToolboxDecoder::RegisterDecodeCompleteCallback( |
- DecodedImageCallback* callback) { |
- RTC_DCHECK(!callback_); |
- callback_ = callback; |
- return WEBRTC_VIDEO_CODEC_OK; |
+- (void)setCallback:(RTCVideoDecoderCallback)callback { |
+ _callback = callback; |
} |
-int H264VideoToolboxDecoder::Release() { |
+- (NSInteger)releaseDecoder { |
// Need to invalidate the session so that callbacks no longer occur and it |
// is safe to null out the callback. |
- DestroyDecompressionSession(); |
- SetVideoFormat(nullptr); |
- callback_ = nullptr; |
+ [self destroyDecompressionSession]; |
+ [self setVideoFormat:nullptr]; |
+ _callback = nullptr; |
return WEBRTC_VIDEO_CODEC_OK; |
} |
-int H264VideoToolboxDecoder::ResetDecompressionSession() { |
- DestroyDecompressionSession(); |
+#pragma mark - Private |
+ |
+- (int)resetDecompressionSession { |
+ [self destroyDecompressionSession]; |
// Need to wait for the first SPS to initialize decoder. |
- if (!video_format_) { |
+ if (!_videoFormat) { |
return WEBRTC_VIDEO_CODEC_OK; |
} |
@@ -198,8 +179,8 @@ int H264VideoToolboxDecoder::ResetDecompressionSession() { |
// CVPixelBuffers directly to the renderer. |
// TODO(tkchin): Maybe only set OpenGL/IOSurface keys if we know that that |
// we can pass CVPixelBuffers as native handles in decoder output. |
- static size_t const attributes_size = 3; |
- CFTypeRef keys[attributes_size] = { |
+ static size_t const attributesSize = 3; |
+ CFTypeRef keys[attributesSize] = { |
#if defined(WEBRTC_IOS) |
kCVPixelBufferOpenGLESCompatibilityKey, |
#elif defined(WEBRTC_MAC) |
@@ -208,70 +189,64 @@ int H264VideoToolboxDecoder::ResetDecompressionSession() { |
kCVPixelBufferIOSurfacePropertiesKey, |
kCVPixelBufferPixelFormatTypeKey |
}; |
- CFDictionaryRef io_surface_value = CreateCFDictionary(nullptr, nullptr, 0); |
+ CFDictionaryRef ioSurfaceValue = CreateCFTypeDictionary(nullptr, nullptr, 0); |
int64_t nv12type = kCVPixelFormatType_420YpCbCr8BiPlanarFullRange; |
- CFNumberRef pixel_format = |
- CFNumberCreate(nullptr, kCFNumberLongType, &nv12type); |
- CFTypeRef values[attributes_size] = {kCFBooleanTrue, io_surface_value, |
- pixel_format}; |
- CFDictionaryRef attributes = |
- CreateCFDictionary(keys, values, attributes_size); |
- if (io_surface_value) { |
- CFRelease(io_surface_value); |
- io_surface_value = nullptr; |
+ CFNumberRef pixelFormat = CFNumberCreate(nullptr, kCFNumberLongType, &nv12type); |
+ CFTypeRef values[attributesSize] = {kCFBooleanTrue, ioSurfaceValue, pixelFormat}; |
+ CFDictionaryRef attributes = CreateCFTypeDictionary(keys, values, attributesSize); |
+ if (ioSurfaceValue) { |
+ CFRelease(ioSurfaceValue); |
+ ioSurfaceValue = nullptr; |
} |
- if (pixel_format) { |
- CFRelease(pixel_format); |
- pixel_format = nullptr; |
+ if (pixelFormat) { |
+ CFRelease(pixelFormat); |
+ pixelFormat = nullptr; |
} |
VTDecompressionOutputCallbackRecord record = { |
- VTDecompressionOutputCallback, this, |
+ decompressionOutputCallback, nullptr, |
}; |
- OSStatus status = |
- VTDecompressionSessionCreate(nullptr, video_format_, nullptr, attributes, |
- &record, &decompression_session_); |
+ OSStatus status = VTDecompressionSessionCreate( |
+ nullptr, _videoFormat, nullptr, attributes, &record, &_decompressionSession); |
CFRelease(attributes); |
if (status != noErr) { |
- DestroyDecompressionSession(); |
+ [self destroyDecompressionSession]; |
return WEBRTC_VIDEO_CODEC_ERROR; |
} |
- ConfigureDecompressionSession(); |
+ [self configureDecompressionSession]; |
return WEBRTC_VIDEO_CODEC_OK; |
} |
-void H264VideoToolboxDecoder::ConfigureDecompressionSession() { |
- RTC_DCHECK(decompression_session_); |
+- (void)configureDecompressionSession { |
+ RTC_DCHECK(_decompressionSession); |
#if defined(WEBRTC_IOS) |
- VTSessionSetProperty(decompression_session_, |
- kVTDecompressionPropertyKey_RealTime, kCFBooleanTrue); |
+ VTSessionSetProperty(_decompressionSession, kVTDecompressionPropertyKey_RealTime, kCFBooleanTrue); |
#endif |
} |
-void H264VideoToolboxDecoder::DestroyDecompressionSession() { |
- if (decompression_session_) { |
- VTDecompressionSessionInvalidate(decompression_session_); |
- CFRelease(decompression_session_); |
- decompression_session_ = nullptr; |
+- (void)destroyDecompressionSession { |
+ if (_decompressionSession) { |
+ VTDecompressionSessionInvalidate(_decompressionSession); |
+ CFRelease(_decompressionSession); |
+ _decompressionSession = nullptr; |
} |
} |
-void H264VideoToolboxDecoder::SetVideoFormat( |
- CMVideoFormatDescriptionRef video_format) { |
- if (video_format_ == video_format) { |
+- (void)setVideoFormat:(CMVideoFormatDescriptionRef)videoFormat { |
+ if (_videoFormat == videoFormat) { |
return; |
} |
- if (video_format_) { |
- CFRelease(video_format_); |
+ if (_videoFormat) { |
+ CFRelease(_videoFormat); |
} |
- video_format_ = video_format; |
- if (video_format_) { |
- CFRetain(video_format_); |
+ _videoFormat = videoFormat; |
+ if (_videoFormat) { |
+ CFRetain(_videoFormat); |
} |
} |
-const char* H264VideoToolboxDecoder::ImplementationName() const { |
- return "VideoToolbox"; |
+- (NSString *)implementationName { |
+ return @"VideoToolbox"; |
} |
-} // namespace webrtc |
+@end |