Chromium Code Reviews| Index: webrtc/sdk/objc/Framework/Classes/RTCOpenGLVideoRenderer.mm |
| diff --git a/webrtc/sdk/objc/Framework/Classes/RTCOpenGLVideoRenderer.mm b/webrtc/sdk/objc/Framework/Classes/RTCOpenGLVideoRenderer.mm |
| index 7d7b416b888843e4028d5068de1362c57fc5f383..5c082f0fb8bc0f1871d91cf6305896eaba022d0c 100644 |
| --- a/webrtc/sdk/objc/Framework/Classes/RTCOpenGLVideoRenderer.mm |
| +++ b/webrtc/sdk/objc/Framework/Classes/RTCOpenGLVideoRenderer.mm |
| @@ -11,15 +11,19 @@ |
| #import "RTCOpenGLVideoRenderer.h" |
| #if TARGET_OS_IPHONE |
| +#import <CoreVideo/CVOpenGLESTextureCache.h> |
| #import <OpenGLES/ES3/gl.h> |
| #else |
| #import <OpenGL/gl3.h> |
| #endif |
| -#include <string.h> |
| #include <memory> |
| +#include <string.h> |
| +#include <vector> |
| #import "WebRTC/RTCVideoFrame.h" |
| +#include "webrtc/base/checks.h" |
| +#include "webrtc/base/logging.h" |
| // TODO(tkchin): check and log openGL errors. Methods here return BOOLs in |
| // anticipation of that happening in the future. |
| @@ -33,6 +37,8 @@ |
| #define FRAGMENT_SHADER_OUT |
| #define FRAGMENT_SHADER_COLOR "gl_FragColor" |
| #define FRAGMENT_SHADER_TEXTURE "texture2D" |
| + |
| +typedef EAGLContext GlContextType; |
| #else |
| #define RTC_PIXEL_FORMAT GL_RED |
| #define SHADER_VERSION "#version 150\n" |
| @@ -42,8 +48,14 @@ |
| #define FRAGMENT_SHADER_OUT "out vec4 fragColor;\n" |
| #define FRAGMENT_SHADER_COLOR "fragColor" |
| #define FRAGMENT_SHADER_TEXTURE "texture" |
| + |
| +typedef NSOpenGLContext GlContextType; |
| #endif |
| +@protocol Shader |
|
tkchin_webrtc
2016/08/02 00:17:18
nit: everything in ObjC must be prefixed withthe
|
| +- (BOOL)drawFrame:(RTCVideoFrame*)frame; |
|
tkchin_webrtc
2016/08/02 00:17:18
nit: (RTCVideoFrame *)
|
| +@end |
| + |
| // Vertex shader doesn't do anything except pass coordinates through. |
| static const char kVertexShaderSource[] = |
| SHADER_VERSION |
| @@ -57,7 +69,7 @@ static const char kVertexShaderSource[] = |
| // Fragment shader converts YUV values from input textures into a final RGB |
| // pixel. The conversion formula is from http://www.fourcc.org/fccyvrgb.php. |
| -static const char kFragmentShaderSource[] = |
| +static const char kI420FragmentShaderSource[] = |
| SHADER_VERSION |
| "precision highp float;" |
| FRAGMENT_SHADER_IN " vec2 v_texcoord;\n" |
| @@ -90,6 +102,15 @@ GLuint CreateShader(GLenum type, const GLchar *source) { |
| GLint compileStatus = GL_FALSE; |
| glGetShaderiv(shader, GL_COMPILE_STATUS, &compileStatus); |
| if (compileStatus == GL_FALSE) { |
| + GLint logLength = 0; |
| + // The null termination character is included in the returned log length. |
| + glGetShaderiv(shader, GL_INFO_LOG_LENGTH, &logLength); |
| + if (logLength > 0) { |
| + std::unique_ptr<char[]> compileLog(new char[logLength]); |
| + // The returned string is null terminated. |
| + glGetShaderInfoLog(shader, logLength, NULL, compileLog.get()); |
| + LOG(LS_ERROR) << "Shader compile error: " << compileLog.get(); |
| + } |
| glDeleteShader(shader); |
| shader = 0; |
| } |
| @@ -118,6 +139,25 @@ GLuint CreateProgram(GLuint vertexShader, GLuint fragmentShader) { |
| return program; |
| } |
| +// Creates and links a shader program with the given fragment shader source and |
| +// a plain vertex shader. Returns the program handle or 0 on error. |
| +GLuint CreateProgramFromFragmentSource(const char fragmentShaderSource[]) { |
| + GLuint vertexShader = CreateShader(GL_VERTEX_SHADER, kVertexShaderSource); |
| + RTC_CHECK(vertexShader) << "failed to create vertex shader"; |
| + GLuint fragmentShader = |
| + CreateShader(GL_FRAGMENT_SHADER, fragmentShaderSource); |
| + RTC_CHECK(fragmentShader) << "failed to create fragment shader"; |
| + GLuint program = CreateProgram(vertexShader, fragmentShader); |
| + // Shaders are created only to generate program. |
| + if (vertexShader) { |
| + glDeleteShader(vertexShader); |
| + } |
| + if (fragmentShader) { |
| + glDeleteShader(fragmentShader); |
| + } |
| + return program; |
| +} |
| + |
| // When modelview and projection matrices are identity (default) the world is |
| // contained in the square around origin with unit size 2. Drawing to these |
| // coordinates is equivalent to drawing to the entire screen. The texture is |
| @@ -125,7 +165,7 @@ GLuint CreateProgram(GLuint vertexShader, GLuint fragmentShader) { |
| // from (0, 0) to (1, 1) inclusive. Texture coordinates are flipped vertically |
| // here because the incoming frame has origin in upper left hand corner but |
| // OpenGL expects origin in bottom left corner. |
| -const GLfloat gVertices[] = { |
| +static const GLfloat gVertices[] = { |
| // X, Y, U, V. |
| -1, -1, 0, 1, // Bottom left. |
| 1, -1, 1, 1, // Bottom right. |
| @@ -133,6 +173,46 @@ const GLfloat gVertices[] = { |
| -1, 1, 0, 0, // Top left. |
| }; |
| +// Set vertex shader variables 'position' and 'texcoord' in |program| to the |
| +// |gVertices| array above. It will use |vertexBuffer| and |vertexArray| to |
| +// store the vertex data. |
| +BOOL setupVerticesForProgram(GLuint program, GLuint* vertexBuffer, GLuint* vertexArray) { |
| + GLint position = glGetAttribLocation(program, "position"); |
| + GLint texcoord = glGetAttribLocation(program, "texcoord"); |
| + if (position < 0 || texcoord < 0) { |
| + return NO; |
| + } |
| +#if !TARGET_OS_IPHONE |
| + glGenVertexArrays(1, vertexArray); |
| + if (*vertexArray == 0) { |
| + return NO; |
| + } |
| + glBindVertexArray(*vertexArray); |
| +#endif |
| + glGenBuffers(1, vertexBuffer); |
| + if (*vertexBuffer == 0) { |
| + return NO; |
| + } |
| + glBindBuffer(GL_ARRAY_BUFFER, *vertexBuffer); |
| + glBufferData(GL_ARRAY_BUFFER, sizeof(gVertices), gVertices, GL_DYNAMIC_DRAW); |
| + |
| + // Read position attribute from |gVertices| with size of 2 and stride of 4 |
| + // beginning at the start of the array. The last argument indicates offset |
| + // of data within |gVertices| as supplied to the vertex buffer. |
| + glVertexAttribPointer(position, 2, GL_FLOAT, GL_FALSE, 4 * sizeof(GLfloat), |
| + (void*)0); |
|
tkchin_webrtc
2016/08/02 00:17:18
nit: (void *)0
since we're in an ObjC++ file we ca
|
| + glEnableVertexAttribArray(position); |
| + |
| + // Read texcoord attribute from |gVertices| with size of 2 and stride of 4 |
| + // beginning at the first texcoord in the array. The last argument indicates |
| + // offset of data within |gVertices| as supplied to the vertex buffer. |
| + glVertexAttribPointer(texcoord, 2, GL_FLOAT, GL_FALSE, 4 * sizeof(GLfloat), |
| + (void*)(2 * sizeof(GLfloat))); |
|
tkchin_webrtc
2016/08/02 00:17:19
ditto
|
| + glEnableVertexAttribArray(texcoord); |
| + |
| + return YES; |
| +} |
| + |
| // |kNumTextures| must not exceed 8, which is the limit in OpenGLES2. Two sets |
| // of 3 textures are used here, one for each of the Y, U and V planes. Having |
| // two sets alleviates CPU blockage in the event that the GPU is asked to render |
| @@ -140,155 +220,68 @@ const GLfloat gVertices[] = { |
| static const GLsizei kNumTextureSets = 2; |
| static const GLsizei kNumTextures = 3 * kNumTextureSets; |
| -@implementation RTCOpenGLVideoRenderer { |
| -#if TARGET_OS_IPHONE |
| - EAGLContext *_context; |
| -#else |
| - NSOpenGLContext *_context; |
| -#endif |
| - BOOL _isInitialized; |
| +// Shader for non-native I420 frames. |
| +@interface I420Shader : NSObject<Shader> |
|
tkchin_webrtc
2016/08/02 00:17:18
nit: style is NSObject <Shader>
together with pref
|
| +- (BOOL)drawFrame:(RTCVideoFrame*)frame; |
|
tkchin_webrtc
2016/08/02 00:17:19
don't need to redeclare methods from protocol
|
| +@end |
| + |
| +@implementation I420Shader { |
|
tkchin_webrtc
2016/08/02 00:17:17
Can we place the various shaders in their own file
|
| + BOOL _hasUnpackRowLength; |
| GLint _currentTextureSet; |
| // Handles for OpenGL constructs. |
| GLuint _textures[kNumTextures]; |
| - GLuint _program; |
| -#if !TARGET_OS_IPHONE |
| + GLuint _i420Program; |
| GLuint _vertexArray; |
| -#endif |
| GLuint _vertexBuffer; |
| - GLint _position; |
| - GLint _texcoord; |
| GLint _ySampler; |
| GLint _uSampler; |
| GLint _vSampler; |
| // Used to create a non-padded plane for GPU upload when we receive padded |
| // frames. |
| - std::unique_ptr<uint8_t[]> _planeBuffer; |
| -} |
| - |
| -@synthesize lastDrawnFrame = _lastDrawnFrame; |
| - |
| -+ (void)initialize { |
| - // Disable dithering for performance. |
| - glDisable(GL_DITHER); |
| + std::vector<uint8_t> _planeBuffer; |
| } |
| +- (instancetype)initWithContext:(GlContextType *)context { |
| + if (self = [super init]) { |
| #if TARGET_OS_IPHONE |
| -- (instancetype)initWithContext:(EAGLContext *)context { |
| + _hasUnpackRowLength = (context.API == kEAGLRenderingAPIOpenGLES3); |
| #else |
| -- (instancetype)initWithContext:(NSOpenGLContext *)context { |
| + _hasUnpackRowLength = YES; |
| #endif |
| - NSAssert(context != nil, @"context cannot be nil"); |
| - if (self = [super init]) { |
| - _context = context; |
| - } |
| - return self; |
| -} |
| - |
| -- (BOOL)drawFrame:(RTCVideoFrame *)frame { |
| - if (!_isInitialized) { |
| - return NO; |
| - } |
| - if (_lastDrawnFrame == frame) { |
| - return NO; |
| - } |
| - [self ensureGLContext]; |
| - glClear(GL_COLOR_BUFFER_BIT); |
| - if (frame) { |
| - if (![self updateTextureSizesForFrame:frame] || |
| - ![self updateTextureDataForFrame:frame]) { |
| - return NO; |
| + glPixelStorei(GL_UNPACK_ALIGNMENT, 1); |
| + if (![self setupI420Program] || ![self setupTextures] || |
| + !setupVerticesForProgram(_i420Program, &_vertexBuffer, &_vertexArray)) { |
| + self = nil; |
| } |
| -#if !TARGET_OS_IPHONE |
| - glBindVertexArray(_vertexArray); |
| -#endif |
| - glBindBuffer(GL_ARRAY_BUFFER, _vertexBuffer); |
| - glDrawArrays(GL_TRIANGLE_FAN, 0, 4); |
| } |
| -#if !TARGET_OS_IPHONE |
| - [_context flushBuffer]; |
| -#endif |
| - _lastDrawnFrame = frame; |
| - return YES; |
| + return self; |
| } |
| -- (void)setupGL { |
| - if (_isInitialized) { |
| - return; |
| +- (void)dealloc { |
| + if (_i420Program != 0) { |
| + glDeleteProgram(_i420Program); |
| } |
| - [self ensureGLContext]; |
| - if (![self setupProgram]) { |
| - return; |
| - } |
| - if (![self setupTextures]) { |
| - return; |
| - } |
| - if (![self setupVertices]) { |
| - return; |
| + if (_textures[0] != 0) { |
| + glDeleteTextures(kNumTextures, _textures); |
| } |
| - glUseProgram(_program); |
| - glPixelStorei(GL_UNPACK_ALIGNMENT, 1); |
| - _isInitialized = YES; |
| -} |
| - |
| -- (void)teardownGL { |
| - if (!_isInitialized) { |
| - return; |
| + if (_vertexBuffer != 0) { |
| + glDeleteBuffers(1, &_vertexBuffer); |
| } |
| - [self ensureGLContext]; |
| - glDeleteProgram(_program); |
| - _program = 0; |
| - glDeleteTextures(kNumTextures, _textures); |
| - glDeleteBuffers(1, &_vertexBuffer); |
| - _vertexBuffer = 0; |
| -#if !TARGET_OS_IPHONE |
| - glDeleteVertexArrays(1, &_vertexArray); |
| -#endif |
| - _isInitialized = NO; |
| -} |
| - |
| -#pragma mark - Private |
| - |
| -- (void)ensureGLContext { |
| - NSAssert(_context, @"context shouldn't be nil"); |
| -#if TARGET_OS_IPHONE |
| - if ([EAGLContext currentContext] != _context) { |
| - [EAGLContext setCurrentContext:_context]; |
| - } |
| -#else |
| - if ([NSOpenGLContext currentContext] != _context) { |
| - [_context makeCurrentContext]; |
| + if (_vertexArray != 0) { |
| + glDeleteVertexArrays(1, &_vertexArray); |
| } |
| -#endif |
| } |
| -- (BOOL)setupProgram { |
| - NSAssert(!_program, @"program already set up"); |
| - GLuint vertexShader = CreateShader(GL_VERTEX_SHADER, kVertexShaderSource); |
| - NSAssert(vertexShader, @"failed to create vertex shader"); |
| - GLuint fragmentShader = |
| - CreateShader(GL_FRAGMENT_SHADER, kFragmentShaderSource); |
| - NSAssert(fragmentShader, @"failed to create fragment shader"); |
| - _program = CreateProgram(vertexShader, fragmentShader); |
| - // Shaders are created only to generate program. |
| - if (vertexShader) { |
| - glDeleteShader(vertexShader); |
| - } |
| - if (fragmentShader) { |
| - glDeleteShader(fragmentShader); |
| - } |
| - if (!_program) { |
| - return NO; |
| - } |
| - _position = glGetAttribLocation(_program, "position"); |
| - _texcoord = glGetAttribLocation(_program, "texcoord"); |
| - _ySampler = glGetUniformLocation(_program, "s_textureY"); |
| - _uSampler = glGetUniformLocation(_program, "s_textureU"); |
| - _vSampler = glGetUniformLocation(_program, "s_textureV"); |
| - if (_position < 0 || _texcoord < 0 || _ySampler < 0 || _uSampler < 0 || |
| - _vSampler < 0) { |
| +- (BOOL)setupI420Program { |
| + _i420Program = CreateProgramFromFragmentSource(kI420FragmentShaderSource); |
| + if (!_i420Program) { |
| return NO; |
| } |
| - return YES; |
| + _ySampler = glGetUniformLocation(_i420Program, "s_textureY"); |
| + _uSampler = glGetUniformLocation(_i420Program, "s_textureU"); |
| + _vSampler = glGetUniformLocation(_i420Program, "s_textureV"); |
| + |
| + return (_ySampler >= 0 && _uSampler >= 0 && _vSampler >= 0); |
|
tkchin_webrtc
2016/08/02 00:17:19
these ivars are init-ed to zero by default after i
|
| } |
| - (BOOL)setupTextures { |
| @@ -305,56 +298,17 @@ static const GLsizei kNumTextures = 3 * kNumTextureSets; |
| return YES; |
| } |
| -- (BOOL)updateTextureSizesForFrame:(RTCVideoFrame *)frame { |
| - if (frame.height == _lastDrawnFrame.height && |
| - frame.width == _lastDrawnFrame.width && |
| - frame.chromaWidth == _lastDrawnFrame.chromaWidth && |
| - frame.chromaHeight == _lastDrawnFrame.chromaHeight) { |
| - return YES; |
| - } |
| - GLsizei lumaWidth = static_cast<GLsizei>(frame.width); |
| - GLsizei lumaHeight = static_cast<GLsizei>(frame.height); |
| - GLsizei chromaWidth = static_cast<GLsizei>(frame.chromaWidth); |
| - GLsizei chromaHeight = static_cast<GLsizei>(frame.chromaHeight); |
| - for (GLint i = 0; i < kNumTextureSets; i++) { |
| - glActiveTexture(GL_TEXTURE0 + i * 3); |
| - glTexImage2D(GL_TEXTURE_2D, |
| - 0, |
| - RTC_PIXEL_FORMAT, |
| - lumaWidth, |
| - lumaHeight, |
| - 0, |
| - RTC_PIXEL_FORMAT, |
| - GL_UNSIGNED_BYTE, |
| - 0); |
| - glActiveTexture(GL_TEXTURE0 + i * 3 + 1); |
| - glTexImage2D(GL_TEXTURE_2D, |
| - 0, |
| - RTC_PIXEL_FORMAT, |
| - chromaWidth, |
| - chromaHeight, |
| - 0, |
| - RTC_PIXEL_FORMAT, |
| - GL_UNSIGNED_BYTE, |
| - 0); |
| - glActiveTexture(GL_TEXTURE0 + i * 3 + 2); |
| - glTexImage2D(GL_TEXTURE_2D, |
| - 0, |
| - RTC_PIXEL_FORMAT, |
| - chromaWidth, |
| - chromaHeight, |
| - 0, |
| - RTC_PIXEL_FORMAT, |
| - GL_UNSIGNED_BYTE, |
| - 0); |
| - } |
| - if ((NSUInteger)frame.yPitch != frame.width || |
| - (NSUInteger)frame.uPitch != frame.chromaWidth || |
| - (NSUInteger)frame.vPitch != frame.chromaWidth) { |
| - _planeBuffer.reset(new uint8_t[frame.width * frame.height]); |
| - } else { |
| - _planeBuffer.reset(); |
| +- (BOOL)drawFrame:(RTCVideoFrame*)frame { |
| + glUseProgram(_i420Program); |
| + if (![self updateTextureDataForFrame:frame]) { |
| + return NO; |
| } |
| +#if !TARGET_OS_IPHONE |
| + glBindVertexArray(_vertexArray); |
| +#endif |
| + glBindBuffer(GL_ARRAY_BUFFER, _vertexBuffer); |
| + glDrawArrays(GL_TRIANGLE_FAN, 0, 4); |
| + |
| return YES; |
| } |
| @@ -365,17 +319,14 @@ static const GLsizei kNumTextures = 3 * kNumTextureSets; |
| height:(size_t)height |
| stride:(int32_t)stride { |
| glActiveTexture(static_cast<GLenum>(GL_TEXTURE0 + offset)); |
| + glBindTexture(GL_TEXTURE_2D, _textures[offset]); |
| + |
| // When setting texture sampler uniforms, the texture index is used not |
| // the texture handle. |
| glUniform1i(sampler, offset); |
| -#if TARGET_OS_IPHONE |
| - BOOL hasUnpackRowLength = _context.API == kEAGLRenderingAPIOpenGLES3; |
| -#else |
| - BOOL hasUnpackRowLength = YES; |
| -#endif |
| const uint8_t *uploadPlane = plane; |
| if ((size_t)stride != width) { |
| - if (hasUnpackRowLength) { |
| + if (_hasUnpackRowLength) { |
| // GLES3 allows us to specify stride. |
| glPixelStorei(GL_UNPACK_ROW_LENGTH, stride); |
| glTexImage2D(GL_TEXTURE_2D, |
| @@ -392,7 +343,7 @@ static const GLsizei kNumTextures = 3 * kNumTextureSets; |
| } else { |
| // Make an unpadded copy and upload that instead. Quick profiling showed |
| // that this is faster than uploading row by row using glTexSubImage2D. |
| - uint8_t *unpaddedPlane = _planeBuffer.get(); |
| + uint8_t *unpaddedPlane = _planeBuffer.data(); |
| for (size_t y = 0; y < height; ++y) { |
| memcpy(unpaddedPlane + y * width, plane + y * stride, width); |
| } |
| @@ -414,6 +365,12 @@ static const GLsizei kNumTextures = 3 * kNumTextureSets; |
| GLint textureOffset = _currentTextureSet * 3; |
| NSAssert(textureOffset + 3 <= kNumTextures, @"invalid offset"); |
| + if (frame.yPitch != static_cast<int32_t>(frame.width) || |
| + frame.uPitch != static_cast<int32_t>(frame.chromaWidth) || |
| + frame.vPitch != static_cast<int32_t>(frame.chromaWidth)) { |
| + _planeBuffer.resize(frame.width * frame.height); |
| + } |
| + |
| [self uploadPlane:frame.yPlane |
| sampler:_ySampler |
| offset:textureOffset |
| @@ -439,46 +396,246 @@ static const GLsizei kNumTextures = 3 * kNumTextureSets; |
| return YES; |
| } |
| -- (BOOL)setupVertices { |
| -#if !TARGET_OS_IPHONE |
| - NSAssert(!_vertexArray, @"vertex array already set up"); |
| - glGenVertexArrays(1, &_vertexArray); |
| - if (!_vertexArray) { |
| +@end |
| + |
| +// Native CVPixelBufferRef rendering is only supported on iPhone because it |
| +// depends on CVOpenGLESTextureCacheCreate. |
| +#if TARGET_OS_IPHONE |
| +static const char kNV12FragmentShaderSource[] = |
| + SHADER_VERSION |
| + "precision mediump float;" |
| + FRAGMENT_SHADER_IN " vec2 v_texcoord;\n" |
| + "uniform lowp sampler2D s_textureY;\n" |
| + "uniform lowp sampler2D s_textureUV;\n" |
| + FRAGMENT_SHADER_OUT |
| + "void main() {\n" |
| + " mediump float y;\n" |
| + " mediump vec2 uv;\n" |
| + " y = " FRAGMENT_SHADER_TEXTURE "(s_textureY, v_texcoord).r;\n" |
| + " uv = " FRAGMENT_SHADER_TEXTURE "(s_textureUV, v_texcoord).ra -\n" |
| + " vec2(0.5, 0.5);\n" |
| + " " FRAGMENT_SHADER_COLOR " = vec4(y + 1.403 * uv.y,\n" |
| + " y - 0.344 * uv.x - 0.714 * uv.y,\n" |
| + " y + 1.770 * uv.x,\n" |
| + " 1.0);\n" |
| + " }\n"; |
| + |
| +// Shader for native NV12 frames. |
| +@interface NativeNV12Shader : NSObject<Shader> |
| +- (BOOL)drawFrame:(RTCVideoFrame*)frame; |
| +@end |
| + |
| +@implementation NativeNV12Shader { |
| + GLuint _vertexBuffer; |
| + GLuint _nv12Program; |
| + GLint _ySampler; |
| + GLint _uvSampler; |
| + CVOpenGLESTextureCacheRef _textureCache; |
| +} |
| + |
| +- (instancetype)initWithContext:(GlContextType *)context { |
| + if (self = [super init]) { |
| + if (![self setupNV12Program] || ![self setupTextureCacheWithContext:context] || |
| + !setupVerticesForProgram(_nv12Program, &_vertexBuffer, nullptr)) { |
| + self = nil; |
| + } |
| + } |
| + return self; |
| +} |
| + |
| +- (void)dealloc { |
| + if (_nv12Program != 0) { |
| + glDeleteProgram(_nv12Program); |
| + } |
| + if (_vertexBuffer != 0) { |
| + glDeleteBuffers(1, &_vertexBuffer); |
| + } |
| + if (_textureCache) { |
| + CFRelease(_textureCache); |
|
tkchin_webrtc
2016/08/02 00:17:18
nit: assign nullptr after release
|
| + } |
| +} |
| + |
| +- (BOOL)setupNV12Program { |
| + _nv12Program = CreateProgramFromFragmentSource(kNV12FragmentShaderSource); |
| + if (!_nv12Program) { |
| return NO; |
| } |
| - glBindVertexArray(_vertexArray); |
| -#endif |
| - NSAssert(!_vertexBuffer, @"vertex buffer already set up"); |
| - glGenBuffers(1, &_vertexBuffer); |
| - if (!_vertexBuffer) { |
| -#if !TARGET_OS_IPHONE |
| - glDeleteVertexArrays(1, &_vertexArray); |
| - _vertexArray = 0; |
| + _ySampler = glGetUniformLocation(_nv12Program, "s_textureY"); |
| + _uvSampler = glGetUniformLocation(_nv12Program, "s_textureUV"); |
| + |
| + return (_ySampler >= 0 && _uvSampler >= 0); |
| +} |
| + |
| +- (BOOL)setupTextureCacheWithContext:(GlContextType *)context { |
| + CVReturn err = CVOpenGLESTextureCacheCreate( |
| + kCFAllocatorDefault, NULL, |
| +#if COREVIDEO_USE_EAGLCONTEXT_CLASS_IN_API |
| + context, |
| +#else |
| + (__bridge void*)context, |
|
tkchin_webrtc
2016/08/02 00:17:17
nit: void *)
|
| #endif |
| + NULL, &_textureCache); |
| + return !err; |
|
tkchin_webrtc
2016/08/02 00:17:17
err == noErr or CVReturnTrue ? I remember there wa
|
| +} |
| + |
| +- (BOOL)drawFrame:(RTCVideoFrame*)frame { |
|
tkchin_webrtc
2016/08/02 00:17:17
nit: RTCVideoFrame *)
|
| + CVPixelBufferRef pixelBuffer = frame.nativeHandle; |
| + RTC_CHECK(pixelBuffer); |
| + glUseProgram(_nv12Program); |
| + const OSType pixelFormat = CVPixelBufferGetPixelFormatType(pixelBuffer); |
| + RTC_CHECK(pixelFormat == kCVPixelFormatType_420YpCbCr8BiPlanarFullRange || |
| + pixelFormat == kCVPixelFormatType_420YpCbCr8BiPlanarVideoRange) |
| + << "Unsupported native pixel format: " << pixelFormat; |
| + |
| + // Y-plane. |
| + const int luma_width = CVPixelBufferGetWidthOfPlane(pixelBuffer, 0); |
|
tkchin_webrtc
2016/08/02 00:17:18
nit: camel case instead of underscore for ObjC fun
|
| + const int luma_height = CVPixelBufferGetHeightOfPlane(pixelBuffer, 0); |
| + |
| + CVOpenGLESTextureRef lumaTexture; |
|
tkchin_webrtc
2016/08/02 00:17:18
nit: init to nullptr
|
| + glActiveTexture(GL_TEXTURE0); |
| + glUniform1i(_ySampler, 0); |
| + CVReturn err = CVOpenGLESTextureCacheCreateTextureFromImage( |
| + kCFAllocatorDefault, _textureCache, pixelBuffer, NULL, GL_TEXTURE_2D, |
| + RTC_PIXEL_FORMAT, luma_width, luma_height, RTC_PIXEL_FORMAT, |
| + GL_UNSIGNED_BYTE, 0, &lumaTexture); |
| + if (err) { |
| + CFRelease(lumaTexture); |
| + return NO; |
| + } |
| + |
| + RTC_CHECK_EQ(static_cast<GLenum>(GL_TEXTURE_2D), |
| + CVOpenGLESTextureGetTarget(lumaTexture)); |
| + glBindTexture(GL_TEXTURE_2D, CVOpenGLESTextureGetName(lumaTexture)); |
| + glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_LINEAR); |
| + glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_LINEAR); |
| + glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE); |
| + glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE); |
| + |
| + // UV-plane. |
| + const int chroma_width = CVPixelBufferGetWidthOfPlane(pixelBuffer, 1); |
|
tkchin_webrtc
2016/08/02 00:17:19
ditto chromeWidth / chromaHeight
|
| + const int chroma_height = CVPixelBufferGetHeightOfPlane(pixelBuffer, 1); |
| + |
| + CVOpenGLESTextureRef chromaTexture; |
| + glActiveTexture(GL_TEXTURE1); |
| + glUniform1i(_uvSampler, 1); |
| + err = CVOpenGLESTextureCacheCreateTextureFromImage( |
| + kCFAllocatorDefault, _textureCache, pixelBuffer, NULL, GL_TEXTURE_2D, |
| + GL_LUMINANCE_ALPHA, chroma_width, chroma_height, GL_LUMINANCE_ALPHA, |
| + GL_UNSIGNED_BYTE, 1, &chromaTexture); |
| + if (err) { |
| + CFRelease(chromaTexture); |
| + CFRelease(lumaTexture); |
| return NO; |
| } |
| + |
| + RTC_CHECK_EQ(static_cast<GLenum>(GL_TEXTURE_2D), |
| + CVOpenGLESTextureGetTarget(chromaTexture)); |
| + glBindTexture(GL_TEXTURE_2D, CVOpenGLESTextureGetName(chromaTexture)); |
| + glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_LINEAR); |
| + glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_LINEAR); |
| + glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE); |
| + glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE); |
| + |
| glBindBuffer(GL_ARRAY_BUFFER, _vertexBuffer); |
| - glBufferData(GL_ARRAY_BUFFER, sizeof(gVertices), gVertices, GL_DYNAMIC_DRAW); |
| + glDrawArrays(GL_TRIANGLE_FAN, 0, 4); |
| - // Read position attribute from |gVertices| with size of 2 and stride of 4 |
| - // beginning at the start of the array. The last argument indicates offset |
| - // of data within |gVertices| as supplied to the vertex buffer. |
| - glVertexAttribPointer( |
| - _position, 2, GL_FLOAT, GL_FALSE, 4 * sizeof(GLfloat), (void *)0); |
| - glEnableVertexAttribArray(_position); |
| + CFRelease(chromaTexture); |
| + CFRelease(lumaTexture); |
| - // Read texcoord attribute from |gVertices| with size of 2 and stride of 4 |
| - // beginning at the first texcoord in the array. The last argument indicates |
| - // offset of data within |gVertices| as supplied to the vertex buffer. |
| - glVertexAttribPointer(_texcoord, |
| - 2, |
| - GL_FLOAT, |
| - GL_FALSE, |
| - 4 * sizeof(GLfloat), |
| - (void *)(2 * sizeof(GLfloat))); |
| - glEnableVertexAttribArray(_texcoord); |
| + return YES; |
| +} |
| + |
| +@end |
| +#endif // TARGET_OS_IPHONE |
| + |
| +@implementation RTCOpenGLVideoRenderer { |
| + GlContextType* _context; |
|
tkchin_webrtc
2016/08/02 00:17:19
nit: GLContextType *_context
|
| + BOOL _isInitialized; |
| + NSObject<Shader>* _i420Shader; |
|
tkchin_webrtc
2016/08/02 00:17:17
typically we declare id instead of saying that it
|
| + NSObject<Shader>* _nv12Shader; |
| +} |
| + |
| +@synthesize lastDrawnFrame = _lastDrawnFrame; |
| + |
| ++ (void)initialize { |
| + // Disable dithering for performance. |
| + glDisable(GL_DITHER); |
| +} |
| + |
| +- (instancetype)initWithContext:(GlContextType *)context { |
| + NSAssert(context != nil, @"context cannot be nil"); |
| + if (self = [super init]) { |
| + _context = context; |
| + } |
| + return self; |
| +} |
| + |
| +- (BOOL)drawFrame:(RTCVideoFrame*)frame { |
| + if (!_isInitialized || !frame || frame == _lastDrawnFrame) { |
| + return NO; |
| + } |
| + [self ensureGLContext]; |
| + glClear(GL_COLOR_BUFFER_BIT); |
| + NSObject<Shader>* shader; |
|
tkchin_webrtc
2016/08/02 00:17:19
id<RTCShader> shader = nil;
|
| +#if TARGET_OS_IPHONE |
| + if (frame.nativeHandle) { |
| + if (!_nv12Shader) { |
| + _nv12Shader = [[NativeNV12Shader alloc] initWithContext:_context]; |
| + } |
| + shader = _nv12Shader; |
| +#else |
| + // Rendering native CVPixelBuffer is not supported on OS X. |
| + if (false) { |
| +#endif |
| + } else { |
| + if (!_i420Shader) { |
| + _i420Shader = [[I420Shader alloc] initWithContext:_context]; |
| + } |
| + shader = _i420Shader; |
| + } |
| + if (!shader || ![shader drawFrame:frame]) { |
| + return NO; |
| + } |
| + |
| +#if !TARGET_OS_IPHONE |
| + [_context flushBuffer]; |
| +#endif |
| + _lastDrawnFrame = frame; |
| return YES; |
| } |
| +- (void)setupGL { |
| + if (_isInitialized) { |
| + return; |
| + } |
| + [self ensureGLContext]; |
| + _isInitialized = YES; |
| +} |
| + |
| +- (void)teardownGL { |
| + if (!_isInitialized) { |
| + return; |
| + } |
| + [self ensureGLContext]; |
| + _i420Shader = nil; |
| + _nv12Shader = nil; |
| + _isInitialized = NO; |
| +} |
| + |
| +#pragma mark - Private |
| + |
| +- (void)ensureGLContext { |
| + NSAssert(_context, @"context shouldn't be nil"); |
| +#if TARGET_OS_IPHONE |
| + if ([EAGLContext currentContext] != _context) { |
| + [EAGLContext setCurrentContext:_context]; |
| + } |
| +#else |
| + if ([NSOpenGLContext currentContext] != _context) { |
| + [_context makeCurrentContext]; |
| + } |
| +#endif |
| +} |
| + |
| @end |