Index: webrtc/modules/video_render/android/video_render_opengles20.cc |
diff --git a/webrtc/modules/video_render/android/video_render_opengles20.cc b/webrtc/modules/video_render/android/video_render_opengles20.cc |
new file mode 100644 |
index 0000000000000000000000000000000000000000..45db56a4f6e6cd93d68abc7d5eb6557f38bb0e43 |
--- /dev/null |
+++ b/webrtc/modules/video_render/android/video_render_opengles20.cc |
@@ -0,0 +1,397 @@ |
+/* |
+ * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved. |
+ * |
+ * Use of this source code is governed by a BSD-style license |
+ * that can be found in the LICENSE file in the root of the source |
+ * tree. An additional intellectual property rights grant can be found |
+ * in the file PATENTS. All contributing project authors may |
+ * be found in the AUTHORS file in the root of the source tree. |
+ */ |
+ |
+#include <GLES2/gl2.h> |
+#include <GLES2/gl2ext.h> |
+ |
+#include <stdio.h> |
+#include <stdlib.h> |
+ |
+#include "webrtc/modules/video_render/android/video_render_opengles20.h" |
+ |
+//#define ANDROID_LOG |
+ |
+#ifdef ANDROID_LOG |
+#include <android/log.h> |
+#include <stdio.h> |
+ |
+#undef WEBRTC_TRACE |
+#define WEBRTC_TRACE(a,b,c,...) __android_log_print(ANDROID_LOG_DEBUG, "*WEBRTCN*", __VA_ARGS__) |
+#else |
+#include "webrtc/system_wrappers/include/trace.h" |
+#endif |
+ |
+namespace webrtc { |
+ |
+const char VideoRenderOpenGles20::g_indices[] = { 0, 3, 2, 0, 2, 1 }; |
+ |
+const char VideoRenderOpenGles20::g_vertextShader[] = { |
+ "attribute vec4 aPosition;\n" |
+ "attribute vec2 aTextureCoord;\n" |
+ "varying vec2 vTextureCoord;\n" |
+ "void main() {\n" |
+ " gl_Position = aPosition;\n" |
+ " vTextureCoord = aTextureCoord;\n" |
+ "}\n" }; |
+ |
+// The fragment shader. |
+// Do YUV to RGB565 conversion. |
+const char VideoRenderOpenGles20::g_fragmentShader[] = { |
+ "precision mediump float;\n" |
+ "uniform sampler2D Ytex;\n" |
+ "uniform sampler2D Utex,Vtex;\n" |
+ "varying vec2 vTextureCoord;\n" |
+ "void main(void) {\n" |
+ " float nx,ny,r,g,b,y,u,v;\n" |
+ " mediump vec4 txl,ux,vx;" |
+ " nx=vTextureCoord[0];\n" |
+ " ny=vTextureCoord[1];\n" |
+ " y=texture2D(Ytex,vec2(nx,ny)).r;\n" |
+ " u=texture2D(Utex,vec2(nx,ny)).r;\n" |
+ " v=texture2D(Vtex,vec2(nx,ny)).r;\n" |
+ |
+ //" y = v;\n"+ |
+ " y=1.1643*(y-0.0625);\n" |
+ " u=u-0.5;\n" |
+ " v=v-0.5;\n" |
+ |
+ " r=y+1.5958*v;\n" |
+ " g=y-0.39173*u-0.81290*v;\n" |
+ " b=y+2.017*u;\n" |
+ " gl_FragColor=vec4(r,g,b,1.0);\n" |
+ "}\n" }; |
+ |
+VideoRenderOpenGles20::VideoRenderOpenGles20(int32_t id) : |
+ _id(id), |
+ _textureWidth(-1), |
+ _textureHeight(-1) { |
+ WEBRTC_TRACE(kTraceDebug, kTraceVideoRenderer, _id, "%s: id %d", |
+ __FUNCTION__, (int) _id); |
+ |
+ const GLfloat vertices[20] = { |
+ // X, Y, Z, U, V |
+ -1, -1, 0, 0, 1, // Bottom Left |
+ 1, -1, 0, 1, 1, //Bottom Right |
+ 1, 1, 0, 1, 0, //Top Right |
+ -1, 1, 0, 0, 0 }; //Top Left |
+ |
+ memcpy(_vertices, vertices, sizeof(_vertices)); |
+} |
+ |
+VideoRenderOpenGles20::~VideoRenderOpenGles20() { |
+} |
+ |
+int32_t VideoRenderOpenGles20::Setup(int32_t width, int32_t height) { |
+ WEBRTC_TRACE(kTraceDebug, kTraceVideoRenderer, _id, |
+ "%s: width %d, height %d", __FUNCTION__, (int) width, |
+ (int) height); |
+ |
+ printGLString("Version", GL_VERSION); |
+ printGLString("Vendor", GL_VENDOR); |
+ printGLString("Renderer", GL_RENDERER); |
+ printGLString("Extensions", GL_EXTENSIONS); |
+ |
+ int maxTextureImageUnits[2]; |
+ int maxTextureSize[2]; |
+ glGetIntegerv(GL_MAX_TEXTURE_IMAGE_UNITS, maxTextureImageUnits); |
+ glGetIntegerv(GL_MAX_TEXTURE_SIZE, maxTextureSize); |
+ |
+ WEBRTC_TRACE(kTraceDebug, kTraceVideoRenderer, _id, |
+ "%s: number of textures %d, size %d", __FUNCTION__, |
+ (int) maxTextureImageUnits[0], (int) maxTextureSize[0]); |
+ |
+ _program = createProgram(g_vertextShader, g_fragmentShader); |
+ if (!_program) { |
+ WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id, |
+ "%s: Could not create program", __FUNCTION__); |
+ return -1; |
+ } |
+ |
+ int positionHandle = glGetAttribLocation(_program, "aPosition"); |
+ checkGlError("glGetAttribLocation aPosition"); |
+ if (positionHandle == -1) { |
+ WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id, |
+ "%s: Could not get aPosition handle", __FUNCTION__); |
+ return -1; |
+ } |
+ |
+ int textureHandle = glGetAttribLocation(_program, "aTextureCoord"); |
+ checkGlError("glGetAttribLocation aTextureCoord"); |
+ if (textureHandle == -1) { |
+ WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id, |
+ "%s: Could not get aTextureCoord handle", __FUNCTION__); |
+ return -1; |
+ } |
+ |
+ // set the vertices array in the shader |
+ // _vertices contains 4 vertices with 5 coordinates. |
+ // 3 for (xyz) for the vertices and 2 for the texture |
+ glVertexAttribPointer(positionHandle, 3, GL_FLOAT, false, |
+ 5 * sizeof(GLfloat), _vertices); |
+ checkGlError("glVertexAttribPointer aPosition"); |
+ |
+ glEnableVertexAttribArray(positionHandle); |
+ checkGlError("glEnableVertexAttribArray positionHandle"); |
+ |
+ // set the texture coordinate array in the shader |
+ // _vertices contains 4 vertices with 5 coordinates. |
+ // 3 for (xyz) for the vertices and 2 for the texture |
+ glVertexAttribPointer(textureHandle, 2, GL_FLOAT, false, 5 |
+ * sizeof(GLfloat), &_vertices[3]); |
+ checkGlError("glVertexAttribPointer maTextureHandle"); |
+ glEnableVertexAttribArray(textureHandle); |
+ checkGlError("glEnableVertexAttribArray textureHandle"); |
+ |
+ glUseProgram(_program); |
+ int i = glGetUniformLocation(_program, "Ytex"); |
+ checkGlError("glGetUniformLocation"); |
+ glUniform1i(i, 0); /* Bind Ytex to texture unit 0 */ |
+ checkGlError("glUniform1i Ytex"); |
+ |
+ i = glGetUniformLocation(_program, "Utex"); |
+ checkGlError("glGetUniformLocation Utex"); |
+ glUniform1i(i, 1); /* Bind Utex to texture unit 1 */ |
+ checkGlError("glUniform1i Utex"); |
+ |
+ i = glGetUniformLocation(_program, "Vtex"); |
+ checkGlError("glGetUniformLocation"); |
+ glUniform1i(i, 2); /* Bind Vtex to texture unit 2 */ |
+ checkGlError("glUniform1i"); |
+ |
+ glViewport(0, 0, width, height); |
+ checkGlError("glViewport"); |
+ return 0; |
+} |
+ |
+// SetCoordinates |
+// Sets the coordinates where the stream shall be rendered. |
+// Values must be between 0 and 1. |
+int32_t VideoRenderOpenGles20::SetCoordinates(int32_t zOrder, |
+ const float left, |
+ const float top, |
+ const float right, |
+ const float bottom) { |
+ if ((top > 1 || top < 0) || (right > 1 || right < 0) || |
+ (bottom > 1 || bottom < 0) || (left > 1 || left < 0)) { |
+ WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id, |
+ "%s: Wrong coordinates", __FUNCTION__); |
+ return -1; |
+ } |
+ |
+ // X, Y, Z, U, V |
+ // -1, -1, 0, 0, 1, // Bottom Left |
+ // 1, -1, 0, 1, 1, //Bottom Right |
+ // 1, 1, 0, 1, 0, //Top Right |
+ // -1, 1, 0, 0, 0 //Top Left |
+ |
+ // Bottom Left |
+ _vertices[0] = (left * 2) - 1; |
+ _vertices[1] = -1 * (2 * bottom) + 1; |
+ _vertices[2] = zOrder; |
+ |
+ //Bottom Right |
+ _vertices[5] = (right * 2) - 1; |
+ _vertices[6] = -1 * (2 * bottom) + 1; |
+ _vertices[7] = zOrder; |
+ |
+ //Top Right |
+ _vertices[10] = (right * 2) - 1; |
+ _vertices[11] = -1 * (2 * top) + 1; |
+ _vertices[12] = zOrder; |
+ |
+ //Top Left |
+ _vertices[15] = (left * 2) - 1; |
+ _vertices[16] = -1 * (2 * top) + 1; |
+ _vertices[17] = zOrder; |
+ |
+ return 0; |
+} |
+ |
+int32_t VideoRenderOpenGles20::Render(const VideoFrame& frameToRender) { |
+ if (frameToRender.IsZeroSize()) { |
+ return -1; |
+ } |
+ |
+ WEBRTC_TRACE(kTraceDebug, kTraceVideoRenderer, _id, "%s: id %d", |
+ __FUNCTION__, (int) _id); |
+ |
+ glUseProgram(_program); |
+ checkGlError("glUseProgram"); |
+ |
+ if (_textureWidth != (GLsizei) frameToRender.width() || |
+ _textureHeight != (GLsizei) frameToRender.height()) { |
+ SetupTextures(frameToRender); |
+ } |
+ UpdateTextures(frameToRender); |
+ |
+ glDrawElements(GL_TRIANGLES, 6, GL_UNSIGNED_BYTE, g_indices); |
+ checkGlError("glDrawArrays"); |
+ |
+ return 0; |
+} |
+ |
+GLuint VideoRenderOpenGles20::loadShader(GLenum shaderType, |
+ const char* pSource) { |
+ GLuint shader = glCreateShader(shaderType); |
+ if (shader) { |
+ glShaderSource(shader, 1, &pSource, NULL); |
+ glCompileShader(shader); |
+ GLint compiled = 0; |
+ glGetShaderiv(shader, GL_COMPILE_STATUS, &compiled); |
+ if (!compiled) { |
+ GLint infoLen = 0; |
+ glGetShaderiv(shader, GL_INFO_LOG_LENGTH, &infoLen); |
+ if (infoLen) { |
+ char* buf = (char*) malloc(infoLen); |
+ if (buf) { |
+ glGetShaderInfoLog(shader, infoLen, NULL, buf); |
+ WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id, |
+ "%s: Could not compile shader %d: %s", |
+ __FUNCTION__, shaderType, buf); |
+ free(buf); |
+ } |
+ glDeleteShader(shader); |
+ shader = 0; |
+ } |
+ } |
+ } |
+ return shader; |
+} |
+ |
+GLuint VideoRenderOpenGles20::createProgram(const char* pVertexSource, |
+ const char* pFragmentSource) { |
+ GLuint vertexShader = loadShader(GL_VERTEX_SHADER, pVertexSource); |
+ if (!vertexShader) { |
+ return 0; |
+ } |
+ |
+ GLuint pixelShader = loadShader(GL_FRAGMENT_SHADER, pFragmentSource); |
+ if (!pixelShader) { |
+ return 0; |
+ } |
+ |
+ GLuint program = glCreateProgram(); |
+ if (program) { |
+ glAttachShader(program, vertexShader); |
+ checkGlError("glAttachShader"); |
+ glAttachShader(program, pixelShader); |
+ checkGlError("glAttachShader"); |
+ glLinkProgram(program); |
+ GLint linkStatus = GL_FALSE; |
+ glGetProgramiv(program, GL_LINK_STATUS, &linkStatus); |
+ if (linkStatus != GL_TRUE) { |
+ GLint bufLength = 0; |
+ glGetProgramiv(program, GL_INFO_LOG_LENGTH, &bufLength); |
+ if (bufLength) { |
+ char* buf = (char*) malloc(bufLength); |
+ if (buf) { |
+ glGetProgramInfoLog(program, bufLength, NULL, buf); |
+ WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id, |
+ "%s: Could not link program: %s", |
+ __FUNCTION__, buf); |
+ free(buf); |
+ } |
+ } |
+ glDeleteProgram(program); |
+ program = 0; |
+ } |
+ } |
+ return program; |
+} |
+ |
+void VideoRenderOpenGles20::printGLString(const char *name, GLenum s) { |
+ const char *v = (const char *) glGetString(s); |
+ WEBRTC_TRACE(kTraceDebug, kTraceVideoRenderer, _id, "GL %s = %s\n", |
+ name, v); |
+} |
+ |
+void VideoRenderOpenGles20::checkGlError(const char* op) { |
+#ifdef ANDROID_LOG |
+ for (GLint error = glGetError(); error; error = glGetError()) { |
+ WEBRTC_TRACE(kTraceError, kTraceVideoRenderer, _id, |
+ "after %s() glError (0x%x)\n", op, error); |
+ } |
+#else |
+ return; |
+#endif |
+} |
+ |
+static void InitializeTexture(int name, int id, int width, int height) { |
+ glActiveTexture(name); |
+ glBindTexture(GL_TEXTURE_2D, id); |
+ glTexParameterf(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_NEAREST); |
+ glTexParameterf(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_LINEAR); |
+ glTexParameterf(GL_TEXTURE_2D, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE); |
+ glTexParameterf(GL_TEXTURE_2D, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE); |
+ glTexImage2D(GL_TEXTURE_2D, 0, GL_LUMINANCE, width, height, 0, |
+ GL_LUMINANCE, GL_UNSIGNED_BYTE, NULL); |
+} |
+ |
+void VideoRenderOpenGles20::SetupTextures(const VideoFrame& frameToRender) { |
+ WEBRTC_TRACE(kTraceDebug, kTraceVideoRenderer, _id, |
+ "%s: width %d, height %d", __FUNCTION__, |
+ frameToRender.width(), frameToRender.height()); |
+ |
+ const GLsizei width = frameToRender.width(); |
+ const GLsizei height = frameToRender.height(); |
+ |
+ glGenTextures(3, _textureIds); //Generate the Y, U and V texture |
+ InitializeTexture(GL_TEXTURE0, _textureIds[0], width, height); |
+ InitializeTexture(GL_TEXTURE1, _textureIds[1], width / 2, height / 2); |
+ InitializeTexture(GL_TEXTURE2, _textureIds[2], width / 2, height / 2); |
+ |
+ checkGlError("SetupTextures"); |
+ |
+ _textureWidth = width; |
+ _textureHeight = height; |
+} |
+ |
+// Uploads a plane of pixel data, accounting for stride != width*bpp. |
+static void GlTexSubImage2D(GLsizei width, GLsizei height, int stride, |
+ const uint8_t* plane) { |
+ if (stride == width) { |
+ // Yay! We can upload the entire plane in a single GL call. |
+ glTexSubImage2D(GL_TEXTURE_2D, 0, 0, 0, width, height, GL_LUMINANCE, |
+ GL_UNSIGNED_BYTE, |
+ static_cast<const GLvoid*>(plane)); |
+ } else { |
+ // Boo! Since GLES2 doesn't have GL_UNPACK_ROW_LENGTH and Android doesn't |
+ // have GL_EXT_unpack_subimage we have to upload a row at a time. Ick. |
+ for (int row = 0; row < height; ++row) { |
+ glTexSubImage2D(GL_TEXTURE_2D, 0, 0, row, width, 1, GL_LUMINANCE, |
+ GL_UNSIGNED_BYTE, |
+ static_cast<const GLvoid*>(plane + (row * stride))); |
+ } |
+ } |
+} |
+ |
+void VideoRenderOpenGles20::UpdateTextures(const VideoFrame& frameToRender) { |
+ const GLsizei width = frameToRender.width(); |
+ const GLsizei height = frameToRender.height(); |
+ |
+ glActiveTexture(GL_TEXTURE0); |
+ glBindTexture(GL_TEXTURE_2D, _textureIds[0]); |
+ GlTexSubImage2D(width, height, frameToRender.stride(kYPlane), |
+ frameToRender.buffer(kYPlane)); |
+ |
+ glActiveTexture(GL_TEXTURE1); |
+ glBindTexture(GL_TEXTURE_2D, _textureIds[1]); |
+ GlTexSubImage2D(width / 2, height / 2, frameToRender.stride(kUPlane), |
+ frameToRender.buffer(kUPlane)); |
+ |
+ glActiveTexture(GL_TEXTURE2); |
+ glBindTexture(GL_TEXTURE_2D, _textureIds[2]); |
+ GlTexSubImage2D(width / 2, height / 2, frameToRender.stride(kVPlane), |
+ frameToRender.buffer(kVPlane)); |
+ |
+ checkGlError("UpdateTextures"); |
+} |
+ |
+} // namespace webrtc |