Chromium Code Reviews| OLD | NEW |
|---|---|
| 1 /* | 1 /* |
| 2 * libjingle | 2 * libjingle |
| 3 * Copyright 2015 Google Inc. | 3 * Copyright 2015 Google Inc. |
| 4 * | 4 * |
| 5 * Redistribution and use in source and binary forms, with or without | 5 * Redistribution and use in source and binary forms, with or without |
| 6 * modification, are permitted provided that the following conditions are met: | 6 * modification, are permitted provided that the following conditions are met: |
| 7 * | 7 * |
| 8 * 1. Redistributions of source code must retain the above copyright notice, | 8 * 1. Redistributions of source code must retain the above copyright notice, |
| 9 * this list of conditions and the following disclaimer. | 9 * this list of conditions and the following disclaimer. |
| 10 * 2. Redistributions in binary form must reproduce the above copyright notice, | 10 * 2. Redistributions in binary form must reproduce the above copyright notice, |
| (...skipping 15 matching lines...) Expand all Loading... | |
| 26 * | 26 * |
| 27 */ | 27 */ |
| 28 | 28 |
| 29 #include <algorithm> | 29 #include <algorithm> |
| 30 #include <vector> | 30 #include <vector> |
| 31 | 31 |
| 32 #include "talk/app/webrtc/java/jni/androidmediadecoder_jni.h" | 32 #include "talk/app/webrtc/java/jni/androidmediadecoder_jni.h" |
| 33 #include "talk/app/webrtc/java/jni/androidmediacodeccommon.h" | 33 #include "talk/app/webrtc/java/jni/androidmediacodeccommon.h" |
| 34 #include "talk/app/webrtc/java/jni/classreferenceholder.h" | 34 #include "talk/app/webrtc/java/jni/classreferenceholder.h" |
| 35 #include "talk/app/webrtc/java/jni/native_handle_impl.h" | 35 #include "talk/app/webrtc/java/jni/native_handle_impl.h" |
| 36 #include "talk/app/webrtc/java/jni/surfacetexturehelper_jni.h" | |
| 36 #include "webrtc/base/bind.h" | 37 #include "webrtc/base/bind.h" |
| 37 #include "webrtc/base/checks.h" | 38 #include "webrtc/base/checks.h" |
| 38 #include "webrtc/base/logging.h" | 39 #include "webrtc/base/logging.h" |
| 39 #include "webrtc/base/scoped_ref_ptr.h" | 40 #include "webrtc/base/scoped_ref_ptr.h" |
| 40 #include "webrtc/base/thread.h" | 41 #include "webrtc/base/thread.h" |
| 41 #include "webrtc/base/timeutils.h" | 42 #include "webrtc/base/timeutils.h" |
| 42 #include "webrtc/common_video/interface/i420_buffer_pool.h" | 43 #include "webrtc/common_video/interface/i420_buffer_pool.h" |
| 43 #include "webrtc/modules/video_coding/codecs/interface/video_codec_interface.h" | 44 #include "webrtc/modules/video_coding/codecs/interface/video_codec_interface.h" |
| 44 #include "webrtc/system_wrappers/include/logcat_trace_context.h" | 45 #include "webrtc/system_wrappers/include/logcat_trace_context.h" |
| 45 #include "webrtc/system_wrappers/include/tick_util.h" | 46 #include "webrtc/system_wrappers/include/tick_util.h" |
| (...skipping 58 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
| 104 | 105 |
| 105 // Type of video codec. | 106 // Type of video codec. |
| 106 VideoCodecType codecType_; | 107 VideoCodecType codecType_; |
| 107 | 108 |
| 108 bool key_frame_required_; | 109 bool key_frame_required_; |
| 109 bool inited_; | 110 bool inited_; |
| 110 bool sw_fallback_required_; | 111 bool sw_fallback_required_; |
| 111 bool use_surface_; | 112 bool use_surface_; |
| 112 VideoCodec codec_; | 113 VideoCodec codec_; |
| 113 webrtc::I420BufferPool decoded_frame_pool_; | 114 webrtc::I420BufferPool decoded_frame_pool_; |
| 114 NativeHandleImpl native_handle_; | 115 rtc::scoped_refptr<SurfaceTextureHelper> surface_texture_helper_; |
| 115 DecodedImageCallback* callback_; | 116 DecodedImageCallback* callback_; |
| 116 int frames_received_; // Number of frames received by decoder. | 117 int frames_received_; // Number of frames received by decoder. |
| 117 int frames_decoded_; // Number of frames decoded by decoder. | 118 int frames_decoded_; // Number of frames decoded by decoder. |
| 118 int64_t start_time_ms_; // Start time for statistics. | 119 int64_t start_time_ms_; // Start time for statistics. |
| 119 int current_frames_; // Number of frames in the current statistics interval. | 120 int current_frames_; // Number of frames in the current statistics interval. |
| 120 int current_bytes_; // Encoded bytes in the current statistics interval. | 121 int current_bytes_; // Encoded bytes in the current statistics interval. |
| 121 int current_decoding_time_ms_; // Overall decoding time in the current second | 122 int current_decoding_time_ms_; // Overall decoding time in the current second |
| 122 uint32_t max_pending_frames_; // Maximum number of pending input frames | 123 uint32_t max_pending_frames_; // Maximum number of pending input frames |
| 123 std::vector<int32_t> timestamps_; | 124 std::vector<int32_t> timestamps_; |
| 124 std::vector<int64_t> ntp_times_ms_; | 125 std::vector<int64_t> ntp_times_ms_; |
| 125 std::vector<int64_t> frame_rtc_times_ms_; // Time when video frame is sent to | 126 std::vector<int64_t> frame_rtc_times_ms_; // Time when video frame is sent to |
| 126 // decoder input. | 127 // decoder input. |
| 127 | 128 |
| 128 // State that is constant for the lifetime of this object once the ctor | 129 // State that is constant for the lifetime of this object once the ctor |
| 129 // returns. | 130 // returns. |
| 130 scoped_ptr<Thread> codec_thread_; // Thread on which to operate MediaCodec. | 131 scoped_ptr<Thread> codec_thread_; // Thread on which to operate MediaCodec. |
| 131 ScopedGlobalRef<jclass> j_media_codec_video_decoder_class_; | 132 ScopedGlobalRef<jclass> j_media_codec_video_decoder_class_; |
| 132 ScopedGlobalRef<jobject> j_media_codec_video_decoder_; | 133 ScopedGlobalRef<jobject> j_media_codec_video_decoder_; |
| 133 jmethodID j_init_decode_method_; | 134 jmethodID j_init_decode_method_; |
| 134 jmethodID j_release_method_; | 135 jmethodID j_release_method_; |
| 135 jmethodID j_dequeue_input_buffer_method_; | 136 jmethodID j_dequeue_input_buffer_method_; |
| 136 jmethodID j_queue_input_buffer_method_; | 137 jmethodID j_queue_input_buffer_method_; |
| 137 jmethodID j_dequeue_output_buffer_method_; | 138 jmethodID j_dequeue_byte_buffer_method_; |
| 139 jmethodID j_dequeue_texture_buffer_method_; | |
| 138 jmethodID j_return_decoded_byte_buffer_method_; | 140 jmethodID j_return_decoded_byte_buffer_method_; |
| 139 // MediaCodecVideoDecoder fields. | 141 // MediaCodecVideoDecoder fields. |
| 140 jfieldID j_input_buffers_field_; | 142 jfieldID j_input_buffers_field_; |
| 141 jfieldID j_output_buffers_field_; | 143 jfieldID j_output_buffers_field_; |
| 142 jfieldID j_color_format_field_; | 144 jfieldID j_color_format_field_; |
| 143 jfieldID j_width_field_; | 145 jfieldID j_width_field_; |
| 144 jfieldID j_height_field_; | 146 jfieldID j_height_field_; |
| 145 jfieldID j_stride_field_; | 147 jfieldID j_stride_field_; |
| 146 jfieldID j_slice_height_field_; | 148 jfieldID j_slice_height_field_; |
| 147 jfieldID j_surface_texture_field_; | |
| 148 // MediaCodecVideoDecoder.DecodedTextureBuffer fields. | 149 // MediaCodecVideoDecoder.DecodedTextureBuffer fields. |
| 149 jfieldID j_textureID_field_; | 150 jfieldID j_texture_id_field_; |
| 151 jfieldID j_transform_matrix_field_; | |
| 150 jfieldID j_texture_presentation_timestamp_us_field_; | 152 jfieldID j_texture_presentation_timestamp_us_field_; |
| 151 // MediaCodecVideoDecoder.DecodedByteBuffer fields. | 153 jfieldID j_texture_decode_time_ms_field_; |
| 154 // MediaCodecVideoDecoder.DecodedOutputBuffer fields. | |
| 152 jfieldID j_info_index_field_; | 155 jfieldID j_info_index_field_; |
| 153 jfieldID j_info_offset_field_; | 156 jfieldID j_info_offset_field_; |
| 154 jfieldID j_info_size_field_; | 157 jfieldID j_info_size_field_; |
| 155 jfieldID j_info_presentation_timestamp_us_field_; | 158 jfieldID j_info_presentation_timestamp_us_field_; |
| 159 jfieldID j_byte_buffer_decode_time_ms_field_; | |
| 156 | 160 |
| 157 // Global references; must be deleted in Release(). | 161 // Global references; must be deleted in Release(). |
| 158 std::vector<jobject> input_buffers_; | 162 std::vector<jobject> input_buffers_; |
| 159 jobject surface_texture_; | |
| 160 jobject previous_surface_texture_; | |
| 161 | 163 |
| 162 // Render EGL context - owned by factory, should not be allocated/destroyed | 164 // Render EGL context - owned by factory, should not be allocated/destroyed |
| 163 // by VideoDecoder. | 165 // by VideoDecoder. |
| 164 jobject render_egl_context_; | 166 jobject render_egl_context_; |
| 165 }; | 167 }; |
| 166 | 168 |
| 167 MediaCodecVideoDecoder::MediaCodecVideoDecoder( | 169 MediaCodecVideoDecoder::MediaCodecVideoDecoder( |
| 168 JNIEnv* jni, VideoCodecType codecType, jobject render_egl_context) : | 170 JNIEnv* jni, VideoCodecType codecType, jobject render_egl_context) : |
| 169 codecType_(codecType), | 171 codecType_(codecType), |
| 170 render_egl_context_(render_egl_context), | 172 render_egl_context_(render_egl_context), |
| 171 key_frame_required_(true), | 173 key_frame_required_(true), |
| 172 inited_(false), | 174 inited_(false), |
| 173 sw_fallback_required_(false), | 175 sw_fallback_required_(false), |
| 174 surface_texture_(NULL), | |
| 175 previous_surface_texture_(NULL), | |
| 176 codec_thread_(new Thread()), | 176 codec_thread_(new Thread()), |
| 177 j_media_codec_video_decoder_class_( | 177 j_media_codec_video_decoder_class_( |
| 178 jni, | 178 jni, |
| 179 FindClass(jni, "org/webrtc/MediaCodecVideoDecoder")), | 179 FindClass(jni, "org/webrtc/MediaCodecVideoDecoder")), |
| 180 j_media_codec_video_decoder_( | 180 j_media_codec_video_decoder_( |
| 181 jni, | 181 jni, |
| 182 jni->NewObject(*j_media_codec_video_decoder_class_, | 182 jni->NewObject(*j_media_codec_video_decoder_class_, |
| 183 GetMethodID(jni, | 183 GetMethodID(jni, |
| 184 *j_media_codec_video_decoder_class_, | 184 *j_media_codec_video_decoder_class_, |
| 185 "<init>", | 185 "<init>", |
| 186 "()V"))) { | 186 "()V"))) { |
| 187 ScopedLocalRefFrame local_ref_frame(jni); | 187 ScopedLocalRefFrame local_ref_frame(jni); |
| 188 codec_thread_->SetName("MediaCodecVideoDecoder", NULL); | 188 codec_thread_->SetName("MediaCodecVideoDecoder", NULL); |
| 189 RTC_CHECK(codec_thread_->Start()) << "Failed to start MediaCodecVideoDecoder"; | 189 RTC_CHECK(codec_thread_->Start()) << "Failed to start MediaCodecVideoDecoder"; |
| 190 | 190 |
| 191 j_init_decode_method_ = GetMethodID( | 191 j_init_decode_method_ = GetMethodID( |
| 192 jni, *j_media_codec_video_decoder_class_, "initDecode", | 192 jni, *j_media_codec_video_decoder_class_, "initDecode", |
| 193 "(Lorg/webrtc/MediaCodecVideoDecoder$VideoCodecType;" | 193 "(Lorg/webrtc/MediaCodecVideoDecoder$VideoCodecType;" |
| 194 "IILjavax/microedition/khronos/egl/EGLContext;)Z"); | 194 "IILorg/webrtc/SurfaceTextureHelper;)Z"); |
| 195 j_release_method_ = | 195 j_release_method_ = |
| 196 GetMethodID(jni, *j_media_codec_video_decoder_class_, "release", "()V"); | 196 GetMethodID(jni, *j_media_codec_video_decoder_class_, "release", "()V"); |
| 197 j_dequeue_input_buffer_method_ = GetMethodID( | 197 j_dequeue_input_buffer_method_ = GetMethodID( |
| 198 jni, *j_media_codec_video_decoder_class_, "dequeueInputBuffer", "()I"); | 198 jni, *j_media_codec_video_decoder_class_, "dequeueInputBuffer", "()I"); |
| 199 j_queue_input_buffer_method_ = GetMethodID( | 199 j_queue_input_buffer_method_ = GetMethodID( |
| 200 jni, *j_media_codec_video_decoder_class_, "queueInputBuffer", "(IIJ)Z"); | 200 jni, *j_media_codec_video_decoder_class_, "queueInputBuffer", "(IIJ)Z"); |
| 201 j_dequeue_output_buffer_method_ = GetMethodID( | 201 j_dequeue_byte_buffer_method_ = GetMethodID( |
| 202 jni, *j_media_codec_video_decoder_class_, "dequeueOutputBuffer", | 202 jni, *j_media_codec_video_decoder_class_, "dequeueOutputBuffer", |
| 203 "(I)Ljava/lang/Object;"); | 203 "(I)Lorg/webrtc/MediaCodecVideoDecoder$DecodedOutputBuffer;"); |
| 204 j_dequeue_texture_buffer_method_ = GetMethodID( | |
| 205 jni, *j_media_codec_video_decoder_class_, "dequeueTextureBuffer", | |
| 206 "(I)Lorg/webrtc/MediaCodecVideoDecoder$DecodedTextureBuffer;"); | |
| 204 j_return_decoded_byte_buffer_method_ = | 207 j_return_decoded_byte_buffer_method_ = |
| 205 GetMethodID(jni, *j_media_codec_video_decoder_class_, | 208 GetMethodID(jni, *j_media_codec_video_decoder_class_, |
| 206 "returnDecodedByteBuffer", "(I)V"); | 209 "returnDecodedOutputBuffer", "(I)V"); |
| 207 | 210 |
| 208 j_input_buffers_field_ = GetFieldID( | 211 j_input_buffers_field_ = GetFieldID( |
| 209 jni, *j_media_codec_video_decoder_class_, | 212 jni, *j_media_codec_video_decoder_class_, |
| 210 "inputBuffers", "[Ljava/nio/ByteBuffer;"); | 213 "inputBuffers", "[Ljava/nio/ByteBuffer;"); |
| 211 j_output_buffers_field_ = GetFieldID( | 214 j_output_buffers_field_ = GetFieldID( |
| 212 jni, *j_media_codec_video_decoder_class_, | 215 jni, *j_media_codec_video_decoder_class_, |
| 213 "outputBuffers", "[Ljava/nio/ByteBuffer;"); | 216 "outputBuffers", "[Ljava/nio/ByteBuffer;"); |
| 214 j_color_format_field_ = GetFieldID( | 217 j_color_format_field_ = GetFieldID( |
| 215 jni, *j_media_codec_video_decoder_class_, "colorFormat", "I"); | 218 jni, *j_media_codec_video_decoder_class_, "colorFormat", "I"); |
| 216 j_width_field_ = GetFieldID( | 219 j_width_field_ = GetFieldID( |
| 217 jni, *j_media_codec_video_decoder_class_, "width", "I"); | 220 jni, *j_media_codec_video_decoder_class_, "width", "I"); |
| 218 j_height_field_ = GetFieldID( | 221 j_height_field_ = GetFieldID( |
| 219 jni, *j_media_codec_video_decoder_class_, "height", "I"); | 222 jni, *j_media_codec_video_decoder_class_, "height", "I"); |
| 220 j_stride_field_ = GetFieldID( | 223 j_stride_field_ = GetFieldID( |
| 221 jni, *j_media_codec_video_decoder_class_, "stride", "I"); | 224 jni, *j_media_codec_video_decoder_class_, "stride", "I"); |
| 222 j_slice_height_field_ = GetFieldID( | 225 j_slice_height_field_ = GetFieldID( |
| 223 jni, *j_media_codec_video_decoder_class_, "sliceHeight", "I"); | 226 jni, *j_media_codec_video_decoder_class_, "sliceHeight", "I"); |
| 224 j_surface_texture_field_ = GetFieldID( | |
| 225 jni, *j_media_codec_video_decoder_class_, "surfaceTexture", | |
| 226 "Landroid/graphics/SurfaceTexture;"); | |
| 227 | 227 |
| 228 jclass j_decoder_decoded_texture_buffer_class = FindClass(jni, | 228 jclass j_decoded_texture_buffer_class = FindClass(jni, |
| 229 "org/webrtc/MediaCodecVideoDecoder$DecodedTextureBuffer"); | 229 "org/webrtc/MediaCodecVideoDecoder$DecodedTextureBuffer"); |
| 230 j_textureID_field_ = GetFieldID( | 230 j_texture_id_field_ = GetFieldID( |
| 231 jni, j_decoder_decoded_texture_buffer_class, "textureID", "I"); | 231 jni, j_decoded_texture_buffer_class, "textureID", "I"); |
| 232 j_texture_presentation_timestamp_us_field_ = | 232 j_transform_matrix_field_ = GetFieldID( |
| 233 GetFieldID(jni, j_decoder_decoded_texture_buffer_class, | 233 jni, j_decoded_texture_buffer_class, "transformMatrix", "[F"); |
| 234 "presentationTimestampUs", "J"); | 234 j_texture_presentation_timestamp_us_field_ = GetFieldID( |
| 235 jni, j_decoded_texture_buffer_class, "presentationTimestampUs", "J"); | |
| 236 j_texture_decode_time_ms_field_ = GetFieldID( | |
| 237 jni, j_decoded_texture_buffer_class, "decodeTimeMs", "J"); | |
| 235 | 238 |
| 236 jclass j_decoder_decoded_byte_buffer_class = FindClass(jni, | 239 jclass j_decoded_output_buffer_class = FindClass(jni, |
| 237 "org/webrtc/MediaCodecVideoDecoder$DecodedByteBuffer"); | 240 "org/webrtc/MediaCodecVideoDecoder$DecodedOutputBuffer"); |
| 238 j_info_index_field_ = GetFieldID( | 241 j_info_index_field_ = GetFieldID( |
| 239 jni, j_decoder_decoded_byte_buffer_class, "index", "I"); | 242 jni, j_decoded_output_buffer_class, "index", "I"); |
| 240 j_info_offset_field_ = GetFieldID( | 243 j_info_offset_field_ = GetFieldID( |
| 241 jni, j_decoder_decoded_byte_buffer_class, "offset", "I"); | 244 jni, j_decoded_output_buffer_class, "offset", "I"); |
| 242 j_info_size_field_ = GetFieldID( | 245 j_info_size_field_ = GetFieldID( |
| 243 jni, j_decoder_decoded_byte_buffer_class, "size", "I"); | 246 jni, j_decoded_output_buffer_class, "size", "I"); |
| 244 j_info_presentation_timestamp_us_field_ = GetFieldID( | 247 j_info_presentation_timestamp_us_field_ = GetFieldID( |
| 245 jni, j_decoder_decoded_byte_buffer_class, "presentationTimestampUs", "J"); | 248 jni, j_decoded_output_buffer_class, "presentationTimestampUs", "J"); |
| 249 j_byte_buffer_decode_time_ms_field_ = GetFieldID( | |
| 250 jni, j_decoded_output_buffer_class, "decodeTimeMs", "J"); | |
| 246 | 251 |
| 247 CHECK_EXCEPTION(jni) << "MediaCodecVideoDecoder ctor failed"; | 252 CHECK_EXCEPTION(jni) << "MediaCodecVideoDecoder ctor failed"; |
| 248 use_surface_ = (render_egl_context_ != NULL); | 253 use_surface_ = (render_egl_context_ != NULL); |
| 249 ALOGD << "MediaCodecVideoDecoder ctor. Use surface: " << use_surface_; | 254 ALOGD << "MediaCodecVideoDecoder ctor. Use surface: " << use_surface_; |
| 250 memset(&codec_, 0, sizeof(codec_)); | 255 memset(&codec_, 0, sizeof(codec_)); |
| 251 AllowBlockingCalls(); | 256 AllowBlockingCalls(); |
| 252 } | 257 } |
| 253 | 258 |
| 254 MediaCodecVideoDecoder::~MediaCodecVideoDecoder() { | 259 MediaCodecVideoDecoder::~MediaCodecVideoDecoder() { |
| 255 // Call Release() to ensure no more callbacks to us after we are deleted. | 260 // Call Release() to ensure no more callbacks to us after we are deleted. |
| 256 Release(); | 261 Release(); |
| 257 // Delete global references. | |
| 258 JNIEnv* jni = AttachCurrentThreadIfNeeded(); | |
| 259 if (previous_surface_texture_ != NULL) { | |
| 260 jni->DeleteGlobalRef(previous_surface_texture_); | |
| 261 } | |
| 262 if (surface_texture_ != NULL) { | |
| 263 jni->DeleteGlobalRef(surface_texture_); | |
| 264 } | |
| 265 } | 262 } |
| 266 | 263 |
| 267 int32_t MediaCodecVideoDecoder::InitDecode(const VideoCodec* inst, | 264 int32_t MediaCodecVideoDecoder::InitDecode(const VideoCodec* inst, |
| 268 int32_t numberOfCores) { | 265 int32_t numberOfCores) { |
| 269 ALOGD << "InitDecode."; | 266 ALOGD << "InitDecode."; |
| 270 if (inst == NULL) { | 267 if (inst == NULL) { |
| 271 ALOGE << "NULL VideoCodec instance"; | 268 ALOGE << "NULL VideoCodec instance"; |
| 272 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; | 269 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; |
| 273 } | 270 } |
| 274 // Factory should guard against other codecs being used with us. | 271 // Factory should guard against other codecs being used with us. |
| (...skipping 30 matching lines...) Expand all Loading... | |
| 305 ALOGE << "Release failure: " << ret_val << " - fallback to SW codec"; | 302 ALOGE << "Release failure: " << ret_val << " - fallback to SW codec"; |
| 306 sw_fallback_required_ = true; | 303 sw_fallback_required_ = true; |
| 307 return WEBRTC_VIDEO_CODEC_ERROR; | 304 return WEBRTC_VIDEO_CODEC_ERROR; |
| 308 } | 305 } |
| 309 | 306 |
| 310 // Always start with a complete key frame. | 307 // Always start with a complete key frame. |
| 311 key_frame_required_ = true; | 308 key_frame_required_ = true; |
| 312 frames_received_ = 0; | 309 frames_received_ = 0; |
| 313 frames_decoded_ = 0; | 310 frames_decoded_ = 0; |
| 314 | 311 |
| 312 if (use_surface_) { | |
| 313 surface_texture_helper_ = new rtc::RefCountedObject<SurfaceTextureHelper>( | |
| 314 jni, render_egl_context_); | |
| 315 } | |
| 316 | |
| 315 jobject j_video_codec_enum = JavaEnumFromIndex( | 317 jobject j_video_codec_enum = JavaEnumFromIndex( |
| 316 jni, "MediaCodecVideoDecoder$VideoCodecType", codecType_); | 318 jni, "MediaCodecVideoDecoder$VideoCodecType", codecType_); |
| 317 bool success = jni->CallBooleanMethod( | 319 bool success = jni->CallBooleanMethod( |
| 318 *j_media_codec_video_decoder_, | 320 *j_media_codec_video_decoder_, |
| 319 j_init_decode_method_, | 321 j_init_decode_method_, |
| 320 j_video_codec_enum, | 322 j_video_codec_enum, |
| 321 codec_.width, | 323 codec_.width, |
| 322 codec_.height, | 324 codec_.height, |
| 323 use_surface_ ? render_egl_context_ : nullptr); | 325 use_surface_ ? surface_texture_helper_->GetJavaSurfaceTextureHelper() |
| 326 : nullptr); | |
| 324 if (CheckException(jni) || !success) { | 327 if (CheckException(jni) || !success) { |
| 325 ALOGE << "Codec initialization error - fallback to SW codec."; | 328 ALOGE << "Codec initialization error - fallback to SW codec."; |
| 326 sw_fallback_required_ = true; | 329 sw_fallback_required_ = true; |
| 327 return WEBRTC_VIDEO_CODEC_ERROR; | 330 return WEBRTC_VIDEO_CODEC_ERROR; |
| 328 } | 331 } |
| 329 inited_ = true; | 332 inited_ = true; |
| 330 | 333 |
| 331 switch (codecType_) { | 334 switch (codecType_) { |
| 332 case kVideoCodecVP8: | 335 case kVideoCodecVP8: |
| 333 max_pending_frames_ = kMaxPendingFramesVp8; | 336 max_pending_frames_ = kMaxPendingFramesVp8; |
| (...skipping 20 matching lines...) Expand all Loading... | |
| 354 for (size_t i = 0; i < num_input_buffers; ++i) { | 357 for (size_t i = 0; i < num_input_buffers; ++i) { |
| 355 input_buffers_[i] = | 358 input_buffers_[i] = |
| 356 jni->NewGlobalRef(jni->GetObjectArrayElement(input_buffers, i)); | 359 jni->NewGlobalRef(jni->GetObjectArrayElement(input_buffers, i)); |
| 357 if (CheckException(jni)) { | 360 if (CheckException(jni)) { |
| 358 ALOGE << "NewGlobalRef error - fallback to SW codec."; | 361 ALOGE << "NewGlobalRef error - fallback to SW codec."; |
| 359 sw_fallback_required_ = true; | 362 sw_fallback_required_ = true; |
| 360 return WEBRTC_VIDEO_CODEC_ERROR; | 363 return WEBRTC_VIDEO_CODEC_ERROR; |
| 361 } | 364 } |
| 362 } | 365 } |
| 363 | 366 |
| 364 if (use_surface_) { | |
| 365 jobject surface_texture = GetObjectField( | |
| 366 jni, *j_media_codec_video_decoder_, j_surface_texture_field_); | |
| 367 if (previous_surface_texture_ != NULL) { | |
| 368 jni->DeleteGlobalRef(previous_surface_texture_); | |
| 369 } | |
| 370 previous_surface_texture_ = surface_texture_; | |
| 371 surface_texture_ = jni->NewGlobalRef(surface_texture); | |
| 372 } | |
| 373 codec_thread_->PostDelayed(kMediaCodecPollMs, this); | 367 codec_thread_->PostDelayed(kMediaCodecPollMs, this); |
| 374 | 368 |
| 375 return WEBRTC_VIDEO_CODEC_OK; | 369 return WEBRTC_VIDEO_CODEC_OK; |
| 376 } | 370 } |
| 377 | 371 |
| 378 int32_t MediaCodecVideoDecoder::Release() { | 372 int32_t MediaCodecVideoDecoder::Release() { |
| 379 ALOGD << "DecoderRelease request"; | 373 ALOGD << "DecoderRelease request"; |
| 380 return codec_thread_->Invoke<int32_t>( | 374 return codec_thread_->Invoke<int32_t>( |
| 381 Bind(&MediaCodecVideoDecoder::ReleaseOnCodecThread, this)); | 375 Bind(&MediaCodecVideoDecoder::ReleaseOnCodecThread, this)); |
| 382 } | 376 } |
| 383 | 377 |
| 384 int32_t MediaCodecVideoDecoder::ReleaseOnCodecThread() { | 378 int32_t MediaCodecVideoDecoder::ReleaseOnCodecThread() { |
| 385 if (!inited_) { | 379 if (!inited_) { |
| 386 return WEBRTC_VIDEO_CODEC_OK; | 380 return WEBRTC_VIDEO_CODEC_OK; |
| 387 } | 381 } |
| 388 CheckOnCodecThread(); | 382 CheckOnCodecThread(); |
| 389 JNIEnv* jni = AttachCurrentThreadIfNeeded(); | 383 JNIEnv* jni = AttachCurrentThreadIfNeeded(); |
| 390 ALOGD << "DecoderReleaseOnCodecThread: Frames received: " << | 384 ALOGD << "DecoderReleaseOnCodecThread: Frames received: " << |
| 391 frames_received_ << ". Frames decoded: " << frames_decoded_; | 385 frames_received_ << ". Frames decoded: " << frames_decoded_; |
| 392 ScopedLocalRefFrame local_ref_frame(jni); | 386 ScopedLocalRefFrame local_ref_frame(jni); |
| 393 for (size_t i = 0; i < input_buffers_.size(); i++) { | 387 for (size_t i = 0; i < input_buffers_.size(); i++) { |
| 394 jni->DeleteGlobalRef(input_buffers_[i]); | 388 jni->DeleteGlobalRef(input_buffers_[i]); |
| 395 } | 389 } |
| 396 input_buffers_.clear(); | 390 input_buffers_.clear(); |
| 397 jni->CallVoidMethod(*j_media_codec_video_decoder_, j_release_method_); | 391 jni->CallVoidMethod(*j_media_codec_video_decoder_, j_release_method_); |
| 392 surface_texture_helper_ = nullptr; | |
| 398 inited_ = false; | 393 inited_ = false; |
| 399 rtc::MessageQueueManager::Clear(this); | 394 rtc::MessageQueueManager::Clear(this); |
| 400 if (CheckException(jni)) { | 395 if (CheckException(jni)) { |
| 401 ALOGE << "Decoder release exception"; | 396 ALOGE << "Decoder release exception"; |
| 402 return WEBRTC_VIDEO_CODEC_ERROR; | 397 return WEBRTC_VIDEO_CODEC_ERROR; |
| 403 } | 398 } |
| 404 ALOGD << "DecoderReleaseOnCodecThread done"; | 399 ALOGD << "DecoderReleaseOnCodecThread done"; |
| 405 return WEBRTC_VIDEO_CODEC_OK; | 400 return WEBRTC_VIDEO_CODEC_OK; |
| 406 } | 401 } |
| 407 | 402 |
| (...skipping 88 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
| 496 int32_t MediaCodecVideoDecoder::DecodeOnCodecThread( | 491 int32_t MediaCodecVideoDecoder::DecodeOnCodecThread( |
| 497 const EncodedImage& inputImage) { | 492 const EncodedImage& inputImage) { |
| 498 CheckOnCodecThread(); | 493 CheckOnCodecThread(); |
| 499 JNIEnv* jni = AttachCurrentThreadIfNeeded(); | 494 JNIEnv* jni = AttachCurrentThreadIfNeeded(); |
| 500 ScopedLocalRefFrame local_ref_frame(jni); | 495 ScopedLocalRefFrame local_ref_frame(jni); |
| 501 | 496 |
| 502 // Try to drain the decoder and wait until output is not too | 497 // Try to drain the decoder and wait until output is not too |
| 503 // much behind the input. | 498 // much behind the input. |
| 504 if (frames_received_ > frames_decoded_ + max_pending_frames_) { | 499 if (frames_received_ > frames_decoded_ + max_pending_frames_) { |
| 505 ALOGV("Received: %d. Decoded: %d. Wait for output...", | 500 ALOGV("Received: %d. Decoded: %d. Wait for output...", |
| 506 frames_received_, frames_decoded_); | 501 frames_received_, frames_decoded_); |
|
magjed_webrtc
2015/10/30 12:28:10
Revert indentation change?
perkj_webrtc
2015/11/06 14:31:12
Done.
| |
| 507 if (!DeliverPendingOutputs(jni, kMediaCodecTimeoutMs * 1000)) { | 502 if (!DeliverPendingOutputs(jni, kMediaCodecTimeoutMs)) { |
| 508 ALOGE << "DeliverPendingOutputs error. Frames received: " << | 503 ALOGE << "DeliverPendingOutputs error. Frames received: " << |
| 509 frames_received_ << ". Frames decoded: " << frames_decoded_; | 504 frames_received_ << ". Frames decoded: " << frames_decoded_; |
| 510 return ProcessHWErrorOnCodecThread(); | 505 return ProcessHWErrorOnCodecThread(); |
| 511 } | 506 } |
| 512 if (frames_received_ > frames_decoded_ + max_pending_frames_) { | 507 if (frames_received_ > frames_decoded_ + max_pending_frames_) { |
| 513 ALOGE << "Output buffer dequeue timeout. Frames received: " << | 508 ALOGE << "Output buffer dequeue timeout. Frames received: " << |
| 514 frames_received_ << ". Frames decoded: " << frames_decoded_; | 509 frames_received_ << ". Frames decoded: " << frames_decoded_; |
| 515 return ProcessHWErrorOnCodecThread(); | 510 return ProcessHWErrorOnCodecThread(); |
| 516 } | 511 } |
| 517 } | 512 } |
| (...skipping 47 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
| 565 // Try to drain the decoder | 560 // Try to drain the decoder |
| 566 if (!DeliverPendingOutputs(jni, 0)) { | 561 if (!DeliverPendingOutputs(jni, 0)) { |
| 567 ALOGE << "DeliverPendingOutputs error"; | 562 ALOGE << "DeliverPendingOutputs error"; |
| 568 return ProcessHWErrorOnCodecThread(); | 563 return ProcessHWErrorOnCodecThread(); |
| 569 } | 564 } |
| 570 | 565 |
| 571 return WEBRTC_VIDEO_CODEC_OK; | 566 return WEBRTC_VIDEO_CODEC_OK; |
| 572 } | 567 } |
| 573 | 568 |
| 574 bool MediaCodecVideoDecoder::DeliverPendingOutputs( | 569 bool MediaCodecVideoDecoder::DeliverPendingOutputs( |
| 575 JNIEnv* jni, int dequeue_timeout_us) { | 570 JNIEnv* jni, int dequeue_timeout_ms) { |
| 576 if (frames_received_ <= frames_decoded_) { | 571 if (frames_received_ <= frames_decoded_) { |
| 577 // No need to query for output buffers - decoder is drained. | 572 // No need to query for output buffers - decoder is drained. |
| 578 return true; | 573 return true; |
| 579 } | 574 } |
| 580 // Get decoder output. | 575 // Get decoder output. |
| 581 jobject j_decoder_output_buffer = jni->CallObjectMethod( | 576 jobject j_decoder_output_buffer = |
| 582 *j_media_codec_video_decoder_, | 577 jni->CallObjectMethod(*j_media_codec_video_decoder_, |
| 583 j_dequeue_output_buffer_method_, | 578 use_surface_ ? j_dequeue_texture_buffer_method_ |
| 584 dequeue_timeout_us); | 579 : j_dequeue_byte_buffer_method_, |
| 580 dequeue_timeout_ms); | |
| 581 | |
| 585 if (CheckException(jni)) { | 582 if (CheckException(jni)) { |
| 586 ALOGE << "dequeueOutputBuffer() error"; | 583 ALOGE << "dequeueOutputBuffer() error"; |
| 587 return false; | 584 return false; |
| 588 } | 585 } |
| 589 if (IsNull(jni, j_decoder_output_buffer)) { | 586 if (IsNull(jni, j_decoder_output_buffer)) { |
| 590 // No decoded frame ready. | 587 // No decoded frame ready. |
| 591 return true; | 588 return true; |
| 592 } | 589 } |
| 593 | 590 |
| 594 // Get decoded video frame properties. | 591 // Get decoded video frame properties. |
| 595 int color_format = GetIntField(jni, *j_media_codec_video_decoder_, | 592 int color_format = GetIntField(jni, *j_media_codec_video_decoder_, |
| 596 j_color_format_field_); | 593 j_color_format_field_); |
| 597 int width = GetIntField(jni, *j_media_codec_video_decoder_, j_width_field_); | 594 int width = GetIntField(jni, *j_media_codec_video_decoder_, j_width_field_); |
| 598 int height = GetIntField(jni, *j_media_codec_video_decoder_, j_height_field_); | 595 int height = GetIntField(jni, *j_media_codec_video_decoder_, j_height_field_); |
| 599 int stride = GetIntField(jni, *j_media_codec_video_decoder_, j_stride_field_); | 596 int stride = GetIntField(jni, *j_media_codec_video_decoder_, j_stride_field_); |
| 600 int slice_height = GetIntField(jni, *j_media_codec_video_decoder_, | 597 int slice_height = GetIntField(jni, *j_media_codec_video_decoder_, |
| 601 j_slice_height_field_); | 598 j_slice_height_field_); |
| 602 | 599 |
| 603 rtc::scoped_refptr<webrtc::VideoFrameBuffer> frame_buffer; | 600 rtc::scoped_refptr<webrtc::VideoFrameBuffer> frame_buffer; |
| 604 long output_timestamps_ms = 0; | 601 int64_t output_timestamps_ms = 0; |
| 602 int decode_time_ms = 0; | |
| 605 if (use_surface_) { | 603 if (use_surface_) { |
| 606 // Extract data from Java DecodedTextureBuffer. | 604 // Extract data from Java DecodedTextureBuffer. |
| 607 const int texture_id = | 605 const int texture_id = |
| 608 GetIntField(jni, j_decoder_output_buffer, j_textureID_field_); | 606 GetIntField(jni, j_decoder_output_buffer, j_texture_id_field_); |
| 607 const jfloatArray j_transform_matrix = | |
| 608 reinterpret_cast<jfloatArray>(GetObjectField( | |
| 609 jni, j_decoder_output_buffer, j_transform_matrix_field_)); | |
| 609 const int64_t timestamp_us = | 610 const int64_t timestamp_us = |
| 610 GetLongField(jni, j_decoder_output_buffer, | 611 GetLongField(jni, j_decoder_output_buffer, |
| 611 j_texture_presentation_timestamp_us_field_); | 612 j_texture_presentation_timestamp_us_field_); |
| 612 output_timestamps_ms = timestamp_us / rtc::kNumMicrosecsPerMillisec; | 613 output_timestamps_ms = timestamp_us / rtc::kNumMicrosecsPerMillisec; |
| 614 decode_time_ms = GetLongField(jni, j_decoder_output_buffer, | |
| 615 j_texture_decode_time_ms_field_); | |
| 616 | |
| 613 // Create webrtc::VideoFrameBuffer with native texture handle. | 617 // Create webrtc::VideoFrameBuffer with native texture handle. |
| 614 native_handle_.SetTextureObject(surface_texture_, texture_id); | 618 frame_buffer = surface_texture_helper_->CreateTextureFrame( |
| 615 frame_buffer = new rtc::RefCountedObject<JniNativeHandleBuffer>( | 619 width, height, NativeHandleImpl(jni, texture_id, j_transform_matrix)); |
| 616 &native_handle_, width, height); | |
| 617 } else { | 620 } else { |
| 618 // Extract data from Java ByteBuffer and create output yuv420 frame - | 621 // Extract data from Java ByteBuffer and create output yuv420 frame - |
| 619 // for non surface decoding only. | 622 // for non surface decoding only. |
| 620 const int output_buffer_index = | 623 const int output_buffer_index = |
| 621 GetIntField(jni, j_decoder_output_buffer, j_info_index_field_); | 624 GetIntField(jni, j_decoder_output_buffer, j_info_index_field_); |
| 622 const int output_buffer_offset = | 625 const int output_buffer_offset = |
| 623 GetIntField(jni, j_decoder_output_buffer, j_info_offset_field_); | 626 GetIntField(jni, j_decoder_output_buffer, j_info_offset_field_); |
| 624 const int output_buffer_size = | 627 const int output_buffer_size = |
| 625 GetIntField(jni, j_decoder_output_buffer, j_info_size_field_); | 628 GetIntField(jni, j_decoder_output_buffer, j_info_size_field_); |
| 626 const int64_t timestamp_us = GetLongField( | 629 const int64_t timestamp_us = GetLongField( |
| 627 jni, j_decoder_output_buffer, j_info_presentation_timestamp_us_field_); | 630 jni, j_decoder_output_buffer, j_info_presentation_timestamp_us_field_); |
| 628 output_timestamps_ms = timestamp_us / rtc::kNumMicrosecsPerMillisec; | 631 output_timestamps_ms = timestamp_us / rtc::kNumMicrosecsPerMillisec; |
| 632 decode_time_ms = GetLongField(jni, j_decoder_output_buffer, | |
| 633 j_byte_buffer_decode_time_ms_field_); | |
| 629 | 634 |
| 630 if (output_buffer_size < width * height * 3 / 2) { | 635 if (output_buffer_size < width * height * 3 / 2) { |
| 631 ALOGE << "Insufficient output buffer size: " << output_buffer_size; | 636 ALOGE << "Insufficient output buffer size: " << output_buffer_size; |
| 632 return false; | 637 return false; |
| 633 } | 638 } |
| 634 jobjectArray output_buffers = reinterpret_cast<jobjectArray>(GetObjectField( | 639 jobjectArray output_buffers = reinterpret_cast<jobjectArray>(GetObjectField( |
| 635 jni, *j_media_codec_video_decoder_, j_output_buffers_field_)); | 640 jni, *j_media_codec_video_decoder_, j_output_buffers_field_)); |
| 636 jobject output_buffer = | 641 jobject output_buffer = |
| 637 jni->GetObjectArrayElement(output_buffers, output_buffer_index); | 642 jni->GetObjectArrayElement(output_buffers, output_buffer_index); |
| 638 uint8_t* payload = reinterpret_cast<uint8_t*>(jni->GetDirectBufferAddress( | 643 uint8_t* payload = reinterpret_cast<uint8_t*>(jni->GetDirectBufferAddress( |
| (...skipping 37 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
| 676 frame_buffer->MutableData(webrtc::kVPlane), | 681 frame_buffer->MutableData(webrtc::kVPlane), |
| 677 frame_buffer->stride(webrtc::kVPlane), | 682 frame_buffer->stride(webrtc::kVPlane), |
| 678 width, height); | 683 width, height); |
| 679 } | 684 } |
| 680 // Return output byte buffer back to codec. | 685 // Return output byte buffer back to codec. |
| 681 jni->CallVoidMethod( | 686 jni->CallVoidMethod( |
| 682 *j_media_codec_video_decoder_, | 687 *j_media_codec_video_decoder_, |
| 683 j_return_decoded_byte_buffer_method_, | 688 j_return_decoded_byte_buffer_method_, |
| 684 output_buffer_index); | 689 output_buffer_index); |
| 685 if (CheckException(jni)) { | 690 if (CheckException(jni)) { |
| 686 ALOGE << "returnDecodedByteBuffer error"; | 691 ALOGE << "returnDecodedOutputBuffer error"; |
| 687 return false; | 692 return false; |
| 688 } | 693 } |
| 689 } | 694 } |
| 690 VideoFrame decoded_frame(frame_buffer, 0, 0, webrtc::kVideoRotation_0); | 695 VideoFrame decoded_frame(frame_buffer, 0, 0, webrtc::kVideoRotation_0); |
| 691 | 696 |
| 692 // Get frame timestamps from a queue. | 697 // Get frame timestamps from a queue. |
| 693 if (timestamps_.size() > 0) { | 698 if (timestamps_.size() > 0) { |
| 694 decoded_frame.set_timestamp(timestamps_.front()); | 699 decoded_frame.set_timestamp(timestamps_.front()); |
| 695 timestamps_.erase(timestamps_.begin()); | 700 timestamps_.erase(timestamps_.begin()); |
| 696 } | 701 } |
| 697 if (ntp_times_ms_.size() > 0) { | 702 if (ntp_times_ms_.size() > 0) { |
| 698 decoded_frame.set_ntp_time_ms(ntp_times_ms_.front()); | 703 decoded_frame.set_ntp_time_ms(ntp_times_ms_.front()); |
| 699 ntp_times_ms_.erase(ntp_times_ms_.begin()); | 704 ntp_times_ms_.erase(ntp_times_ms_.begin()); |
| 700 } | 705 } |
| 701 int64_t frame_decoding_time_ms = 0; | 706 int64_t frame_delayed_ms = 0; |
| 702 if (frame_rtc_times_ms_.size() > 0) { | 707 if (frame_rtc_times_ms_.size() > 0) { |
| 703 frame_decoding_time_ms = GetCurrentTimeMs() - frame_rtc_times_ms_.front(); | 708 frame_delayed_ms = GetCurrentTimeMs() - frame_rtc_times_ms_.front(); |
| 704 frame_rtc_times_ms_.erase(frame_rtc_times_ms_.begin()); | 709 frame_rtc_times_ms_.erase(frame_rtc_times_ms_.begin()); |
| 705 } | 710 } |
| 706 if (frames_decoded_ < kMaxDecodedLogFrames) { | 711 if (frames_decoded_ < kMaxDecodedLogFrames) { |
| 707 ALOGD << "Decoder frame out # " << frames_decoded_ << ". " << width << | 712 ALOGD << "Decoder frame out # " << frames_decoded_ << ". " << width << |
| 708 " x " << height << ". " << stride << " x " << slice_height << | 713 " x " << height << ". " << stride << " x " << slice_height << |
| 709 ". Color: " << color_format << ". TS:" << (int)output_timestamps_ms << | 714 ". Color: " << color_format << ". TS:" << (int)output_timestamps_ms << |
| 710 ". DecTime: " << (int)frame_decoding_time_ms; | 715 ". DecTime: " << (int)decode_time_ms << |
| 716 ". DelayTime: " << (int)frame_delayed_ms; | |
| 711 } | 717 } |
| 712 | 718 |
| 719 ALOGV("Decoder frame out # %d. %d x %d. %d x %d. Color: 0x%x. TS: %lld." | |
|
magjed_webrtc
2015/10/30 12:28:10
Has this log been removed from master?
AlexG
2015/10/30 21:06:10
I think this can be added back or alternatively in
perkj_webrtc
2015/11/06 14:31:12
Right, this looks like a merge mistake. I will rem
| |
| 720 " DecodeTime: %d. DelayTime: %lld.", frames_decoded_, width, height, | |
| 721 stride, slice_height, color_format, output_timestamps_ms, decode_time_ms, | |
| 722 frame_delayed_ms); | |
| 723 | |
| 713 // Calculate and print decoding statistics - every 3 seconds. | 724 // Calculate and print decoding statistics - every 3 seconds. |
| 714 frames_decoded_++; | 725 frames_decoded_++; |
| 715 current_frames_++; | 726 current_frames_++; |
| 716 current_decoding_time_ms_ += frame_decoding_time_ms; | 727 current_decoding_time_ms_ += decode_time_ms; |
| 717 int statistic_time_ms = GetCurrentTimeMs() - start_time_ms_; | 728 int statistic_time_ms = GetCurrentTimeMs() - start_time_ms_; |
| 718 if (statistic_time_ms >= kMediaCodecStatisticsIntervalMs && | 729 if (statistic_time_ms >= kMediaCodecStatisticsIntervalMs && |
| 719 current_frames_ > 0) { | 730 current_frames_ > 0) { |
| 720 ALOGD << "Decoded frames: " << frames_decoded_ << ". Bitrate: " << | 731 ALOGD << "Decoded frames: " << frames_decoded_ << ". Received frames: " |
| 732 << frames_received_ << ". Bitrate: " << | |
| 721 (current_bytes_ * 8 / statistic_time_ms) << " kbps, fps: " << | 733 (current_bytes_ * 8 / statistic_time_ms) << " kbps, fps: " << |
| 722 ((current_frames_ * 1000 + statistic_time_ms / 2) / statistic_time_ms) | 734 ((current_frames_ * 1000 + statistic_time_ms / 2) / statistic_time_ms) |
| 723 << ". decTime: " << (current_decoding_time_ms_ / current_frames_) << | 735 << ". decTime: " << (current_decoding_time_ms_ / current_frames_) << |
| 724 " for last " << statistic_time_ms << " ms."; | 736 " for last " << statistic_time_ms << " ms."; |
| 725 start_time_ms_ = GetCurrentTimeMs(); | 737 start_time_ms_ = GetCurrentTimeMs(); |
| 726 current_frames_ = 0; | 738 current_frames_ = 0; |
| 727 current_bytes_ = 0; | 739 current_bytes_ = 0; |
| 728 current_decoding_time_ms_ = 0; | 740 current_decoding_time_ms_ = 0; |
| 729 } | 741 } |
| 730 | 742 |
| 731 // Callback - output decoded frame. | 743 // Callback - output decoded frame. |
| 732 const int32_t callback_status = callback_->Decoded(decoded_frame); | 744 const int32_t callback_status = |
| 745 callback_->Decoded(decoded_frame, decode_time_ms); | |
|
AlexG
2015/10/30 21:06:10
I am not sure if this is taken care of, but someti
perkj_webrtc
2015/11/06 14:31:12
So if that is just an initial problem we should re
| |
| 733 if (callback_status > 0) { | 746 if (callback_status > 0) { |
| 734 ALOGE << "callback error"; | 747 ALOGE << "callback error"; |
| 735 } | 748 } |
| 736 | 749 |
| 737 return true; | 750 return true; |
| 738 } | 751 } |
| 739 | 752 |
| 740 int32_t MediaCodecVideoDecoder::RegisterDecodeCompleteCallback( | 753 int32_t MediaCodecVideoDecoder::RegisterDecodeCompleteCallback( |
| 741 DecodedImageCallback* callback) { | 754 DecodedImageCallback* callback) { |
| 742 callback_ = callback; | 755 callback_ = callback; |
| (...skipping 113 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
| 856 } | 869 } |
| 857 | 870 |
| 858 void MediaCodecVideoDecoderFactory::DestroyVideoDecoder( | 871 void MediaCodecVideoDecoderFactory::DestroyVideoDecoder( |
| 859 webrtc::VideoDecoder* decoder) { | 872 webrtc::VideoDecoder* decoder) { |
| 860 ALOGD << "Destroy video decoder."; | 873 ALOGD << "Destroy video decoder."; |
| 861 delete decoder; | 874 delete decoder; |
| 862 } | 875 } |
| 863 | 876 |
| 864 } // namespace webrtc_jni | 877 } // namespace webrtc_jni |
| 865 | 878 |
| OLD | NEW |