OLD | NEW |
1 /* | 1 /* |
2 * Copyright (c) 2014 The WebRTC project authors. All Rights Reserved. | 2 * Copyright (c) 2014 The WebRTC project authors. All Rights Reserved. |
3 * | 3 * |
4 * Use of this source code is governed by a BSD-style license | 4 * Use of this source code is governed by a BSD-style license |
5 * that can be found in the LICENSE file in the root of the source | 5 * that can be found in the LICENSE file in the root of the source |
6 * tree. An additional intellectual property rights grant can be found | 6 * tree. An additional intellectual property rights grant can be found |
7 * in the file PATENTS. All contributing project authors may | 7 * in the file PATENTS. All contributing project authors may |
8 * be found in the AUTHORS file in the root of the source tree. | 8 * be found in the AUTHORS file in the root of the source tree. |
9 */ | 9 */ |
10 | 10 |
(...skipping 275 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
286 | 286 |
287 int dst_width = streaminfos_[stream_idx].width; | 287 int dst_width = streaminfos_[stream_idx].width; |
288 int dst_height = streaminfos_[stream_idx].height; | 288 int dst_height = streaminfos_[stream_idx].height; |
289 // If scaling isn't required, because the input resolution | 289 // If scaling isn't required, because the input resolution |
290 // matches the destination or the input image is empty (e.g. | 290 // matches the destination or the input image is empty (e.g. |
291 // a keyframe request for encoders with internal camera | 291 // a keyframe request for encoders with internal camera |
292 // sources), pass the image on directly. Otherwise, we'll | 292 // sources), pass the image on directly. Otherwise, we'll |
293 // scale it to match what the encoder expects (below). | 293 // scale it to match what the encoder expects (below). |
294 if ((dst_width == src_width && dst_height == src_height) || | 294 if ((dst_width == src_width && dst_height == src_height) || |
295 input_image.IsZeroSize()) { | 295 input_image.IsZeroSize()) { |
296 streaminfos_[stream_idx].encoder->Encode(input_image, codec_specific_info, | 296 int ret = streaminfos_[stream_idx].encoder->Encode( |
297 &stream_frame_types); | 297 input_image, codec_specific_info, &stream_frame_types); |
| 298 if (ret != WEBRTC_VIDEO_CODEC_OK) { |
| 299 return ret; |
| 300 } |
298 } else { | 301 } else { |
299 VideoFrame dst_frame; | 302 VideoFrame dst_frame; |
300 // Making sure that destination frame is of sufficient size. | 303 // Making sure that destination frame is of sufficient size. |
301 // Aligning stride values based on width. | 304 // Aligning stride values based on width. |
302 dst_frame.CreateEmptyFrame(dst_width, dst_height, dst_width, | 305 dst_frame.CreateEmptyFrame(dst_width, dst_height, dst_width, |
303 (dst_width + 1) / 2, (dst_width + 1) / 2); | 306 (dst_width + 1) / 2, (dst_width + 1) / 2); |
304 libyuv::I420Scale(input_image.video_frame_buffer()->DataY(), | 307 libyuv::I420Scale(input_image.video_frame_buffer()->DataY(), |
305 input_image.video_frame_buffer()->StrideY(), | 308 input_image.video_frame_buffer()->StrideY(), |
306 input_image.video_frame_buffer()->DataU(), | 309 input_image.video_frame_buffer()->DataU(), |
307 input_image.video_frame_buffer()->StrideU(), | 310 input_image.video_frame_buffer()->StrideU(), |
308 input_image.video_frame_buffer()->DataV(), | 311 input_image.video_frame_buffer()->DataV(), |
309 input_image.video_frame_buffer()->StrideV(), | 312 input_image.video_frame_buffer()->StrideV(), |
310 src_width, src_height, | 313 src_width, src_height, |
311 dst_frame.video_frame_buffer()->MutableDataY(), | 314 dst_frame.video_frame_buffer()->MutableDataY(), |
312 dst_frame.video_frame_buffer()->StrideY(), | 315 dst_frame.video_frame_buffer()->StrideY(), |
313 dst_frame.video_frame_buffer()->MutableDataU(), | 316 dst_frame.video_frame_buffer()->MutableDataU(), |
314 dst_frame.video_frame_buffer()->StrideU(), | 317 dst_frame.video_frame_buffer()->StrideU(), |
315 dst_frame.video_frame_buffer()->MutableDataV(), | 318 dst_frame.video_frame_buffer()->MutableDataV(), |
316 dst_frame.video_frame_buffer()->StrideV(), | 319 dst_frame.video_frame_buffer()->StrideV(), |
317 dst_width, dst_height, | 320 dst_width, dst_height, |
318 libyuv::kFilterBilinear); | 321 libyuv::kFilterBilinear); |
319 dst_frame.set_timestamp(input_image.timestamp()); | 322 dst_frame.set_timestamp(input_image.timestamp()); |
320 dst_frame.set_render_time_ms(input_image.render_time_ms()); | 323 dst_frame.set_render_time_ms(input_image.render_time_ms()); |
321 streaminfos_[stream_idx].encoder->Encode(dst_frame, codec_specific_info, | 324 int ret = streaminfos_[stream_idx].encoder->Encode( |
322 &stream_frame_types); | 325 dst_frame, codec_specific_info, &stream_frame_types); |
| 326 if (ret != WEBRTC_VIDEO_CODEC_OK) { |
| 327 return ret; |
| 328 } |
323 } | 329 } |
324 } | 330 } |
325 | 331 |
326 return WEBRTC_VIDEO_CODEC_OK; | 332 return WEBRTC_VIDEO_CODEC_OK; |
327 } | 333 } |
328 | 334 |
329 int SimulcastEncoderAdapter::RegisterEncodeCompleteCallback( | 335 int SimulcastEncoderAdapter::RegisterEncodeCompleteCallback( |
330 EncodedImageCallback* callback) { | 336 EncodedImageCallback* callback) { |
331 encoded_complete_callback_ = callback; | 337 encoded_complete_callback_ = callback; |
332 return WEBRTC_VIDEO_CODEC_OK; | 338 return WEBRTC_VIDEO_CODEC_OK; |
(...skipping 175 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
508 if (streaminfos_.size() != 1) | 514 if (streaminfos_.size() != 1) |
509 return false; | 515 return false; |
510 return streaminfos_[0].encoder->SupportsNativeHandle(); | 516 return streaminfos_[0].encoder->SupportsNativeHandle(); |
511 } | 517 } |
512 | 518 |
513 const char* SimulcastEncoderAdapter::ImplementationName() const { | 519 const char* SimulcastEncoderAdapter::ImplementationName() const { |
514 return implementation_name_.c_str(); | 520 return implementation_name_.c_str(); |
515 } | 521 } |
516 | 522 |
517 } // namespace webrtc | 523 } // namespace webrtc |
OLD | NEW |