| Index: webrtc/api/androidvideotracksource.cc
|
| diff --git a/webrtc/api/androidvideotracksource.cc b/webrtc/api/androidvideotracksource.cc
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..c52e4f84ff6e3bdeb66d7aea12022a8ea715dd75
|
| --- /dev/null
|
| +++ b/webrtc/api/androidvideotracksource.cc
|
| @@ -0,0 +1,260 @@
|
| +/*
|
| + * Copyright (c) 2016 The WebRTC project authors. All Rights Reserved.
|
| + *
|
| + * Use of this source code is governed by a BSD-style license
|
| + * that can be found in the LICENSE file in the root of the source
|
| + * tree. An additional intellectual property rights grant can be found
|
| + * in the file PATENTS. All contributing project authors may
|
| + * be found in the AUTHORS file in the root of the source tree.
|
| + */
|
| +
|
| +#include "webrtc/api/androidvideotracksource.h"
|
| +
|
| +#include <utility>
|
| +
|
| +namespace webrtc {
|
| +
|
| +AndroidVideoTrackSource::AndroidVideoTrackSource(rtc::Thread* signaling_thread,
|
| + JNIEnv* jni,
|
| + jobject j_egl_context)
|
| + : signaling_thread_(signaling_thread),
|
| + surface_texture_helper_(webrtc_jni::SurfaceTextureHelper::create(
|
| + jni,
|
| + "Camera SurfaceTextureHelper",
|
| + j_egl_context)) {
|
| + LOG(LS_INFO) << "AndroidVideoTrackSource ctor";
|
| + worker_thread_checker_.DetachFromThread();
|
| + camera_thread_checker_.DetachFromThread();
|
| +}
|
| +
|
| +bool AndroidVideoTrackSource::GetStats(AndroidVideoTrackSource::Stats* stats) {
|
| + rtc::CritScope lock(&stats_crit_);
|
| +
|
| + if (!stats_) {
|
| + return false;
|
| + }
|
| +
|
| + *stats = *stats_;
|
| + return true;
|
| +}
|
| +
|
| +void AndroidVideoTrackSource::SetState(SourceState state) {
|
| + if (rtc::Thread::Current() != signaling_thread_) {
|
| + invoker_.AsyncInvoke<void>(
|
| + RTC_FROM_HERE, signaling_thread_,
|
| + rtc::Bind(&AndroidVideoTrackSource::SetState, this, state));
|
| + return;
|
| + }
|
| +
|
| + if (state_ != state) {
|
| + state_ = state;
|
| + FireOnChanged();
|
| + }
|
| +}
|
| +
|
| +void AndroidVideoTrackSource::AddOrUpdateSink(
|
| + rtc::VideoSinkInterface<cricket::VideoFrame>* sink,
|
| + const rtc::VideoSinkWants& wants) {
|
| + RTC_DCHECK(worker_thread_checker_.CalledOnValidThread());
|
| +
|
| + broadcaster_.AddOrUpdateSink(sink, wants);
|
| + OnSinkWantsChanged(broadcaster_.wants());
|
| +}
|
| +
|
| +void AndroidVideoTrackSource::RemoveSink(
|
| + rtc::VideoSinkInterface<cricket::VideoFrame>* sink) {
|
| + RTC_DCHECK(worker_thread_checker_.CalledOnValidThread());
|
| +
|
| + broadcaster_.RemoveSink(sink);
|
| + OnSinkWantsChanged(broadcaster_.wants());
|
| +}
|
| +
|
| +void AndroidVideoTrackSource::OnSinkWantsChanged(
|
| + const rtc::VideoSinkWants& wants) {
|
| + {
|
| + rtc::CritScope lock(&apply_rotation_crit_);
|
| + apply_rotation_ = wants.rotation_applied;
|
| + }
|
| +
|
| + video_adapter_.OnResolutionRequest(wants.max_pixel_count,
|
| + wants.max_pixel_count_step_up);
|
| +}
|
| +
|
| +void AndroidVideoTrackSource::OnByteBufferFrameCaptured(const void* frame_data,
|
| + int length,
|
| + int width,
|
| + int height,
|
| + int rotation,
|
| + int64_t timestamp_ns) {
|
| + RTC_DCHECK(camera_thread_checker_.CalledOnValidThread());
|
| + RTC_DCHECK(rotation == 0 || rotation == 90 || rotation == 180 ||
|
| + rotation == 270);
|
| +
|
| + int adapted_width;
|
| + int adapted_height;
|
| + int crop_width;
|
| + int crop_height;
|
| + int crop_x;
|
| + int crop_y;
|
| + int64_t translated_camera_time_us;
|
| +
|
| + if (!AdaptFrame(width, height, timestamp_ns / rtc::kNumNanosecsPerMicrosec,
|
| + &adapted_width, &adapted_height, &crop_width, &crop_height,
|
| + &crop_x, &crop_y, &translated_camera_time_us)) {
|
| + return;
|
| + }
|
| +
|
| + int rotated_width = crop_width;
|
| + int rotated_height = crop_height;
|
| +
|
| + rtc::CritScope lock(&apply_rotation_crit_);
|
| + if (apply_rotation_ && (rotation == 90 || rotation == 270)) {
|
| + std::swap(adapted_width, adapted_height);
|
| + std::swap(rotated_width, rotated_height);
|
| + }
|
| +
|
| + rtc::scoped_refptr<webrtc::VideoFrameBuffer> buffer =
|
| + pre_scale_pool_.CreateBuffer(rotated_width, rotated_height);
|
| +
|
| + const uint8_t* y_plane = static_cast<const uint8_t*>(frame_data);
|
| + const uint8_t* uv_plane = y_plane + width * height;
|
| + int uv_width = (width + 1) / 2;
|
| +
|
| + RTC_CHECK_GE(length, width * height + 2 * uv_width * ((height + 1) / 2));
|
| +
|
| + // Can only crop at even pixels.
|
| + crop_x &= ~1;
|
| + crop_y &= ~1;
|
| +
|
| + libyuv::NV12ToI420Rotate(
|
| + y_plane + width * crop_y + crop_x, width,
|
| + uv_plane + uv_width * crop_y + crop_x, width, buffer->MutableDataY(),
|
| + buffer->StrideY(),
|
| + // Swap U and V, since we have NV21, not NV12.
|
| + buffer->MutableDataV(), buffer->StrideV(), buffer->MutableDataU(),
|
| + buffer->StrideU(), crop_width, crop_height,
|
| + static_cast<libyuv::RotationMode>(apply_rotation_ ? rotation : 0));
|
| +
|
| + if (adapted_width != buffer->width() || adapted_height != buffer->height()) {
|
| + rtc::scoped_refptr<webrtc::I420Buffer> scaled_buffer(
|
| + post_scale_pool_.CreateBuffer(adapted_width, adapted_height));
|
| + scaled_buffer->ScaleFrom(buffer);
|
| + buffer = scaled_buffer;
|
| + }
|
| +
|
| + OnFrame(cricket::WebRtcVideoFrame(
|
| + buffer,
|
| + apply_rotation_ ? webrtc::kVideoRotation_0
|
| + : static_cast<webrtc::VideoRotation>(rotation),
|
| + translated_camera_time_us),
|
| + width, height);
|
| +}
|
| +
|
| +void AndroidVideoTrackSource::OnTextureFrameCaptured(
|
| + int width,
|
| + int height,
|
| + int rotation,
|
| + int64_t timestamp_ns,
|
| + const webrtc_jni::NativeHandleImpl& handle) {
|
| + RTC_DCHECK(camera_thread_checker_.CalledOnValidThread());
|
| + RTC_DCHECK(rotation == 0 || rotation == 90 || rotation == 180 ||
|
| + rotation == 270);
|
| +
|
| + int adapted_width;
|
| + int adapted_height;
|
| + int crop_width;
|
| + int crop_height;
|
| + int crop_x;
|
| + int crop_y;
|
| + int64_t translated_camera_time_us;
|
| +
|
| + if (!AdaptFrame(width, height, timestamp_ns / rtc::kNumNanosecsPerMicrosec,
|
| + &adapted_width, &adapted_height, &crop_width, &crop_height,
|
| + &crop_x, &crop_y, &translated_camera_time_us)) {
|
| + surface_texture_helper_->ReturnTextureFrame();
|
| + return;
|
| + }
|
| +
|
| + webrtc_jni::Matrix matrix = handle.sampling_matrix;
|
| +
|
| + matrix.Crop(crop_width / static_cast<float>(width),
|
| + crop_height / static_cast<float>(height),
|
| + crop_x / static_cast<float>(width),
|
| + crop_y / static_cast<float>(height));
|
| +
|
| + rtc::CritScope lock(&apply_rotation_crit_);
|
| + if (apply_rotation_) {
|
| + if (rotation == webrtc::kVideoRotation_90 ||
|
| + rotation == webrtc::kVideoRotation_270) {
|
| + std::swap(adapted_width, adapted_height);
|
| + }
|
| + matrix.Rotate(static_cast<webrtc::VideoRotation>(rotation));
|
| + }
|
| +
|
| + OnFrame(cricket::WebRtcVideoFrame(
|
| + surface_texture_helper_->CreateTextureFrame(
|
| + adapted_width, adapted_height,
|
| + webrtc_jni::NativeHandleImpl(handle.oes_texture_id, matrix)),
|
| + apply_rotation_ ? webrtc::kVideoRotation_0
|
| + : static_cast<webrtc::VideoRotation>(rotation),
|
| + translated_camera_time_us),
|
| + width, height);
|
| +}
|
| +
|
| +void AndroidVideoTrackSource::OnFrame(const cricket::VideoFrame& frame,
|
| + int width,
|
| + int height) {
|
| + {
|
| + rtc::CritScope lock(&stats_crit_);
|
| + stats_ = rtc::Optional<AndroidVideoTrackSource::Stats>({width, height});
|
| + }
|
| +
|
| + broadcaster_.OnFrame(frame);
|
| +}
|
| +
|
| +void AndroidVideoTrackSource::OnOutputFormatRequest(int width,
|
| + int height,
|
| + int fps) {
|
| + RTC_DCHECK(camera_thread_checker_.CalledOnValidThread());
|
| +
|
| + cricket::VideoFormat format(width, height,
|
| + cricket::VideoFormat::FpsToInterval(fps), 0);
|
| + video_adapter_.OnOutputFormatRequest(format);
|
| +}
|
| +
|
| +bool AndroidVideoTrackSource::AdaptFrame(int width,
|
| + int height,
|
| + int64_t camera_time_us,
|
| + int* out_width,
|
| + int* out_height,
|
| + int* crop_width,
|
| + int* crop_height,
|
| + int* crop_x,
|
| + int* crop_y,
|
| + int64_t* translated_camera_time_us) {
|
| + RTC_DCHECK(camera_thread_checker_.CalledOnValidThread());
|
| +
|
| + int64_t system_time_us = rtc::TimeMicros();
|
| +
|
| + int64_t offset_us =
|
| + timestamp_aligner_.UpdateOffset(camera_time_us, system_time_us);
|
| +
|
| + if (!broadcaster_.frame_wanted()) {
|
| + return false;
|
| + }
|
| +
|
| + if (!video_adapter_.AdaptFrameResolution(
|
| + width, height, camera_time_us * rtc::kNumNanosecsPerMicrosec,
|
| + crop_width, crop_height, out_width, out_height)) {
|
| + // VideoAdapter dropped the frame.
|
| + return false;
|
| + }
|
| + *crop_x = (width - *crop_width) / 2;
|
| + *crop_y = (height - *crop_height) / 2;
|
| +
|
| + *translated_camera_time_us = timestamp_aligner_.ClipTimestamp(
|
| + camera_time_us + offset_us, system_time_us);
|
| + return true;
|
| +}
|
| +
|
| +} // namespace webrtc
|
|
|