Index: webrtc/api/android/jni/androidvideotracksource.cc |
diff --git a/webrtc/api/android/jni/androidvideotracksource.cc b/webrtc/api/android/jni/androidvideotracksource.cc |
deleted file mode 100644 |
index 4f708a6b4f7cb5d9d1bb59a957c0c4f4bc7c6f8b..0000000000000000000000000000000000000000 |
--- a/webrtc/api/android/jni/androidvideotracksource.cc |
+++ /dev/null |
@@ -1,163 +0,0 @@ |
-/* |
- * Copyright (c) 2016 The WebRTC project authors. All Rights Reserved. |
- * |
- * Use of this source code is governed by a BSD-style license |
- * that can be found in the LICENSE file in the root of the source |
- * tree. An additional intellectual property rights grant can be found |
- * in the file PATENTS. All contributing project authors may |
- * be found in the AUTHORS file in the root of the source tree. |
- */ |
- |
-#include "webrtc/api/android/jni/androidvideotracksource.h" |
- |
-#include <utility> |
- |
-namespace webrtc { |
- |
-AndroidVideoTrackSource::AndroidVideoTrackSource(rtc::Thread* signaling_thread, |
- JNIEnv* jni, |
- jobject j_egl_context, |
- bool is_screencast) |
- : signaling_thread_(signaling_thread), |
- surface_texture_helper_(webrtc_jni::SurfaceTextureHelper::create( |
- jni, |
- "Camera SurfaceTextureHelper", |
- j_egl_context)), |
- is_screencast_(is_screencast) { |
- LOG(LS_INFO) << "AndroidVideoTrackSource ctor"; |
- camera_thread_checker_.DetachFromThread(); |
-} |
- |
-void AndroidVideoTrackSource::SetState(SourceState state) { |
- if (rtc::Thread::Current() != signaling_thread_) { |
- invoker_.AsyncInvoke<void>( |
- RTC_FROM_HERE, signaling_thread_, |
- rtc::Bind(&AndroidVideoTrackSource::SetState, this, state)); |
- return; |
- } |
- |
- if (state_ != state) { |
- state_ = state; |
- FireOnChanged(); |
- } |
-} |
- |
-void AndroidVideoTrackSource::OnByteBufferFrameCaptured(const void* frame_data, |
- int length, |
- int width, |
- int height, |
- int rotation, |
- int64_t timestamp_ns) { |
- RTC_DCHECK(camera_thread_checker_.CalledOnValidThread()); |
- RTC_DCHECK(rotation == 0 || rotation == 90 || rotation == 180 || |
- rotation == 270); |
- |
- int64_t camera_time_us = timestamp_ns / rtc::kNumNanosecsPerMicrosec; |
- int64_t translated_camera_time_us = |
- timestamp_aligner_.TranslateTimestamp(camera_time_us, rtc::TimeMicros()); |
- |
- int adapted_width; |
- int adapted_height; |
- int crop_width; |
- int crop_height; |
- int crop_x; |
- int crop_y; |
- |
- if (!AdaptFrame(width, height, camera_time_us, &adapted_width, |
- &adapted_height, &crop_width, &crop_height, &crop_x, |
- &crop_y)) { |
- return; |
- } |
- |
- const uint8_t* y_plane = static_cast<const uint8_t*>(frame_data); |
- const uint8_t* uv_plane = y_plane + width * height; |
- const int uv_width = (width + 1) / 2; |
- |
- RTC_CHECK_GE(length, width * height + 2 * uv_width * ((height + 1) / 2)); |
- |
- // Can only crop at even pixels. |
- crop_x &= ~1; |
- crop_y &= ~1; |
- // Crop just by modifying pointers. |
- y_plane += width * crop_y + crop_x; |
- uv_plane += uv_width * crop_y + crop_x; |
- |
- rtc::scoped_refptr<webrtc::I420Buffer> buffer = |
- buffer_pool_.CreateBuffer(adapted_width, adapted_height); |
- |
- nv12toi420_scaler_.NV12ToI420Scale( |
- y_plane, width, uv_plane, uv_width * 2, crop_width, crop_height, |
- buffer->MutableDataY(), buffer->StrideY(), |
- // Swap U and V, since we have NV21, not NV12. |
- buffer->MutableDataV(), buffer->StrideV(), buffer->MutableDataU(), |
- buffer->StrideU(), buffer->width(), buffer->height()); |
- |
- OnFrame(VideoFrame(buffer, static_cast<webrtc::VideoRotation>(rotation), |
- translated_camera_time_us)); |
-} |
- |
-void AndroidVideoTrackSource::OnTextureFrameCaptured( |
- int width, |
- int height, |
- int rotation, |
- int64_t timestamp_ns, |
- const webrtc_jni::NativeHandleImpl& handle) { |
- RTC_DCHECK(camera_thread_checker_.CalledOnValidThread()); |
- RTC_DCHECK(rotation == 0 || rotation == 90 || rotation == 180 || |
- rotation == 270); |
- |
- int64_t camera_time_us = timestamp_ns / rtc::kNumNanosecsPerMicrosec; |
- int64_t translated_camera_time_us = |
- timestamp_aligner_.TranslateTimestamp(camera_time_us, rtc::TimeMicros()); |
- |
- int adapted_width; |
- int adapted_height; |
- int crop_width; |
- int crop_height; |
- int crop_x; |
- int crop_y; |
- |
- if (!AdaptFrame(width, height, camera_time_us, &adapted_width, |
- &adapted_height, &crop_width, &crop_height, &crop_x, |
- &crop_y)) { |
- surface_texture_helper_->ReturnTextureFrame(); |
- return; |
- } |
- |
- webrtc_jni::Matrix matrix = handle.sampling_matrix; |
- |
- matrix.Crop(crop_width / static_cast<float>(width), |
- crop_height / static_cast<float>(height), |
- crop_x / static_cast<float>(width), |
- crop_y / static_cast<float>(height)); |
- |
- // Make a local copy, since value of apply_rotation() may change |
- // under our feet. |
- bool do_rotate = apply_rotation(); |
- |
- if (do_rotate) { |
- if (rotation == webrtc::kVideoRotation_90 || |
- rotation == webrtc::kVideoRotation_270) { |
- std::swap(adapted_width, adapted_height); |
- } |
- matrix.Rotate(static_cast<webrtc::VideoRotation>(rotation)); |
- } |
- |
- OnFrame(VideoFrame( |
- surface_texture_helper_->CreateTextureFrame( |
- adapted_width, adapted_height, |
- webrtc_jni::NativeHandleImpl(handle.oes_texture_id, matrix)), |
- do_rotate ? webrtc::kVideoRotation_0 |
- : static_cast<webrtc::VideoRotation>(rotation), |
- translated_camera_time_us)); |
-} |
- |
-void AndroidVideoTrackSource::OnOutputFormatRequest(int width, |
- int height, |
- int fps) { |
- cricket::VideoFormat format(width, height, |
- cricket::VideoFormat::FpsToInterval(fps), 0); |
- video_adapter()->OnOutputFormatRequest(format); |
-} |
- |
-} // namespace webrtc |