OLD | NEW |
1 /* | 1 /* |
2 * Copyright (c) 2011 The WebRTC project authors. All Rights Reserved. | 2 * Copyright (c) 2011 The WebRTC project authors. All Rights Reserved. |
3 * | 3 * |
4 * Use of this source code is governed by a BSD-style license | 4 * Use of this source code is governed by a BSD-style license |
5 * that can be found in the LICENSE file in the root of the source | 5 * that can be found in the LICENSE file in the root of the source |
6 * tree. An additional intellectual property rights grant can be found | 6 * tree. An additional intellectual property rights grant can be found |
7 * in the file PATENTS. All contributing project authors may | 7 * in the file PATENTS. All contributing project authors may |
8 * be found in the AUTHORS file in the root of the source tree. | 8 * be found in the AUTHORS file in the root of the source tree. |
9 */ | 9 */ |
10 | 10 |
11 #include "webrtc/media/engine/webrtcvideoframe.h" | 11 #include "webrtc/media/engine/webrtcvideoframe.h" |
12 | 12 |
13 #include "libyuv/convert.h" | 13 #include "libyuv/convert.h" |
14 #include "webrtc/base/logging.h" | 14 #include "webrtc/base/logging.h" |
15 #include "webrtc/media/base/videocapturer.h" | 15 #include "webrtc/media/base/videocapturer.h" |
16 #include "webrtc/media/base/videocommon.h" | 16 #include "webrtc/media/base/videocommon.h" |
17 #include "webrtc/video_frame.h" | 17 #include "webrtc/video_frame.h" |
18 | 18 |
19 using webrtc::kYPlane; | 19 using webrtc::kYPlane; |
20 using webrtc::kUPlane; | 20 using webrtc::kUPlane; |
21 using webrtc::kVPlane; | 21 using webrtc::kVPlane; |
22 | 22 |
23 namespace cricket { | 23 namespace cricket { |
24 | 24 |
25 WebRtcVideoFrame::WebRtcVideoFrame(): | 25 WebRtcVideoFrame::WebRtcVideoFrame() |
26 time_stamp_ns_(0), | 26 : timestamp_us_(0), rotation_(webrtc::kVideoRotation_0) {} |
27 rotation_(webrtc::kVideoRotation_0) {} | |
28 | 27 |
29 WebRtcVideoFrame::WebRtcVideoFrame( | 28 WebRtcVideoFrame::WebRtcVideoFrame( |
30 const rtc::scoped_refptr<webrtc::VideoFrameBuffer>& buffer, | 29 const rtc::scoped_refptr<webrtc::VideoFrameBuffer>& buffer, |
31 int64_t time_stamp_ns, | 30 int64_t time_stamp_ns, |
32 webrtc::VideoRotation rotation) | 31 webrtc::VideoRotation rotation) |
33 : video_frame_buffer_(buffer), | 32 : video_frame_buffer_(buffer), |
34 time_stamp_ns_(time_stamp_ns), | |
35 rotation_(rotation) { | 33 rotation_(rotation) { |
| 34 // Convert to usecs. |
| 35 SetTimeStamp(time_stamp_ns); |
36 } | 36 } |
37 | 37 |
38 WebRtcVideoFrame::~WebRtcVideoFrame() {} | 38 WebRtcVideoFrame::~WebRtcVideoFrame() {} |
39 | 39 |
40 bool WebRtcVideoFrame::Init(uint32_t format, | 40 bool WebRtcVideoFrame::Init(uint32_t format, |
41 int w, | 41 int w, |
42 int h, | 42 int h, |
43 int dw, | 43 int dw, |
44 int dh, | 44 int dh, |
45 uint8_t* sample, | 45 uint8_t* sample, |
46 size_t sample_size, | 46 size_t sample_size, |
47 int64_t time_stamp_ns, | 47 int64_t time_stamp_ns, |
48 webrtc::VideoRotation rotation) { | 48 webrtc::VideoRotation rotation) { |
49 return Reset(format, w, h, dw, dh, sample, sample_size, | 49 if (!Reset(format, w, h, dw, dh, sample, sample_size, rotation, |
50 time_stamp_ns, rotation, | 50 true /*apply_rotation*/)) |
51 true /*apply_rotation*/); | 51 return false; |
| 52 |
| 53 SetTimeStamp(time_stamp_ns); |
| 54 return true; |
52 } | 55 } |
53 | 56 |
54 bool WebRtcVideoFrame::Init(const CapturedFrame* frame, int dw, int dh, | 57 bool WebRtcVideoFrame::Init(const CapturedFrame* frame, int dw, int dh, |
55 bool apply_rotation) { | 58 bool apply_rotation) { |
56 return Reset(frame->fourcc, frame->width, frame->height, dw, dh, | 59 if (!Reset(frame->fourcc, frame->width, frame->height, dw, dh, |
57 static_cast<uint8_t*>(frame->data), frame->data_size, | 60 static_cast<uint8_t*>(frame->data), frame->data_size, |
58 frame->time_stamp, | 61 frame->rotation, apply_rotation)) { |
59 frame->rotation, apply_rotation); | 62 return false; |
| 63 } |
| 64 SetTimeStamp(frame->time_stamp); |
| 65 |
| 66 // Check that we always get the right epoch (with a 30s margin). If |
| 67 // this is false for some camera, we need conversion logic adding |
| 68 // the camera's clock offset. |
| 69 RTC_DCHECK_LT( |
| 70 std::abs(timestamp_us_ - static_cast<int64_t>(rtc::TimeMicros())), |
| 71 30 * rtc::kNumMicrosecsPerSec); |
| 72 |
| 73 return true; |
60 } | 74 } |
61 | 75 |
62 bool WebRtcVideoFrame::InitToBlack(int w, int h, | 76 bool WebRtcVideoFrame::InitToBlack(int w, int h, |
63 int64_t time_stamp_ns) { | 77 int64_t time_stamp_ns) { |
64 InitToEmptyBuffer(w, h, time_stamp_ns); | 78 InitToEmptyBuffer(w, h, time_stamp_ns); |
65 return SetToBlack(); | 79 return SetToBlack(); |
66 } | 80 } |
67 | 81 |
68 int WebRtcVideoFrame::width() const { | 82 int WebRtcVideoFrame::width() const { |
69 return video_frame_buffer_ ? video_frame_buffer_->width() : 0; | 83 return video_frame_buffer_ ? video_frame_buffer_->width() : 0; |
(...skipping 50 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
120 return video_frame_buffer_ ? video_frame_buffer_->native_handle() : nullptr; | 134 return video_frame_buffer_ ? video_frame_buffer_->native_handle() : nullptr; |
121 } | 135 } |
122 | 136 |
123 rtc::scoped_refptr<webrtc::VideoFrameBuffer> | 137 rtc::scoped_refptr<webrtc::VideoFrameBuffer> |
124 WebRtcVideoFrame::GetVideoFrameBuffer() const { | 138 WebRtcVideoFrame::GetVideoFrameBuffer() const { |
125 return video_frame_buffer_; | 139 return video_frame_buffer_; |
126 } | 140 } |
127 | 141 |
128 VideoFrame* WebRtcVideoFrame::Copy() const { | 142 VideoFrame* WebRtcVideoFrame::Copy() const { |
129 WebRtcVideoFrame* new_frame = new WebRtcVideoFrame( | 143 WebRtcVideoFrame* new_frame = new WebRtcVideoFrame( |
130 video_frame_buffer_, time_stamp_ns_, rotation_); | 144 video_frame_buffer_, 0 /* Dummy timestamp, overwridden below */, |
| 145 rotation_); |
| 146 new_frame->set_timestamp_us(timestamp_us_); |
131 return new_frame; | 147 return new_frame; |
132 } | 148 } |
133 | 149 |
134 size_t WebRtcVideoFrame::ConvertToRgbBuffer(uint32_t to_fourcc, | 150 size_t WebRtcVideoFrame::ConvertToRgbBuffer(uint32_t to_fourcc, |
135 uint8_t* buffer, | 151 uint8_t* buffer, |
136 size_t size, | 152 size_t size, |
137 int stride_rgb) const { | 153 int stride_rgb) const { |
138 RTC_CHECK(video_frame_buffer_); | 154 RTC_CHECK(video_frame_buffer_); |
139 RTC_CHECK(video_frame_buffer_->native_handle() == nullptr); | 155 RTC_CHECK(video_frame_buffer_->native_handle() == nullptr); |
140 return VideoFrame::ConvertToRgbBuffer(to_fourcc, buffer, size, stride_rgb); | 156 return VideoFrame::ConvertToRgbBuffer(to_fourcc, buffer, size, stride_rgb); |
141 } | 157 } |
142 | 158 |
143 bool WebRtcVideoFrame::Reset(uint32_t format, | 159 bool WebRtcVideoFrame::Reset(uint32_t format, |
144 int w, | 160 int w, |
145 int h, | 161 int h, |
146 int dw, | 162 int dw, |
147 int dh, | 163 int dh, |
148 uint8_t* sample, | 164 uint8_t* sample, |
149 size_t sample_size, | 165 size_t sample_size, |
150 int64_t time_stamp_ns, | |
151 webrtc::VideoRotation rotation, | 166 webrtc::VideoRotation rotation, |
152 bool apply_rotation) { | 167 bool apply_rotation) { |
153 if (!Validate(format, w, h, sample, sample_size)) { | 168 if (!Validate(format, w, h, sample, sample_size)) { |
154 return false; | 169 return false; |
155 } | 170 } |
156 // Translate aliases to standard enums (e.g., IYUV -> I420). | 171 // Translate aliases to standard enums (e.g., IYUV -> I420). |
157 format = CanonicalFourCC(format); | 172 format = CanonicalFourCC(format); |
158 | 173 |
159 // Set up a new buffer. | 174 // Set up a new buffer. |
160 // TODO(fbarchard): Support lazy allocation. | 175 // TODO(fbarchard): Support lazy allocation. |
161 int new_width = dw; | 176 int new_width = dw; |
162 int new_height = dh; | 177 int new_height = dh; |
163 // If rotated swap width, height. | 178 // If rotated swap width, height. |
164 if (apply_rotation && (rotation == 90 || rotation == 270)) { | 179 if (apply_rotation && (rotation == 90 || rotation == 270)) { |
165 new_width = dh; | 180 new_width = dh; |
166 new_height = dw; | 181 new_height = dw; |
167 } | 182 } |
168 | 183 |
169 InitToEmptyBuffer(new_width, new_height, | 184 InitToEmptyBuffer(new_width, new_height); |
170 time_stamp_ns); | |
171 rotation_ = apply_rotation ? webrtc::kVideoRotation_0 : rotation; | 185 rotation_ = apply_rotation ? webrtc::kVideoRotation_0 : rotation; |
172 | 186 |
173 int horiz_crop = ((w - dw) / 2) & ~1; | 187 int horiz_crop = ((w - dw) / 2) & ~1; |
174 // ARGB on Windows has negative height. | 188 // ARGB on Windows has negative height. |
175 // The sample's layout in memory is normal, so just correct crop. | 189 // The sample's layout in memory is normal, so just correct crop. |
176 int vert_crop = ((abs(h) - dh) / 2) & ~1; | 190 int vert_crop = ((abs(h) - dh) / 2) & ~1; |
177 // Conversion functions expect negative height to flip the image. | 191 // Conversion functions expect negative height to flip the image. |
178 int idh = (h < 0) ? -dh : dh; | 192 int idh = (h < 0) ? -dh : dh; |
179 int r = libyuv::ConvertToI420( | 193 int r = libyuv::ConvertToI420( |
180 sample, sample_size, | 194 sample, sample_size, |
181 GetYPlane(), GetYPitch(), | 195 GetYPlane(), GetYPitch(), |
182 GetUPlane(), GetUPitch(), | 196 GetUPlane(), GetUPitch(), |
183 GetVPlane(), GetVPitch(), | 197 GetVPlane(), GetVPitch(), |
184 horiz_crop, vert_crop, | 198 horiz_crop, vert_crop, |
185 w, h, | 199 w, h, |
186 dw, idh, | 200 dw, idh, |
187 static_cast<libyuv::RotationMode>( | 201 static_cast<libyuv::RotationMode>( |
188 apply_rotation ? rotation : webrtc::kVideoRotation_0), | 202 apply_rotation ? rotation : webrtc::kVideoRotation_0), |
189 format); | 203 format); |
190 if (r) { | 204 if (r) { |
191 LOG(LS_ERROR) << "Error parsing format: " << GetFourccName(format) | 205 LOG(LS_ERROR) << "Error parsing format: " << GetFourccName(format) |
192 << " return code : " << r; | 206 << " return code : " << r; |
193 return false; | 207 return false; |
194 } | 208 } |
195 return true; | 209 return true; |
196 } | 210 } |
197 | 211 |
198 VideoFrame* WebRtcVideoFrame::CreateEmptyFrame( | 212 VideoFrame* WebRtcVideoFrame::CreateEmptyFrame( |
199 int w, int h, | 213 int w, int h, |
200 int64_t time_stamp_ns) const { | 214 int64_t timestamp_us) const { |
201 WebRtcVideoFrame* frame = new WebRtcVideoFrame(); | 215 WebRtcVideoFrame* frame = new WebRtcVideoFrame(); |
202 frame->InitToEmptyBuffer(w, h, time_stamp_ns); | 216 frame->InitToEmptyBuffer(w, h, rtc::kNumNanosecsPerMicrosec * timestamp_us); |
203 return frame; | 217 return frame; |
204 } | 218 } |
205 | 219 |
| 220 void WebRtcVideoFrame::InitToEmptyBuffer(int w, int h) { |
| 221 video_frame_buffer_ = new rtc::RefCountedObject<webrtc::I420Buffer>(w, h); |
| 222 rotation_ = webrtc::kVideoRotation_0; |
| 223 } |
| 224 |
206 void WebRtcVideoFrame::InitToEmptyBuffer(int w, int h, | 225 void WebRtcVideoFrame::InitToEmptyBuffer(int w, int h, |
207 int64_t time_stamp_ns) { | 226 int64_t time_stamp_ns) { |
208 video_frame_buffer_ = new rtc::RefCountedObject<webrtc::I420Buffer>(w, h); | 227 video_frame_buffer_ = new rtc::RefCountedObject<webrtc::I420Buffer>(w, h); |
209 time_stamp_ns_ = time_stamp_ns; | 228 SetTimeStamp(time_stamp_ns); |
210 rotation_ = webrtc::kVideoRotation_0; | 229 rotation_ = webrtc::kVideoRotation_0; |
211 } | 230 } |
212 | 231 |
213 const VideoFrame* WebRtcVideoFrame::GetCopyWithRotationApplied() const { | 232 const VideoFrame* WebRtcVideoFrame::GetCopyWithRotationApplied() const { |
214 // If the frame is not rotated, the caller should reuse this frame instead of | 233 // If the frame is not rotated, the caller should reuse this frame instead of |
215 // making a redundant copy. | 234 // making a redundant copy. |
216 if (GetVideoRotation() == webrtc::kVideoRotation_0) { | 235 if (GetVideoRotation() == webrtc::kVideoRotation_0) { |
217 return this; | 236 return this; |
218 } | 237 } |
219 | 238 |
(...skipping 10 matching lines...) Expand all Loading... |
230 int orig_height = height(); | 249 int orig_height = height(); |
231 | 250 |
232 int rotated_width = orig_width; | 251 int rotated_width = orig_width; |
233 int rotated_height = orig_height; | 252 int rotated_height = orig_height; |
234 if (GetVideoRotation() == webrtc::kVideoRotation_90 || | 253 if (GetVideoRotation() == webrtc::kVideoRotation_90 || |
235 GetVideoRotation() == webrtc::kVideoRotation_270) { | 254 GetVideoRotation() == webrtc::kVideoRotation_270) { |
236 rotated_width = orig_height; | 255 rotated_width = orig_height; |
237 rotated_height = orig_width; | 256 rotated_height = orig_width; |
238 } | 257 } |
239 | 258 |
240 rotated_frame_.reset(CreateEmptyFrame(rotated_width, rotated_height, | 259 rotated_frame_.reset( |
241 GetTimeStamp())); | 260 CreateEmptyFrame(rotated_width, rotated_height, timestamp_us_)); |
242 | 261 |
243 // TODO(guoweis): Add a function in webrtc_libyuv.cc to convert from | 262 // TODO(guoweis): Add a function in webrtc_libyuv.cc to convert from |
244 // VideoRotation to libyuv::RotationMode. | 263 // VideoRotation to libyuv::RotationMode. |
245 int ret = libyuv::I420Rotate( | 264 int ret = libyuv::I420Rotate( |
246 GetYPlane(), GetYPitch(), GetUPlane(), GetUPitch(), GetVPlane(), | 265 GetYPlane(), GetYPitch(), GetUPlane(), GetUPitch(), GetVPlane(), |
247 GetVPitch(), rotated_frame_->GetYPlane(), rotated_frame_->GetYPitch(), | 266 GetVPitch(), rotated_frame_->GetYPlane(), rotated_frame_->GetYPitch(), |
248 rotated_frame_->GetUPlane(), rotated_frame_->GetUPitch(), | 267 rotated_frame_->GetUPlane(), rotated_frame_->GetUPitch(), |
249 rotated_frame_->GetVPlane(), rotated_frame_->GetVPitch(), | 268 rotated_frame_->GetVPlane(), rotated_frame_->GetVPitch(), |
250 orig_width, orig_height, | 269 orig_width, orig_height, |
251 static_cast<libyuv::RotationMode>(GetVideoRotation())); | 270 static_cast<libyuv::RotationMode>(GetVideoRotation())); |
252 if (ret == 0) { | 271 if (ret == 0) { |
253 return rotated_frame_.get(); | 272 return rotated_frame_.get(); |
254 } | 273 } |
255 return nullptr; | 274 return nullptr; |
256 } | 275 } |
257 | 276 |
258 } // namespace cricket | 277 } // namespace cricket |
OLD | NEW |