Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(174)

Side by Side Diff: webrtc/modules/video_coding/codecs/vp9/vp9_impl.cc

Issue 1211353002: Integration of VP9 packetization. (Closed) Base URL: https://chromium.googlesource.com/external/webrtc.git@master
Patch Set: Created 5 years, 4 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch
OLDNEW
1 /* 1 /*
2 * Copyright (c) 2014 The WebRTC project authors. All Rights Reserved. 2 * Copyright (c) 2014 The WebRTC project authors. All Rights Reserved.
3 * 3 *
4 * Use of this source code is governed by a BSD-style license 4 * Use of this source code is governed by a BSD-style license
5 * that can be found in the LICENSE file in the root of the source 5 * that can be found in the LICENSE file in the root of the source
6 * tree. An additional intellectual property rights grant can be found 6 * tree. An additional intellectual property rights grant can be found
7 * in the file PATENTS. All contributing project authors may 7 * in the file PATENTS. All contributing project authors may
8 * be found in the AUTHORS file in the root of the source tree. 8 * be found in the AUTHORS file in the root of the source tree.
9 * 9 *
10 */ 10 */
(...skipping 39 matching lines...) Expand 10 before | Expand all | Expand 10 after
50 if (width * height <= 352 * 288) 50 if (width * height <= 352 * 288)
51 return 5; 51 return 5;
52 else 52 else
53 return 7; 53 return 7;
54 } 54 }
55 55
56 VP9Encoder* VP9Encoder::Create() { 56 VP9Encoder* VP9Encoder::Create() {
57 return new VP9EncoderImpl(); 57 return new VP9EncoderImpl();
58 } 58 }
59 59
60 void VP9EncoderImpl::EncoderOutputCodedPacketCallback(vpx_codec_cx_pkt* pkt,
61 void* user_data) {
62 VP9EncoderImpl* enc = (VP9EncoderImpl*)(user_data);
63 enc->GetEncodedLayerFrame(pkt);
64 }
65
60 VP9EncoderImpl::VP9EncoderImpl() 66 VP9EncoderImpl::VP9EncoderImpl()
61 : encoded_image_(), 67 : encoded_image_(),
62 encoded_complete_callback_(NULL), 68 encoded_complete_callback_(NULL),
63 inited_(false), 69 inited_(false),
64 timestamp_(0), 70 timestamp_(0),
65 picture_id_(0), 71 picture_id_(0),
66 cpu_speed_(3), 72 cpu_speed_(3),
67 rc_max_intra_target_(0), 73 rc_max_intra_target_(0),
68 encoder_(NULL), 74 encoder_(NULL),
69 config_(NULL), 75 config_(NULL),
70 raw_(NULL) { 76 raw_(NULL),
77 input_image_(NULL),
78 tl0_pic_idx_(0),
79 gof_idx_(0),
80 num_temporal_layers_(0),
81 num_spatial_layers_(0) {
71 memset(&codec_, 0, sizeof(codec_)); 82 memset(&codec_, 0, sizeof(codec_));
72 uint32_t seed = static_cast<uint32_t>(TickTime::MillisecondTimestamp()); 83 uint32_t seed = static_cast<uint32_t>(TickTime::MillisecondTimestamp());
73 srand(seed); 84 srand(seed);
74 } 85 }
75 86
76 VP9EncoderImpl::~VP9EncoderImpl() { 87 VP9EncoderImpl::~VP9EncoderImpl() {
77 Release(); 88 Release();
78 } 89 }
79 90
80 int VP9EncoderImpl::Release() { 91 int VP9EncoderImpl::Release() {
(...skipping 13 matching lines...) Expand all
94 config_ = NULL; 105 config_ = NULL;
95 } 106 }
96 if (raw_ != NULL) { 107 if (raw_ != NULL) {
97 vpx_img_free(raw_); 108 vpx_img_free(raw_);
98 raw_ = NULL; 109 raw_ = NULL;
99 } 110 }
100 inited_ = false; 111 inited_ = false;
101 return WEBRTC_VIDEO_CODEC_OK; 112 return WEBRTC_VIDEO_CODEC_OK;
102 } 113 }
103 114
115 bool VP9EncoderImpl::SetSvcRates() {
116 float rate_ratio[VPX_MAX_LAYERS] = {0};
117 float total = 0;
118 uint8_t i = 0;
119
120 if (num_spatial_layers_ > 1) {
121 for (i = 0; i < num_spatial_layers_; ++i) {
122 if (svc_internal_.svc_params.scaling_factor_num[i] <= 0 ||
123 svc_internal_.svc_params.scaling_factor_den[i] <= 0) {
124 return false;
125 }
126 rate_ratio[i] = static_cast<float>(
127 svc_internal_.svc_params.scaling_factor_num[i]) /
128 svc_internal_.svc_params.scaling_factor_den[i];
129 total += rate_ratio[i];
130 }
131 } else {
132 rate_ratio[0] = total = 1;
133 }
134 assert(total >= 1.0f);
stefan-webrtc 2015/07/29 13:13:05 Hm, seems like there could be a risk that this doe
åsapersson 2015/07/30 11:42:19 Done.
135
136 for (i = 0; i < num_spatial_layers_; ++i) {
137 config_->ss_target_bitrate[i] = static_cast<unsigned int>(
138 config_->rc_target_bitrate * rate_ratio[i] / total);
stefan-webrtc 2015/07/29 13:13:05 Should this really divide by total? total is expec
åsapersson 2015/07/30 11:42:19 total will be >1 for num_spatial_layers_ > 1...
stefan-webrtc 2015/07/30 12:05:25 Right, my mistake.
139 if (num_temporal_layers_ == 1) {
140 config_->layer_target_bitrate[0] = config_->ss_target_bitrate[i];
141 } else if (num_temporal_layers_ == 2) {
142 config_->layer_target_bitrate[i * num_temporal_layers_] =
143 config_->ss_target_bitrate[i] * 2 / 3;
144 config_->layer_target_bitrate[i * num_temporal_layers_ + 1] =
145 config_->ss_target_bitrate[i];
146 } else if (num_temporal_layers_ == 3) {
147 config_->layer_target_bitrate[i * num_temporal_layers_] =
148 config_->ss_target_bitrate[i] / 2;
149 config_->layer_target_bitrate[i * num_temporal_layers_ + 1] =
150 config_->layer_target_bitrate[i * num_temporal_layers_] +
151 (config_->ss_target_bitrate[i] / 4);
152 config_->layer_target_bitrate[i * num_temporal_layers_ + 2] =
153 config_->ss_target_bitrate[i];
154 } else {
155 return false;
156 }
157 }
158
159 if (num_spatial_layers_ == 1) {
stefan-webrtc 2015/07/29 13:13:05 Comment: Currently only support temporal layers if
åsapersson 2015/07/30 11:42:19 Added a comment.
160 for (i = 0; i < num_temporal_layers_; ++i) {
161 config_->ts_target_bitrate[i] = config_->layer_target_bitrate[i];
162 }
163 }
164
165 return true;
166 }
167
104 int VP9EncoderImpl::SetRates(uint32_t new_bitrate_kbit, 168 int VP9EncoderImpl::SetRates(uint32_t new_bitrate_kbit,
105 uint32_t new_framerate) { 169 uint32_t new_framerate) {
106 if (!inited_) { 170 if (!inited_) {
107 return WEBRTC_VIDEO_CODEC_UNINITIALIZED; 171 return WEBRTC_VIDEO_CODEC_UNINITIALIZED;
108 } 172 }
109 if (encoder_->err) { 173 if (encoder_->err) {
110 return WEBRTC_VIDEO_CODEC_ERROR; 174 return WEBRTC_VIDEO_CODEC_ERROR;
111 } 175 }
112 if (new_framerate < 1) { 176 if (new_framerate < 1) {
113 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; 177 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER;
114 } 178 }
115 // Update bit rate 179 // Update bit rate
116 if (codec_.maxBitrate > 0 && new_bitrate_kbit > codec_.maxBitrate) { 180 if (codec_.maxBitrate > 0 && new_bitrate_kbit > codec_.maxBitrate) {
117 new_bitrate_kbit = codec_.maxBitrate; 181 new_bitrate_kbit = codec_.maxBitrate;
118 } 182 }
119 config_->rc_target_bitrate = new_bitrate_kbit; 183 config_->rc_target_bitrate = new_bitrate_kbit;
120 codec_.maxFramerate = new_framerate; 184 codec_.maxFramerate = new_framerate;
185
186 if (!SetSvcRates()) {
187 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER;
188 }
189
121 // Update encoder context 190 // Update encoder context
122 if (vpx_codec_enc_config_set(encoder_, config_)) { 191 if (vpx_codec_enc_config_set(encoder_, config_)) {
123 return WEBRTC_VIDEO_CODEC_ERROR; 192 return WEBRTC_VIDEO_CODEC_ERROR;
124 } 193 }
125 return WEBRTC_VIDEO_CODEC_OK; 194 return WEBRTC_VIDEO_CODEC_OK;
126 } 195 }
127 196
128 int VP9EncoderImpl::InitEncode(const VideoCodec* inst, 197 int VP9EncoderImpl::InitEncode(const VideoCodec* inst,
129 int number_of_cores, 198 int number_of_cores,
130 size_t /*max_payload_size*/) { 199 size_t /*max_payload_size*/) {
(...skipping 20 matching lines...) Expand all
151 if (encoder_ == NULL) { 220 if (encoder_ == NULL) {
152 encoder_ = new vpx_codec_ctx_t; 221 encoder_ = new vpx_codec_ctx_t;
153 } 222 }
154 if (config_ == NULL) { 223 if (config_ == NULL) {
155 config_ = new vpx_codec_enc_cfg_t; 224 config_ = new vpx_codec_enc_cfg_t;
156 } 225 }
157 timestamp_ = 0; 226 timestamp_ = 0;
158 if (&codec_ != inst) { 227 if (&codec_ != inst) {
159 codec_ = *inst; 228 codec_ = *inst;
160 } 229 }
230
231 num_temporal_layers_ = inst->codecSpecific.VP9.numberOfTemporalLayers;
232 num_spatial_layers_ = inst->codecSpecific.VP9.numberOfSpatialLayers;
233 // For now, only support 1 spatial layer.
234 if (num_spatial_layers_ != 1) {
stefan-webrtc 2015/07/29 13:13:05 Does this mean one more than the base layer?
åsapersson 2015/07/30 11:42:19 Only one spatial layer.
235 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER;
236 }
237
161 // Random start 16 bits is enough. 238 // Random start 16 bits is enough.
162 picture_id_ = static_cast<uint16_t>(rand()) & 0x7FFF; 239 picture_id_ = static_cast<uint16_t>(rand()) & 0x7FFF;
163 // Allocate memory for encoded image 240 // Allocate memory for encoded image
164 if (encoded_image_._buffer != NULL) { 241 if (encoded_image_._buffer != NULL) {
165 delete [] encoded_image_._buffer; 242 delete [] encoded_image_._buffer;
166 } 243 }
167 encoded_image_._size = CalcBufferSize(kI420, codec_.width, codec_.height); 244 encoded_image_._size = CalcBufferSize(kI420, codec_.width, codec_.height);
168 encoded_image_._buffer = new uint8_t[encoded_image_._size]; 245 encoded_image_._buffer = new uint8_t[encoded_image_._size];
169 encoded_image_._completeFrame = true; 246 encoded_image_._completeFrame = true;
170 // Creating a wrapper to the image - setting image data to NULL. Actual 247 // Creating a wrapper to the image - setting image data to NULL. Actual
(...skipping 31 matching lines...) Expand 10 before | Expand all | Expand 10 after
202 if (inst->codecSpecific.VP9.keyFrameInterval > 0) { 279 if (inst->codecSpecific.VP9.keyFrameInterval > 0) {
203 config_->kf_mode = VPX_KF_AUTO; 280 config_->kf_mode = VPX_KF_AUTO;
204 config_->kf_max_dist = inst->codecSpecific.VP9.keyFrameInterval; 281 config_->kf_max_dist = inst->codecSpecific.VP9.keyFrameInterval;
205 } else { 282 } else {
206 config_->kf_mode = VPX_KF_DISABLED; 283 config_->kf_mode = VPX_KF_DISABLED;
207 } 284 }
208 // Determine number of threads based on the image size and #cores. 285 // Determine number of threads based on the image size and #cores.
209 config_->g_threads = NumberOfThreads(config_->g_w, 286 config_->g_threads = NumberOfThreads(config_->g_w,
210 config_->g_h, 287 config_->g_h,
211 number_of_cores); 288 number_of_cores);
289
212 cpu_speed_ = GetCpuSpeed(config_->g_w, config_->g_h); 290 cpu_speed_ = GetCpuSpeed(config_->g_w, config_->g_h);
291
292 // TODO(asapersson): Check configuration of temporal switch up.
293 if (num_temporal_layers_ == 1) {
294 gof_.SetGofInfoVP9(kTemporalStructureMode1);
295 config_->temporal_layering_mode = VP9E_TEMPORAL_LAYERING_MODE_NOLAYERING;
296 config_->ts_number_layers = 1;
297 config_->ts_rate_decimator[0] = 1;
298 config_->ts_periodicity = 1;
299 config_->ts_layer_id[0] = 0;
300 } else if (num_temporal_layers_ == 2) {
301 gof_.SetGofInfoVP9(kTemporalStructureMode2);
302 config_->temporal_layering_mode = VP9E_TEMPORAL_LAYERING_MODE_0101;
303 config_->ts_number_layers = 2;
304 config_->ts_rate_decimator[0] = 2;
305 config_->ts_rate_decimator[1] = 1;
306 config_->ts_periodicity = 2;
307 config_->ts_layer_id[0] = 0;
308 config_->ts_layer_id[1] = 1;
309 } else if (num_temporal_layers_ == 3) {
310 gof_.SetGofInfoVP9(kTemporalStructureMode3);
311 config_->temporal_layering_mode = VP9E_TEMPORAL_LAYERING_MODE_0212;
312 config_->ts_number_layers = 3;
313 config_->ts_rate_decimator[0] = 4;
314 config_->ts_rate_decimator[1] = 2;
315 config_->ts_rate_decimator[2] = 1;
316 config_->ts_periodicity = 4;
317 config_->ts_layer_id[0] = 0;
318 config_->ts_layer_id[1] = 2;
319 config_->ts_layer_id[2] = 1;
320 config_->ts_layer_id[3] = 2;
321 } else {
322 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER;
323 }
stefan-webrtc 2015/07/29 13:13:05 Ideally I would like this code to be shared with t
åsapersson 2015/07/30 11:42:19 Leave for follow up cl?
stefan-webrtc 2015/07/30 12:05:25 Acknowledged.
324
325 tl0_pic_idx_ = static_cast<uint8_t>(rand());
326
213 return InitAndSetControlSettings(inst); 327 return InitAndSetControlSettings(inst);
214 } 328 }
215 329
216 int VP9EncoderImpl::NumberOfThreads(int width, 330 int VP9EncoderImpl::NumberOfThreads(int width,
217 int height, 331 int height,
218 int number_of_cores) { 332 int number_of_cores) {
333 // For the current libvpx library, only 1 thread is supported when SVC is
334 // turned on.
335 if (num_temporal_layers_ > 1 || num_spatial_layers_ > 1) {
336 return 1;
337 }
338
219 // Keep the number of encoder threads equal to the possible number of column 339 // Keep the number of encoder threads equal to the possible number of column
220 // tiles, which is (1, 2, 4, 8). See comments below for VP9E_SET_TILE_COLUMNS. 340 // tiles, which is (1, 2, 4, 8). See comments below for VP9E_SET_TILE_COLUMNS.
221 if (width * height >= 1280 * 720 && number_of_cores > 4) { 341 if (width * height >= 1280 * 720 && number_of_cores > 4) {
222 return 4; 342 return 4;
223 } else if (width * height >= 640 * 480 && number_of_cores > 2) { 343 } else if (width * height >= 640 * 480 && number_of_cores > 2) {
224 return 2; 344 return 2;
225 } else { 345 } else {
226 // 1 thread less than VGA. 346 // 1 thread less than VGA.
227 return 1; 347 return 1;
228 } 348 }
229 } 349 }
230 350
231 int VP9EncoderImpl::InitAndSetControlSettings(const VideoCodec* inst) { 351 int VP9EncoderImpl::InitAndSetControlSettings(const VideoCodec* inst) {
352
353 config_->ss_number_layers = num_spatial_layers_;
354
355 if (num_spatial_layers_ > 1) {
356 config_->rc_min_quantizer = 0;
357 config_->rc_max_quantizer = 63;
358 }
359 int scaling_factor_num = 256;
360 for (int i = num_spatial_layers_ - 1; i >= 0; --i, scaling_factor_num /= 2) {
stefan-webrtc 2015/07/29 13:13:05 Move scaling_factor_num /= 2 to line 366 instead.
åsapersson 2015/07/30 11:42:19 Done.
361 svc_internal_.svc_params.max_quantizers[i] = config_->rc_max_quantizer;
362 svc_internal_.svc_params.min_quantizers[i] = config_->rc_min_quantizer;
363 // 1:2 scaling in each dimension.
364 svc_internal_.svc_params.scaling_factor_num[i] = scaling_factor_num;
365 svc_internal_.svc_params.scaling_factor_den[i] = 256;
366 }
367
368 if (!SetSvcRates()) {
369 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER;
370 }
371
232 if (vpx_codec_enc_init(encoder_, vpx_codec_vp9_cx(), config_, 0)) { 372 if (vpx_codec_enc_init(encoder_, vpx_codec_vp9_cx(), config_, 0)) {
233 return WEBRTC_VIDEO_CODEC_UNINITIALIZED; 373 return WEBRTC_VIDEO_CODEC_UNINITIALIZED;
234 } 374 }
235 vpx_codec_control(encoder_, VP8E_SET_CPUUSED, cpu_speed_); 375 vpx_codec_control(encoder_, VP8E_SET_CPUUSED, cpu_speed_);
236 vpx_codec_control(encoder_, VP8E_SET_MAX_INTRA_BITRATE_PCT, 376 vpx_codec_control(encoder_, VP8E_SET_MAX_INTRA_BITRATE_PCT,
237 rc_max_intra_target_); 377 rc_max_intra_target_);
238 vpx_codec_control(encoder_, VP9E_SET_AQ_MODE, 378 vpx_codec_control(encoder_, VP9E_SET_AQ_MODE,
239 inst->codecSpecific.VP9.adaptiveQpMode ? 3 : 0); 379 inst->codecSpecific.VP9.adaptiveQpMode ? 3 : 0);
380
381 vpx_codec_control(
382 encoder_, VP9E_SET_SVC,
383 (num_temporal_layers_ > 1 || num_spatial_layers_ > 1) ? 1 : 0);
384 if (num_spatial_layers_ > 1) {
385 vpx_codec_control(encoder_, VP9E_SET_SVC_PARAMETERS,
386 &svc_internal_.svc_params);
stefan-webrtc 2015/07/29 13:13:05 Should this not be called if num_temporal_layers_
åsapersson 2015/07/30 11:42:19 Done.
387 }
388 // Register callback for getting each spatial layer.
389 vpx_codec_priv_output_cx_pkt_cb_pair_t cbp = {
390 VP9EncoderImpl::EncoderOutputCodedPacketCallback, (void*)(this)};
391 vpx_codec_control(encoder_, VP9E_REGISTER_CX_CALLBACK, (void*)(&cbp));
392
240 // Control function to set the number of column tiles in encoding a frame, in 393 // Control function to set the number of column tiles in encoding a frame, in
241 // log2 unit: e.g., 0 = 1 tile column, 1 = 2 tile columns, 2 = 4 tile columns. 394 // log2 unit: e.g., 0 = 1 tile column, 1 = 2 tile columns, 2 = 4 tile columns.
242 // The number tile columns will be capped by the encoder based on image size 395 // The number tile columns will be capped by the encoder based on image size
243 // (minimum width of tile column is 256 pixels, maximum is 4096). 396 // (minimum width of tile column is 256 pixels, maximum is 4096).
244 vpx_codec_control(encoder_, VP9E_SET_TILE_COLUMNS, (config_->g_threads >> 1)); 397 vpx_codec_control(encoder_, VP9E_SET_TILE_COLUMNS, (config_->g_threads >> 1));
245 #if !defined(WEBRTC_ARCH_ARM) && !defined(WEBRTC_ARCH_ARM64) 398 #if !defined(WEBRTC_ARCH_ARM) && !defined(WEBRTC_ARCH_ARM64)
246 // Note denoiser is still off by default until further testing/optimization, 399 // Note denoiser is still off by default until further testing/optimization,
247 // i.e., codecSpecific.VP9.denoisingOn == 0. 400 // i.e., codecSpecific.VP9.denoisingOn == 0.
248 vpx_codec_control(encoder_, VP9E_SET_NOISE_SENSITIVITY, 401 vpx_codec_control(encoder_, VP9E_SET_NOISE_SENSITIVITY,
249 inst->codecSpecific.VP9.denoisingOn ? 1 : 0); 402 inst->codecSpecific.VP9.denoisingOn ? 1 : 0);
(...skipping 29 matching lines...) Expand all
279 if (encoded_complete_callback_ == NULL) { 432 if (encoded_complete_callback_ == NULL) {
280 return WEBRTC_VIDEO_CODEC_UNINITIALIZED; 433 return WEBRTC_VIDEO_CODEC_UNINITIALIZED;
281 } 434 }
282 VideoFrameType frame_type = kDeltaFrame; 435 VideoFrameType frame_type = kDeltaFrame;
283 // We only support one stream at the moment. 436 // We only support one stream at the moment.
284 if (frame_types && frame_types->size() > 0) { 437 if (frame_types && frame_types->size() > 0) {
285 frame_type = (*frame_types)[0]; 438 frame_type = (*frame_types)[0];
286 } 439 }
287 DCHECK_EQ(input_image.width(), static_cast<int>(raw_->d_w)); 440 DCHECK_EQ(input_image.width(), static_cast<int>(raw_->d_w));
288 DCHECK_EQ(input_image.height(), static_cast<int>(raw_->d_h)); 441 DCHECK_EQ(input_image.height(), static_cast<int>(raw_->d_h));
442
443 // Set input image for use in the callback.
444 // This was necessary since you need some information from input_image.
445 // You can save only the necessary information (such as timestamp) instead of
446 // doing this.
447 input_image_ = &input_image;
448
289 // Image in vpx_image_t format. 449 // Image in vpx_image_t format.
290 // Input image is const. VPX's raw image is not defined as const. 450 // Input image is const. VPX's raw image is not defined as const.
291 raw_->planes[VPX_PLANE_Y] = const_cast<uint8_t*>(input_image.buffer(kYPlane)); 451 raw_->planes[VPX_PLANE_Y] = const_cast<uint8_t*>(input_image.buffer(kYPlane));
292 raw_->planes[VPX_PLANE_U] = const_cast<uint8_t*>(input_image.buffer(kUPlane)); 452 raw_->planes[VPX_PLANE_U] = const_cast<uint8_t*>(input_image.buffer(kUPlane));
293 raw_->planes[VPX_PLANE_V] = const_cast<uint8_t*>(input_image.buffer(kVPlane)); 453 raw_->planes[VPX_PLANE_V] = const_cast<uint8_t*>(input_image.buffer(kVPlane));
294 raw_->stride[VPX_PLANE_Y] = input_image.stride(kYPlane); 454 raw_->stride[VPX_PLANE_Y] = input_image.stride(kYPlane);
295 raw_->stride[VPX_PLANE_U] = input_image.stride(kUPlane); 455 raw_->stride[VPX_PLANE_U] = input_image.stride(kUPlane);
296 raw_->stride[VPX_PLANE_V] = input_image.stride(kVPlane); 456 raw_->stride[VPX_PLANE_V] = input_image.stride(kVPlane);
297 457
298 int flags = 0; 458 int flags = 0;
299 bool send_keyframe = (frame_type == kKeyFrame); 459 bool send_keyframe = (frame_type == kKeyFrame);
300 if (send_keyframe) { 460 if (send_keyframe) {
301 // Key frame request from caller. 461 // Key frame request from caller.
302 flags = VPX_EFLAG_FORCE_KF; 462 flags = VPX_EFLAG_FORCE_KF;
303 } 463 }
304 assert(codec_.maxFramerate > 0); 464 assert(codec_.maxFramerate > 0);
305 uint32_t duration = 90000 / codec_.maxFramerate; 465 uint32_t duration = 90000 / codec_.maxFramerate;
306 if (vpx_codec_encode(encoder_, raw_, timestamp_, duration, flags, 466 if (vpx_codec_encode(encoder_, raw_, timestamp_, duration, flags,
307 VPX_DL_REALTIME)) { 467 VPX_DL_REALTIME)) {
308 return WEBRTC_VIDEO_CODEC_ERROR; 468 return WEBRTC_VIDEO_CODEC_ERROR;
309 } 469 }
310 timestamp_ += duration; 470 timestamp_ += duration;
311 return GetEncodedPartitions(input_image); 471
472 return WEBRTC_VIDEO_CODEC_OK;
312 } 473 }
313 474
314 void VP9EncoderImpl::PopulateCodecSpecific(CodecSpecificInfo* codec_specific, 475 void VP9EncoderImpl::PopulateCodecSpecific(CodecSpecificInfo* codec_specific,
315 const vpx_codec_cx_pkt& pkt, 476 const vpx_codec_cx_pkt& pkt,
316 uint32_t timestamp) { 477 uint32_t timestamp) {
317 assert(codec_specific != NULL); 478 assert(codec_specific != NULL);
318 codec_specific->codecType = kVideoCodecVP9; 479 codec_specific->codecType = kVideoCodecVP9;
319 CodecSpecificInfoVP9 *vp9_info = &(codec_specific->codecSpecific.VP9); 480 CodecSpecificInfoVP9 *vp9_info = &(codec_specific->codecSpecific.VP9);
320 vp9_info->pictureId = picture_id_; 481 // TODO(asapersson): Set correct values.
321 vp9_info->keyIdx = kNoKeyIdx; 482 vp9_info->inter_pic_predicted =
322 vp9_info->nonReference = (pkt.data.frame.flags & VPX_FRAME_IS_DROPPABLE) != 0; 483 (pkt.data.frame.flags & VPX_FRAME_IS_KEY) ? false : true;
323 // TODO(marpan): Temporal layers are supported in the current VP9 version, 484 vp9_info->flexible_mode = codec_.codecSpecific.VP9.flexibleMode;
324 // but for now use 1 temporal layer encoding. Will update this when temporal 485 vp9_info->beginning_of_frame = true;
325 // layer support for VP9 is added in webrtc. 486 vp9_info->end_of_frame = true;
stefan-webrtc 2015/07/29 13:13:05 Seems like these two flags aren't really needed on
åsapersson 2015/07/30 11:42:19 Removed these settings from CodecSpecificInfo and
326 vp9_info->temporalIdx = kNoTemporalIdx; 487 vp9_info->ss_data_available =
327 vp9_info->layerSync = false; 488 (pkt.data.frame.flags & VPX_FRAME_IS_KEY) ? true : false;
328 vp9_info->tl0PicIdx = kNoTl0PicIdx; 489
329 picture_id_ = (picture_id_ + 1) & 0x7FFF; 490 vpx_svc_layer_id_t layer_id = {0};
491 vpx_codec_control(encoder_, VP9E_GET_SVC_LAYER_ID, &layer_id);
stefan-webrtc 2015/07/29 13:13:05 Do you know how this works? Does it return the spa
åsapersson 2015/07/30 11:42:19 Yes it looks so.
492
493 assert(num_temporal_layers_ > 0);
494 assert(num_spatial_layers_ > 0);
495 if (num_temporal_layers_ == 1) {
496 assert(layer_id.temporal_layer_id == 0);
497 vp9_info->temporal_idx = kNoTemporalIdx;
498 } else {
499 vp9_info->temporal_idx = layer_id.temporal_layer_id;
500 }
501 if (num_spatial_layers_ == 1) {
502 assert(layer_id.spatial_layer_id == 0);
503 vp9_info->spatial_idx = kNoSpatialIdx;
504 } else {
505 vp9_info->spatial_idx = layer_id.spatial_layer_id;
506 }
507 if (layer_id.spatial_layer_id != 0) {
508 vp9_info->ss_data_available = false;
509 }
stefan-webrtc 2015/07/29 13:13:05 Seems like we should DCHECK(layer_id.spatial_layer
åsapersson 2015/07/30 11:42:19 Skipped this as discussed offline.
510
511 if (vp9_info->flexible_mode) {
512 vp9_info->gof_idx = kNoGofIdx;
513 } else {
514 vp9_info->gof_idx = gof_idx_++ % gof_.num_frames_in_gof;
515 }
516
517 // TODO(asapersson): this info has to be obtained from the encoder.
518 vp9_info->temporal_up_switch = true;
519
520 if (layer_id.spatial_layer_id == 0) {
521 picture_id_ = (picture_id_ + 1) & 0x7FFF;
522 // TODO(asapersson): this info has to be obtained from the encoder.
523 vp9_info->inter_layer_predicted = false;
524 } else {
525 // TODO(asapersson): this info has to be obtained from the encoder.
526 vp9_info->inter_layer_predicted = true;
527 }
528
529 vp9_info->picture_id = picture_id_;
530
531 if (!vp9_info->flexible_mode) {
532 if (layer_id.temporal_layer_id == 0 && layer_id.spatial_layer_id == 0) {
533 tl0_pic_idx_++;
534 }
535 vp9_info->tl0_pic_idx = tl0_pic_idx_;
536 }
537
538 if (vp9_info->ss_data_available) {
539 vp9_info->num_spatial_layers = num_spatial_layers_;
540 vp9_info->spatial_layer_resolution_present = true;
541 for (uint8_t i = 0; i < vp9_info->num_spatial_layers; i++) {
stefan-webrtc 2015/07/29 13:13:05 ++i and make i an int.
åsapersson 2015/07/30 11:42:19 Changed to size_t which num_spatial_layers is.
542 vp9_info->width[i] = codec_.width *
543 svc_internal_.svc_params.scaling_factor_num[i] /
544 svc_internal_.svc_params.scaling_factor_den[i];
545 vp9_info->height[i] = codec_.height *
546 svc_internal_.svc_params.scaling_factor_num[i] /
547 svc_internal_.svc_params.scaling_factor_den[i];
548 }
549 if (!vp9_info->flexible_mode) {
550 vp9_info->gof.CopyGofInfoVP9(gof_);
551 }
552 }
330 } 553 }
331 554
332 int VP9EncoderImpl::GetEncodedPartitions(const VideoFrame& input_image) { 555 int VP9EncoderImpl::GetEncodedLayerFrame(const vpx_codec_cx_pkt* pkt) {
333 vpx_codec_iter_t iter = NULL;
334 encoded_image_._length = 0; 556 encoded_image_._length = 0;
335 encoded_image_._frameType = kDeltaFrame; 557 encoded_image_._frameType = kDeltaFrame;
336 RTPFragmentationHeader frag_info; 558 RTPFragmentationHeader frag_info;
337 // Note: no data partitioning in VP9, so 1 partition only. We keep this 559 // Note: no data partitioning in VP9, so 1 partition only. We keep this
338 // fragmentation data for now, until VP9 packetizer is implemented. 560 // fragmentation data for now, until VP9 packetizer is implemented.
339 frag_info.VerifyAndAllocateFragmentationHeader(1); 561 frag_info.VerifyAndAllocateFragmentationHeader(1);
340 int part_idx = 0; 562 int part_idx = 0;
341 CodecSpecificInfo codec_specific; 563 CodecSpecificInfo codec_specific;
342 const vpx_codec_cx_pkt_t *pkt = NULL; 564
343 while ((pkt = vpx_codec_get_cx_data(encoder_, &iter)) != NULL) { 565 assert(pkt->kind == VPX_CODEC_CX_FRAME_PKT);
344 switch (pkt->kind) { 566 memcpy(&encoded_image_._buffer[encoded_image_._length], pkt->data.frame.buf,
345 case VPX_CODEC_CX_FRAME_PKT: { 567 pkt->data.frame.sz);
346 memcpy(&encoded_image_._buffer[encoded_image_._length], 568 frag_info.fragmentationOffset[part_idx] = encoded_image_._length;
347 pkt->data.frame.buf, 569 frag_info.fragmentationLength[part_idx] =
348 pkt->data.frame.sz); 570 static_cast<uint32_t>(pkt->data.frame.sz);
349 frag_info.fragmentationOffset[part_idx] = encoded_image_._length; 571 frag_info.fragmentationPlType[part_idx] = 0;
350 frag_info.fragmentationLength[part_idx] = 572 frag_info.fragmentationTimeDiff[part_idx] = 0;
351 static_cast<uint32_t>(pkt->data.frame.sz); 573 encoded_image_._length += static_cast<uint32_t>(pkt->data.frame.sz);
352 frag_info.fragmentationPlType[part_idx] = 0; 574 assert(encoded_image_._length <= encoded_image_._size);
353 frag_info.fragmentationTimeDiff[part_idx] = 0; 575
354 encoded_image_._length += static_cast<uint32_t>(pkt->data.frame.sz); 576 // End of frame.
355 assert(encoded_image_._length <= encoded_image_._size); 577 if ((pkt->data.frame.flags & VPX_FRAME_IS_FRAGMENT) == 0) {
stefan-webrtc 2015/07/29 13:13:05 I doubt this check is needed. Feel free to try to
åsapersson 2015/07/30 11:42:19 Seems ok to remove, done.
356 break; 578 // Check if encoded frame is a key frame.
357 } 579 if (pkt->data.frame.flags & VPX_FRAME_IS_KEY) {
358 default: { 580 encoded_image_._frameType = kKeyFrame;
359 break;
360 }
361 } 581 }
362 // End of frame. 582 PopulateCodecSpecific(&codec_specific, *pkt, input_image_->timestamp());
363 if ((pkt->data.frame.flags & VPX_FRAME_IS_FRAGMENT) == 0) {
364 // Check if encoded frame is a key frame.
365 if (pkt->data.frame.flags & VPX_FRAME_IS_KEY) {
366 encoded_image_._frameType = kKeyFrame;
367 }
368 PopulateCodecSpecific(&codec_specific, *pkt, input_image.timestamp());
369 break;
370 }
371 } 583 }
372 if (encoded_image_._length > 0) { 584 if (encoded_image_._length > 0) {
373 TRACE_COUNTER1("webrtc", "EncodedFrameSize", encoded_image_._length); 585 TRACE_COUNTER1("webrtc", "EncodedFrameSize", encoded_image_._length);
374 encoded_image_._timeStamp = input_image.timestamp(); 586 encoded_image_._timeStamp = input_image_->timestamp();
375 encoded_image_.capture_time_ms_ = input_image.render_time_ms(); 587 encoded_image_.capture_time_ms_ = input_image_->render_time_ms();
376 encoded_image_._encodedHeight = raw_->d_h; 588 encoded_image_._encodedHeight = raw_->d_h;
377 encoded_image_._encodedWidth = raw_->d_w; 589 encoded_image_._encodedWidth = raw_->d_w;
378 encoded_complete_callback_->Encoded(encoded_image_, &codec_specific, 590 encoded_complete_callback_->Encoded(encoded_image_, &codec_specific,
379 &frag_info); 591 &frag_info);
380 } 592 }
381 return WEBRTC_VIDEO_CODEC_OK; 593 return WEBRTC_VIDEO_CODEC_OK;
382 } 594 }
383 595
384 int VP9EncoderImpl::SetChannelParameters(uint32_t packet_loss, int64_t rtt) { 596 int VP9EncoderImpl::SetChannelParameters(uint32_t packet_loss, int64_t rtt) {
385 return WEBRTC_VIDEO_CODEC_OK; 597 return WEBRTC_VIDEO_CODEC_OK;
386 } 598 }
387 599
388 int VP9EncoderImpl::RegisterEncodeCompleteCallback( 600 int VP9EncoderImpl::RegisterEncodeCompleteCallback(
389 EncodedImageCallback* callback) { 601 EncodedImageCallback* callback) {
(...skipping 169 matching lines...) Expand 10 before | Expand all | Expand 10 after
559 decoder_ = NULL; 771 decoder_ = NULL;
560 } 772 }
561 // Releases buffers from the pool. Any buffers not in use are deleted. Buffers 773 // Releases buffers from the pool. Any buffers not in use are deleted. Buffers
562 // still referenced externally are deleted once fully released, not returning 774 // still referenced externally are deleted once fully released, not returning
563 // to the pool. 775 // to the pool.
564 frame_buffer_pool_.ClearPool(); 776 frame_buffer_pool_.ClearPool();
565 inited_ = false; 777 inited_ = false;
566 return WEBRTC_VIDEO_CODEC_OK; 778 return WEBRTC_VIDEO_CODEC_OK;
567 } 779 }
568 } // namespace webrtc 780 } // namespace webrtc
OLDNEW

Powered by Google App Engine
This is Rietveld 408576698