Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(89)

Side by Side Diff: webrtc/modules/video_coding/codecs/vp9/vp9_impl.cc

Issue 1211353002: Integration of VP9 packetization. (Closed) Base URL: https://chromium.googlesource.com/external/webrtc.git@master
Patch Set: allow 1-3 temporal layers Created 5 years, 4 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch
OLDNEW
1 /* 1 /*
2 * Copyright (c) 2014 The WebRTC project authors. All Rights Reserved. 2 * Copyright (c) 2014 The WebRTC project authors. All Rights Reserved.
3 * 3 *
4 * Use of this source code is governed by a BSD-style license 4 * Use of this source code is governed by a BSD-style license
5 * that can be found in the LICENSE file in the root of the source 5 * that can be found in the LICENSE file in the root of the source
6 * tree. An additional intellectual property rights grant can be found 6 * tree. An additional intellectual property rights grant can be found
7 * in the file PATENTS. All contributing project authors may 7 * in the file PATENTS. All contributing project authors may
8 * be found in the AUTHORS file in the root of the source tree. 8 * be found in the AUTHORS file in the root of the source tree.
9 * 9 *
10 */ 10 */
(...skipping 39 matching lines...) Expand 10 before | Expand all | Expand 10 after
50 if (width * height <= 352 * 288) 50 if (width * height <= 352 * 288)
51 return 5; 51 return 5;
52 else 52 else
53 return 7; 53 return 7;
54 } 54 }
55 55
56 VP9Encoder* VP9Encoder::Create() { 56 VP9Encoder* VP9Encoder::Create() {
57 return new VP9EncoderImpl(); 57 return new VP9EncoderImpl();
58 } 58 }
59 59
60 void VP9EncoderImpl::EncoderOutputCodedPacketCallback(vpx_codec_cx_pkt* pkt,
61 void* user_data) {
62 VP9EncoderImpl* enc = (VP9EncoderImpl*)(user_data);
63 enc->GetEncodedLayerFrame(pkt);
64 }
65
60 VP9EncoderImpl::VP9EncoderImpl() 66 VP9EncoderImpl::VP9EncoderImpl()
61 : encoded_image_(), 67 : encoded_image_(),
62 encoded_complete_callback_(NULL), 68 encoded_complete_callback_(NULL),
63 inited_(false), 69 inited_(false),
64 timestamp_(0), 70 timestamp_(0),
65 picture_id_(0), 71 picture_id_(0),
66 cpu_speed_(3), 72 cpu_speed_(3),
67 rc_max_intra_target_(0), 73 rc_max_intra_target_(0),
68 encoder_(NULL), 74 encoder_(NULL),
69 config_(NULL), 75 config_(NULL),
70 raw_(NULL) { 76 raw_(NULL),
77 input_image_(NULL),
78 tl0_pic_idx_(0),
79 gof_idx_(0),
80 num_temporal_layers_(0),
81 num_spatial_layers_(0) {
71 memset(&codec_, 0, sizeof(codec_)); 82 memset(&codec_, 0, sizeof(codec_));
72 uint32_t seed = static_cast<uint32_t>(TickTime::MillisecondTimestamp()); 83 uint32_t seed = static_cast<uint32_t>(TickTime::MillisecondTimestamp());
73 srand(seed); 84 srand(seed);
74 } 85 }
75 86
76 VP9EncoderImpl::~VP9EncoderImpl() { 87 VP9EncoderImpl::~VP9EncoderImpl() {
77 Release(); 88 Release();
78 } 89 }
79 90
80 int VP9EncoderImpl::Release() { 91 int VP9EncoderImpl::Release() {
(...skipping 13 matching lines...) Expand all
94 config_ = NULL; 105 config_ = NULL;
95 } 106 }
96 if (raw_ != NULL) { 107 if (raw_ != NULL) {
97 vpx_img_free(raw_); 108 vpx_img_free(raw_);
98 raw_ = NULL; 109 raw_ = NULL;
99 } 110 }
100 inited_ = false; 111 inited_ = false;
101 return WEBRTC_VIDEO_CODEC_OK; 112 return WEBRTC_VIDEO_CODEC_OK;
102 } 113 }
103 114
115 bool VP9EncoderImpl::SetSvcRates() {
116 float rate_ratio[VPX_MAX_LAYERS] = {0};
117 float total = 0;
118 uint8_t i = 0;
119
120 for (i = 0; i < num_spatial_layers_; ++i) {
121 if (svc_internal_.svc_params.scaling_factor_num[i] <= 0 ||
122 svc_internal_.svc_params.scaling_factor_den[i] <= 0) {
123 return false;
124 }
125 rate_ratio[i] = static_cast<float>(
126 svc_internal_.svc_params.scaling_factor_num[i]) /
127 svc_internal_.svc_params.scaling_factor_den[i];
128 total += rate_ratio[i];
129 }
130
131 for (i = 0; i < num_spatial_layers_; ++i) {
132 config_->ss_target_bitrate[i] = static_cast<unsigned int>(
133 config_->rc_target_bitrate * rate_ratio[i] / total);
134 if (num_temporal_layers_ == 1) {
135 config_->layer_target_bitrate[i] = config_->ss_target_bitrate[i];
136 } else if (num_temporal_layers_ == 2) {
137 config_->layer_target_bitrate[i * num_temporal_layers_] =
138 config_->ss_target_bitrate[i] * 2 / 3;
139 config_->layer_target_bitrate[i * num_temporal_layers_ + 1] =
140 config_->ss_target_bitrate[i];
141 } else if (num_temporal_layers_ == 3) {
142 config_->layer_target_bitrate[i * num_temporal_layers_] =
143 config_->ss_target_bitrate[i] / 2;
144 config_->layer_target_bitrate[i * num_temporal_layers_ + 1] =
145 config_->layer_target_bitrate[i * num_temporal_layers_] +
146 (config_->ss_target_bitrate[i] / 4);
147 config_->layer_target_bitrate[i * num_temporal_layers_ + 2] =
148 config_->ss_target_bitrate[i];
149 } else {
150 return false;
151 }
152 }
153
154 // For now, temporal layers only supported when having one spatial layer.
155 if (num_spatial_layers_ == 1) {
156 for (i = 0; i < num_temporal_layers_; ++i) {
157 config_->ts_target_bitrate[i] = config_->layer_target_bitrate[i];
158 }
159 }
160
161 return true;
162 }
163
104 int VP9EncoderImpl::SetRates(uint32_t new_bitrate_kbit, 164 int VP9EncoderImpl::SetRates(uint32_t new_bitrate_kbit,
105 uint32_t new_framerate) { 165 uint32_t new_framerate) {
106 if (!inited_) { 166 if (!inited_) {
107 return WEBRTC_VIDEO_CODEC_UNINITIALIZED; 167 return WEBRTC_VIDEO_CODEC_UNINITIALIZED;
108 } 168 }
109 if (encoder_->err) { 169 if (encoder_->err) {
110 return WEBRTC_VIDEO_CODEC_ERROR; 170 return WEBRTC_VIDEO_CODEC_ERROR;
111 } 171 }
112 if (new_framerate < 1) { 172 if (new_framerate < 1) {
113 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; 173 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER;
114 } 174 }
115 // Update bit rate 175 // Update bit rate
116 if (codec_.maxBitrate > 0 && new_bitrate_kbit > codec_.maxBitrate) { 176 if (codec_.maxBitrate > 0 && new_bitrate_kbit > codec_.maxBitrate) {
117 new_bitrate_kbit = codec_.maxBitrate; 177 new_bitrate_kbit = codec_.maxBitrate;
118 } 178 }
119 config_->rc_target_bitrate = new_bitrate_kbit; 179 config_->rc_target_bitrate = new_bitrate_kbit;
120 codec_.maxFramerate = new_framerate; 180 codec_.maxFramerate = new_framerate;
181
182 if (!SetSvcRates()) {
183 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER;
184 }
185
121 // Update encoder context 186 // Update encoder context
122 if (vpx_codec_enc_config_set(encoder_, config_)) { 187 if (vpx_codec_enc_config_set(encoder_, config_)) {
123 return WEBRTC_VIDEO_CODEC_ERROR; 188 return WEBRTC_VIDEO_CODEC_ERROR;
124 } 189 }
125 return WEBRTC_VIDEO_CODEC_OK; 190 return WEBRTC_VIDEO_CODEC_OK;
126 } 191 }
127 192
128 int VP9EncoderImpl::InitEncode(const VideoCodec* inst, 193 int VP9EncoderImpl::InitEncode(const VideoCodec* inst,
129 int number_of_cores, 194 int number_of_cores,
130 size_t /*max_payload_size*/) { 195 size_t /*max_payload_size*/) {
131 if (inst == NULL) { 196 if (inst == NULL) {
132 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; 197 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER;
133 } 198 }
134 if (inst->maxFramerate < 1) { 199 if (inst->maxFramerate < 1) {
135 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; 200 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER;
136 } 201 }
137 // Allow zero to represent an unspecified maxBitRate 202 // Allow zero to represent an unspecified maxBitRate
138 if (inst->maxBitrate > 0 && inst->startBitrate > inst->maxBitrate) { 203 if (inst->maxBitrate > 0 && inst->startBitrate > inst->maxBitrate) {
139 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; 204 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER;
140 } 205 }
141 if (inst->width < 1 || inst->height < 1) { 206 if (inst->width < 1 || inst->height < 1) {
142 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; 207 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER;
143 } 208 }
144 if (number_of_cores < 1) { 209 if (number_of_cores < 1) {
145 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; 210 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER;
146 } 211 }
212 if (inst->codecSpecific.VP9.numberOfTemporalLayers < 1 ||
holmer 2015/07/31 09:05:26 Should 0 be allowed as "no temporal layers"?
åsapersson 2015/07/31 10:09:37 See change.
213 inst->codecSpecific.VP9.numberOfTemporalLayers > 3) {
214 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER;
215 }
216 // For now, only support one spatial layer.
217 if (inst->codecSpecific.VP9.numberOfSpatialLayers != 1) {
218 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER;
219 }
147 int retVal = Release(); 220 int retVal = Release();
148 if (retVal < 0) { 221 if (retVal < 0) {
149 return retVal; 222 return retVal;
150 } 223 }
151 if (encoder_ == NULL) { 224 if (encoder_ == NULL) {
152 encoder_ = new vpx_codec_ctx_t; 225 encoder_ = new vpx_codec_ctx_t;
153 } 226 }
154 if (config_ == NULL) { 227 if (config_ == NULL) {
155 config_ = new vpx_codec_enc_cfg_t; 228 config_ = new vpx_codec_enc_cfg_t;
156 } 229 }
157 timestamp_ = 0; 230 timestamp_ = 0;
158 if (&codec_ != inst) { 231 if (&codec_ != inst) {
159 codec_ = *inst; 232 codec_ = *inst;
160 } 233 }
234
235 num_temporal_layers_ = inst->codecSpecific.VP9.numberOfTemporalLayers;
236 num_spatial_layers_ = inst->codecSpecific.VP9.numberOfSpatialLayers;
237
161 // Random start 16 bits is enough. 238 // Random start 16 bits is enough.
162 picture_id_ = static_cast<uint16_t>(rand()) & 0x7FFF; 239 picture_id_ = static_cast<uint16_t>(rand()) & 0x7FFF;
163 // Allocate memory for encoded image 240 // Allocate memory for encoded image
164 if (encoded_image_._buffer != NULL) { 241 if (encoded_image_._buffer != NULL) {
165 delete [] encoded_image_._buffer; 242 delete [] encoded_image_._buffer;
166 } 243 }
167 encoded_image_._size = CalcBufferSize(kI420, codec_.width, codec_.height); 244 encoded_image_._size = CalcBufferSize(kI420, codec_.width, codec_.height);
168 encoded_image_._buffer = new uint8_t[encoded_image_._size]; 245 encoded_image_._buffer = new uint8_t[encoded_image_._size];
169 encoded_image_._completeFrame = true; 246 encoded_image_._completeFrame = true;
170 // Creating a wrapper to the image - setting image data to NULL. Actual 247 // Creating a wrapper to the image - setting image data to NULL. Actual
(...skipping 31 matching lines...) Expand 10 before | Expand all | Expand 10 after
202 if (inst->codecSpecific.VP9.keyFrameInterval > 0) { 279 if (inst->codecSpecific.VP9.keyFrameInterval > 0) {
203 config_->kf_mode = VPX_KF_AUTO; 280 config_->kf_mode = VPX_KF_AUTO;
204 config_->kf_max_dist = inst->codecSpecific.VP9.keyFrameInterval; 281 config_->kf_max_dist = inst->codecSpecific.VP9.keyFrameInterval;
205 } else { 282 } else {
206 config_->kf_mode = VPX_KF_DISABLED; 283 config_->kf_mode = VPX_KF_DISABLED;
207 } 284 }
208 // Determine number of threads based on the image size and #cores. 285 // Determine number of threads based on the image size and #cores.
209 config_->g_threads = NumberOfThreads(config_->g_w, 286 config_->g_threads = NumberOfThreads(config_->g_w,
210 config_->g_h, 287 config_->g_h,
211 number_of_cores); 288 number_of_cores);
289
212 cpu_speed_ = GetCpuSpeed(config_->g_w, config_->g_h); 290 cpu_speed_ = GetCpuSpeed(config_->g_w, config_->g_h);
291
292 // TODO(asapersson): Check configuration of temporal switch up and increase
293 // pattern length.
294 if (num_temporal_layers_ == 1) {
295 gof_.SetGofInfoVP9(kTemporalStructureMode1);
296 config_->temporal_layering_mode = VP9E_TEMPORAL_LAYERING_MODE_NOLAYERING;
297 config_->ts_number_layers = 1;
298 config_->ts_rate_decimator[0] = 1;
299 config_->ts_periodicity = 1;
300 config_->ts_layer_id[0] = 0;
301 } else if (num_temporal_layers_ == 2) {
302 gof_.SetGofInfoVP9(kTemporalStructureMode2);
303 config_->temporal_layering_mode = VP9E_TEMPORAL_LAYERING_MODE_0101;
304 config_->ts_number_layers = 2;
305 config_->ts_rate_decimator[0] = 2;
306 config_->ts_rate_decimator[1] = 1;
307 config_->ts_periodicity = 2;
308 config_->ts_layer_id[0] = 0;
309 config_->ts_layer_id[1] = 1;
310 } else if (num_temporal_layers_ == 3) {
311 gof_.SetGofInfoVP9(kTemporalStructureMode3);
312 config_->temporal_layering_mode = VP9E_TEMPORAL_LAYERING_MODE_0212;
313 config_->ts_number_layers = 3;
314 config_->ts_rate_decimator[0] = 4;
315 config_->ts_rate_decimator[1] = 2;
316 config_->ts_rate_decimator[2] = 1;
317 config_->ts_periodicity = 4;
318 config_->ts_layer_id[0] = 0;
319 config_->ts_layer_id[1] = 2;
320 config_->ts_layer_id[2] = 1;
321 config_->ts_layer_id[3] = 2;
322 } else {
323 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER;
324 }
325
326 tl0_pic_idx_ = static_cast<uint8_t>(rand());
327
213 return InitAndSetControlSettings(inst); 328 return InitAndSetControlSettings(inst);
214 } 329 }
215 330
216 int VP9EncoderImpl::NumberOfThreads(int width, 331 int VP9EncoderImpl::NumberOfThreads(int width,
217 int height, 332 int height,
218 int number_of_cores) { 333 int number_of_cores) {
334 // For the current libvpx library, only 1 thread is supported when SVC is
335 // turned on.
336 if (num_temporal_layers_ > 1 || num_spatial_layers_ > 1) {
337 return 1;
338 }
339
219 // Keep the number of encoder threads equal to the possible number of column 340 // Keep the number of encoder threads equal to the possible number of column
220 // tiles, which is (1, 2, 4, 8). See comments below for VP9E_SET_TILE_COLUMNS. 341 // tiles, which is (1, 2, 4, 8). See comments below for VP9E_SET_TILE_COLUMNS.
221 if (width * height >= 1280 * 720 && number_of_cores > 4) { 342 if (width * height >= 1280 * 720 && number_of_cores > 4) {
222 return 4; 343 return 4;
223 } else if (width * height >= 640 * 480 && number_of_cores > 2) { 344 } else if (width * height >= 640 * 480 && number_of_cores > 2) {
224 return 2; 345 return 2;
225 } else { 346 } else {
226 // 1 thread less than VGA. 347 // 1 thread less than VGA.
227 return 1; 348 return 1;
228 } 349 }
229 } 350 }
230 351
231 int VP9EncoderImpl::InitAndSetControlSettings(const VideoCodec* inst) { 352 int VP9EncoderImpl::InitAndSetControlSettings(const VideoCodec* inst) {
353
354 config_->ss_number_layers = num_spatial_layers_;
355
356 if (num_spatial_layers_ > 1) {
357 config_->rc_min_quantizer = 0;
358 config_->rc_max_quantizer = 63;
359 }
360 int scaling_factor_num = 256;
361 for (int i = num_spatial_layers_ - 1; i >= 0; --i) {
362 svc_internal_.svc_params.max_quantizers[i] = config_->rc_max_quantizer;
363 svc_internal_.svc_params.min_quantizers[i] = config_->rc_min_quantizer;
364 // 1:2 scaling in each dimension.
365 svc_internal_.svc_params.scaling_factor_num[i] = scaling_factor_num;
366 svc_internal_.svc_params.scaling_factor_den[i] = 256;
367 scaling_factor_num /= 2;
368 }
369
370 if (!SetSvcRates()) {
371 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER;
372 }
373
232 if (vpx_codec_enc_init(encoder_, vpx_codec_vp9_cx(), config_, 0)) { 374 if (vpx_codec_enc_init(encoder_, vpx_codec_vp9_cx(), config_, 0)) {
233 return WEBRTC_VIDEO_CODEC_UNINITIALIZED; 375 return WEBRTC_VIDEO_CODEC_UNINITIALIZED;
234 } 376 }
235 vpx_codec_control(encoder_, VP8E_SET_CPUUSED, cpu_speed_); 377 vpx_codec_control(encoder_, VP8E_SET_CPUUSED, cpu_speed_);
236 vpx_codec_control(encoder_, VP8E_SET_MAX_INTRA_BITRATE_PCT, 378 vpx_codec_control(encoder_, VP8E_SET_MAX_INTRA_BITRATE_PCT,
237 rc_max_intra_target_); 379 rc_max_intra_target_);
238 vpx_codec_control(encoder_, VP9E_SET_AQ_MODE, 380 vpx_codec_control(encoder_, VP9E_SET_AQ_MODE,
239 inst->codecSpecific.VP9.adaptiveQpMode ? 3 : 0); 381 inst->codecSpecific.VP9.adaptiveQpMode ? 3 : 0);
382
383 vpx_codec_control(
384 encoder_, VP9E_SET_SVC,
385 (num_temporal_layers_ > 1 || num_spatial_layers_ > 1) ? 1 : 0);
386 if (num_temporal_layers_ > 1 || num_spatial_layers_ > 1) {
387 vpx_codec_control(encoder_, VP9E_SET_SVC_PARAMETERS,
388 &svc_internal_.svc_params);
389 }
390 // Register callback for getting each spatial layer.
391 vpx_codec_priv_output_cx_pkt_cb_pair_t cbp = {
392 VP9EncoderImpl::EncoderOutputCodedPacketCallback, (void*)(this)};
393 vpx_codec_control(encoder_, VP9E_REGISTER_CX_CALLBACK, (void*)(&cbp));
394
240 // Control function to set the number of column tiles in encoding a frame, in 395 // Control function to set the number of column tiles in encoding a frame, in
241 // log2 unit: e.g., 0 = 1 tile column, 1 = 2 tile columns, 2 = 4 tile columns. 396 // log2 unit: e.g., 0 = 1 tile column, 1 = 2 tile columns, 2 = 4 tile columns.
242 // The number tile columns will be capped by the encoder based on image size 397 // The number tile columns will be capped by the encoder based on image size
243 // (minimum width of tile column is 256 pixels, maximum is 4096). 398 // (minimum width of tile column is 256 pixels, maximum is 4096).
244 vpx_codec_control(encoder_, VP9E_SET_TILE_COLUMNS, (config_->g_threads >> 1)); 399 vpx_codec_control(encoder_, VP9E_SET_TILE_COLUMNS, (config_->g_threads >> 1));
245 #if !defined(WEBRTC_ARCH_ARM) && !defined(WEBRTC_ARCH_ARM64) 400 #if !defined(WEBRTC_ARCH_ARM) && !defined(WEBRTC_ARCH_ARM64)
246 // Note denoiser is still off by default until further testing/optimization, 401 // Note denoiser is still off by default until further testing/optimization,
247 // i.e., codecSpecific.VP9.denoisingOn == 0. 402 // i.e., codecSpecific.VP9.denoisingOn == 0.
248 vpx_codec_control(encoder_, VP9E_SET_NOISE_SENSITIVITY, 403 vpx_codec_control(encoder_, VP9E_SET_NOISE_SENSITIVITY,
249 inst->codecSpecific.VP9.denoisingOn ? 1 : 0); 404 inst->codecSpecific.VP9.denoisingOn ? 1 : 0);
(...skipping 29 matching lines...) Expand all
279 if (encoded_complete_callback_ == NULL) { 434 if (encoded_complete_callback_ == NULL) {
280 return WEBRTC_VIDEO_CODEC_UNINITIALIZED; 435 return WEBRTC_VIDEO_CODEC_UNINITIALIZED;
281 } 436 }
282 VideoFrameType frame_type = kDeltaFrame; 437 VideoFrameType frame_type = kDeltaFrame;
283 // We only support one stream at the moment. 438 // We only support one stream at the moment.
284 if (frame_types && frame_types->size() > 0) { 439 if (frame_types && frame_types->size() > 0) {
285 frame_type = (*frame_types)[0]; 440 frame_type = (*frame_types)[0];
286 } 441 }
287 DCHECK_EQ(input_image.width(), static_cast<int>(raw_->d_w)); 442 DCHECK_EQ(input_image.width(), static_cast<int>(raw_->d_w));
288 DCHECK_EQ(input_image.height(), static_cast<int>(raw_->d_h)); 443 DCHECK_EQ(input_image.height(), static_cast<int>(raw_->d_h));
444
445 // Set input image for use in the callback.
446 // This was necessary since you need some information from input_image.
447 // You can save only the necessary information (such as timestamp) instead of
448 // doing this.
449 input_image_ = &input_image;
450
289 // Image in vpx_image_t format. 451 // Image in vpx_image_t format.
290 // Input image is const. VPX's raw image is not defined as const. 452 // Input image is const. VPX's raw image is not defined as const.
291 raw_->planes[VPX_PLANE_Y] = const_cast<uint8_t*>(input_image.buffer(kYPlane)); 453 raw_->planes[VPX_PLANE_Y] = const_cast<uint8_t*>(input_image.buffer(kYPlane));
292 raw_->planes[VPX_PLANE_U] = const_cast<uint8_t*>(input_image.buffer(kUPlane)); 454 raw_->planes[VPX_PLANE_U] = const_cast<uint8_t*>(input_image.buffer(kUPlane));
293 raw_->planes[VPX_PLANE_V] = const_cast<uint8_t*>(input_image.buffer(kVPlane)); 455 raw_->planes[VPX_PLANE_V] = const_cast<uint8_t*>(input_image.buffer(kVPlane));
294 raw_->stride[VPX_PLANE_Y] = input_image.stride(kYPlane); 456 raw_->stride[VPX_PLANE_Y] = input_image.stride(kYPlane);
295 raw_->stride[VPX_PLANE_U] = input_image.stride(kUPlane); 457 raw_->stride[VPX_PLANE_U] = input_image.stride(kUPlane);
296 raw_->stride[VPX_PLANE_V] = input_image.stride(kVPlane); 458 raw_->stride[VPX_PLANE_V] = input_image.stride(kVPlane);
297 459
298 int flags = 0; 460 int flags = 0;
299 bool send_keyframe = (frame_type == kKeyFrame); 461 bool send_keyframe = (frame_type == kKeyFrame);
300 if (send_keyframe) { 462 if (send_keyframe) {
301 // Key frame request from caller. 463 // Key frame request from caller.
302 flags = VPX_EFLAG_FORCE_KF; 464 flags = VPX_EFLAG_FORCE_KF;
303 } 465 }
304 assert(codec_.maxFramerate > 0); 466 assert(codec_.maxFramerate > 0);
305 uint32_t duration = 90000 / codec_.maxFramerate; 467 uint32_t duration = 90000 / codec_.maxFramerate;
306 if (vpx_codec_encode(encoder_, raw_, timestamp_, duration, flags, 468 if (vpx_codec_encode(encoder_, raw_, timestamp_, duration, flags,
307 VPX_DL_REALTIME)) { 469 VPX_DL_REALTIME)) {
308 return WEBRTC_VIDEO_CODEC_ERROR; 470 return WEBRTC_VIDEO_CODEC_ERROR;
309 } 471 }
310 timestamp_ += duration; 472 timestamp_ += duration;
311 return GetEncodedPartitions(input_image); 473
474 return WEBRTC_VIDEO_CODEC_OK;
312 } 475 }
313 476
314 void VP9EncoderImpl::PopulateCodecSpecific(CodecSpecificInfo* codec_specific, 477 void VP9EncoderImpl::PopulateCodecSpecific(CodecSpecificInfo* codec_specific,
315 const vpx_codec_cx_pkt& pkt, 478 const vpx_codec_cx_pkt& pkt,
316 uint32_t timestamp) { 479 uint32_t timestamp) {
317 assert(codec_specific != NULL); 480 assert(codec_specific != NULL);
318 codec_specific->codecType = kVideoCodecVP9; 481 codec_specific->codecType = kVideoCodecVP9;
319 CodecSpecificInfoVP9 *vp9_info = &(codec_specific->codecSpecific.VP9); 482 CodecSpecificInfoVP9 *vp9_info = &(codec_specific->codecSpecific.VP9);
320 vp9_info->pictureId = picture_id_; 483 // TODO(asapersson): Set correct values.
321 vp9_info->keyIdx = kNoKeyIdx; 484 vp9_info->inter_pic_predicted =
322 vp9_info->nonReference = (pkt.data.frame.flags & VPX_FRAME_IS_DROPPABLE) != 0; 485 (pkt.data.frame.flags & VPX_FRAME_IS_KEY) ? false : true;
323 // TODO(marpan): Temporal layers are supported in the current VP9 version, 486 vp9_info->flexible_mode = codec_.codecSpecific.VP9.flexibleMode;
324 // but for now use 1 temporal layer encoding. Will update this when temporal 487 vp9_info->ss_data_available =
325 // layer support for VP9 is added in webrtc. 488 (pkt.data.frame.flags & VPX_FRAME_IS_KEY) ? true : false;
326 vp9_info->temporalIdx = kNoTemporalIdx; 489 if (pkt.data.frame.flags & VPX_FRAME_IS_KEY) {
327 vp9_info->layerSync = false; 490 gof_idx_ = 0;
328 vp9_info->tl0PicIdx = kNoTl0PicIdx; 491 }
329 picture_id_ = (picture_id_ + 1) & 0x7FFF; 492
493 vpx_svc_layer_id_t layer_id = {0};
494 vpx_codec_control(encoder_, VP9E_GET_SVC_LAYER_ID, &layer_id);
495
496 assert(num_temporal_layers_ > 0);
497 assert(num_spatial_layers_ > 0);
498 if (num_temporal_layers_ == 1) {
499 assert(layer_id.temporal_layer_id == 0);
500 vp9_info->temporal_idx = kNoTemporalIdx;
501 } else {
502 vp9_info->temporal_idx = layer_id.temporal_layer_id;
503 }
504 if (num_spatial_layers_ == 1) {
505 assert(layer_id.spatial_layer_id == 0);
506 vp9_info->spatial_idx = kNoSpatialIdx;
507 } else {
508 vp9_info->spatial_idx = layer_id.spatial_layer_id;
509 }
510 if (layer_id.spatial_layer_id != 0) {
511 vp9_info->ss_data_available = false;
512 }
513
514 if (vp9_info->flexible_mode) {
515 vp9_info->gof_idx = kNoGofIdx;
516 } else {
517 vp9_info->gof_idx =
518 static_cast<uint8_t>(gof_idx_++ % gof_.num_frames_in_gof);
519 }
520
521 // TODO(asapersson): this info has to be obtained from the encoder.
522 vp9_info->temporal_up_switch = true;
523
524 if (layer_id.spatial_layer_id == 0) {
525 picture_id_ = (picture_id_ + 1) & 0x7FFF;
526 // TODO(asapersson): this info has to be obtained from the encoder.
527 vp9_info->inter_layer_predicted = false;
528 } else {
529 // TODO(asapersson): this info has to be obtained from the encoder.
530 vp9_info->inter_layer_predicted = true;
531 }
532
533 vp9_info->picture_id = picture_id_;
534
535 if (!vp9_info->flexible_mode) {
536 if (layer_id.temporal_layer_id == 0 && layer_id.spatial_layer_id == 0) {
537 tl0_pic_idx_++;
538 }
539 vp9_info->tl0_pic_idx = tl0_pic_idx_;
540 }
541
542 if (vp9_info->ss_data_available) {
543 vp9_info->num_spatial_layers = num_spatial_layers_;
544 vp9_info->spatial_layer_resolution_present = true;
545 for (size_t i = 0; i < vp9_info->num_spatial_layers; ++i) {
546 vp9_info->width[i] = codec_.width *
547 svc_internal_.svc_params.scaling_factor_num[i] /
548 svc_internal_.svc_params.scaling_factor_den[i];
549 vp9_info->height[i] = codec_.height *
550 svc_internal_.svc_params.scaling_factor_num[i] /
551 svc_internal_.svc_params.scaling_factor_den[i];
552 }
553 if (!vp9_info->flexible_mode) {
554 vp9_info->gof.CopyGofInfoVP9(gof_);
555 }
556 }
330 } 557 }
331 558
332 int VP9EncoderImpl::GetEncodedPartitions(const VideoFrame& input_image) { 559 int VP9EncoderImpl::GetEncodedLayerFrame(const vpx_codec_cx_pkt* pkt) {
333 vpx_codec_iter_t iter = NULL;
334 encoded_image_._length = 0; 560 encoded_image_._length = 0;
335 encoded_image_._frameType = kDeltaFrame; 561 encoded_image_._frameType = kDeltaFrame;
336 RTPFragmentationHeader frag_info; 562 RTPFragmentationHeader frag_info;
337 // Note: no data partitioning in VP9, so 1 partition only. We keep this 563 // Note: no data partitioning in VP9, so 1 partition only. We keep this
338 // fragmentation data for now, until VP9 packetizer is implemented. 564 // fragmentation data for now, until VP9 packetizer is implemented.
339 frag_info.VerifyAndAllocateFragmentationHeader(1); 565 frag_info.VerifyAndAllocateFragmentationHeader(1);
340 int part_idx = 0; 566 int part_idx = 0;
341 CodecSpecificInfo codec_specific; 567 CodecSpecificInfo codec_specific;
342 const vpx_codec_cx_pkt_t *pkt = NULL; 568
343 while ((pkt = vpx_codec_get_cx_data(encoder_, &iter)) != NULL) { 569 assert(pkt->kind == VPX_CODEC_CX_FRAME_PKT);
344 switch (pkt->kind) { 570 memcpy(&encoded_image_._buffer[encoded_image_._length], pkt->data.frame.buf,
345 case VPX_CODEC_CX_FRAME_PKT: { 571 pkt->data.frame.sz);
346 memcpy(&encoded_image_._buffer[encoded_image_._length], 572 frag_info.fragmentationOffset[part_idx] = encoded_image_._length;
347 pkt->data.frame.buf, 573 frag_info.fragmentationLength[part_idx] =
348 pkt->data.frame.sz); 574 static_cast<uint32_t>(pkt->data.frame.sz);
349 frag_info.fragmentationOffset[part_idx] = encoded_image_._length; 575 frag_info.fragmentationPlType[part_idx] = 0;
350 frag_info.fragmentationLength[part_idx] = 576 frag_info.fragmentationTimeDiff[part_idx] = 0;
351 static_cast<uint32_t>(pkt->data.frame.sz); 577 encoded_image_._length += static_cast<uint32_t>(pkt->data.frame.sz);
352 frag_info.fragmentationPlType[part_idx] = 0; 578 assert(encoded_image_._length <= encoded_image_._size);
353 frag_info.fragmentationTimeDiff[part_idx] = 0; 579
354 encoded_image_._length += static_cast<uint32_t>(pkt->data.frame.sz); 580 // End of frame.
355 assert(encoded_image_._length <= encoded_image_._size); 581 // Check if encoded frame is a key frame.
356 break; 582 if (pkt->data.frame.flags & VPX_FRAME_IS_KEY) {
357 } 583 encoded_image_._frameType = kKeyFrame;
358 default: {
359 break;
360 }
361 }
362 // End of frame.
363 if ((pkt->data.frame.flags & VPX_FRAME_IS_FRAGMENT) == 0) {
364 // Check if encoded frame is a key frame.
365 if (pkt->data.frame.flags & VPX_FRAME_IS_KEY) {
366 encoded_image_._frameType = kKeyFrame;
367 }
368 PopulateCodecSpecific(&codec_specific, *pkt, input_image.timestamp());
369 break;
370 }
371 } 584 }
585 PopulateCodecSpecific(&codec_specific, *pkt, input_image_->timestamp());
586
372 if (encoded_image_._length > 0) { 587 if (encoded_image_._length > 0) {
373 TRACE_COUNTER1("webrtc", "EncodedFrameSize", encoded_image_._length); 588 TRACE_COUNTER1("webrtc", "EncodedFrameSize", encoded_image_._length);
374 encoded_image_._timeStamp = input_image.timestamp(); 589 encoded_image_._timeStamp = input_image_->timestamp();
375 encoded_image_.capture_time_ms_ = input_image.render_time_ms(); 590 encoded_image_.capture_time_ms_ = input_image_->render_time_ms();
376 encoded_image_._encodedHeight = raw_->d_h; 591 encoded_image_._encodedHeight = raw_->d_h;
377 encoded_image_._encodedWidth = raw_->d_w; 592 encoded_image_._encodedWidth = raw_->d_w;
378 encoded_complete_callback_->Encoded(encoded_image_, &codec_specific, 593 encoded_complete_callback_->Encoded(encoded_image_, &codec_specific,
379 &frag_info); 594 &frag_info);
380 } 595 }
381 return WEBRTC_VIDEO_CODEC_OK; 596 return WEBRTC_VIDEO_CODEC_OK;
382 } 597 }
383 598
384 int VP9EncoderImpl::SetChannelParameters(uint32_t packet_loss, int64_t rtt) { 599 int VP9EncoderImpl::SetChannelParameters(uint32_t packet_loss, int64_t rtt) {
385 return WEBRTC_VIDEO_CODEC_OK; 600 return WEBRTC_VIDEO_CODEC_OK;
386 } 601 }
387 602
388 int VP9EncoderImpl::RegisterEncodeCompleteCallback( 603 int VP9EncoderImpl::RegisterEncodeCompleteCallback(
389 EncodedImageCallback* callback) { 604 EncodedImageCallback* callback) {
(...skipping 169 matching lines...) Expand 10 before | Expand all | Expand 10 after
559 decoder_ = NULL; 774 decoder_ = NULL;
560 } 775 }
561 // Releases buffers from the pool. Any buffers not in use are deleted. Buffers 776 // Releases buffers from the pool. Any buffers not in use are deleted. Buffers
562 // still referenced externally are deleted once fully released, not returning 777 // still referenced externally are deleted once fully released, not returning
563 // to the pool. 778 // to the pool.
564 frame_buffer_pool_.ClearPool(); 779 frame_buffer_pool_.ClearPool();
565 inited_ = false; 780 inited_ = false;
566 return WEBRTC_VIDEO_CODEC_OK; 781 return WEBRTC_VIDEO_CODEC_OK;
567 } 782 }
568 } // namespace webrtc 783 } // namespace webrtc
OLDNEW
« no previous file with comments | « webrtc/modules/video_coding/codecs/vp9/vp9_impl.h ('k') | webrtc/modules/video_coding/main/source/codec_database.cc » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698