OLD | NEW |
---|---|
1 /* | 1 /* |
2 * Copyright (c) 2014 The WebRTC project authors. All Rights Reserved. | 2 * Copyright (c) 2014 The WebRTC project authors. All Rights Reserved. |
3 * | 3 * |
4 * Use of this source code is governed by a BSD-style license | 4 * Use of this source code is governed by a BSD-style license |
5 * that can be found in the LICENSE file in the root of the source | 5 * that can be found in the LICENSE file in the root of the source |
6 * tree. An additional intellectual property rights grant can be found | 6 * tree. An additional intellectual property rights grant can be found |
7 * in the file PATENTS. All contributing project authors may | 7 * in the file PATENTS. All contributing project authors may |
8 * be found in the AUTHORS file in the root of the source tree. | 8 * be found in the AUTHORS file in the root of the source tree. |
9 * | 9 * |
10 */ | 10 */ |
(...skipping 39 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
50 if (width * height <= 352 * 288) | 50 if (width * height <= 352 * 288) |
51 return 5; | 51 return 5; |
52 else | 52 else |
53 return 7; | 53 return 7; |
54 } | 54 } |
55 | 55 |
56 VP9Encoder* VP9Encoder::Create() { | 56 VP9Encoder* VP9Encoder::Create() { |
57 return new VP9EncoderImpl(); | 57 return new VP9EncoderImpl(); |
58 } | 58 } |
59 | 59 |
60 void VP9EncoderImpl::EncoderOutputCodedPacketCallback(vpx_codec_cx_pkt* pkt, | |
61 void* user_data) { | |
62 VP9EncoderImpl* enc = (VP9EncoderImpl*)(user_data); | |
63 enc->GetEncodedLayerFrame(pkt); | |
64 } | |
65 | |
60 VP9EncoderImpl::VP9EncoderImpl() | 66 VP9EncoderImpl::VP9EncoderImpl() |
61 : encoded_image_(), | 67 : encoded_image_(), |
62 encoded_complete_callback_(NULL), | 68 encoded_complete_callback_(NULL), |
63 inited_(false), | 69 inited_(false), |
64 timestamp_(0), | 70 timestamp_(0), |
65 picture_id_(0), | 71 picture_id_(0), |
66 cpu_speed_(3), | 72 cpu_speed_(3), |
67 rc_max_intra_target_(0), | 73 rc_max_intra_target_(0), |
68 encoder_(NULL), | 74 encoder_(NULL), |
69 config_(NULL), | 75 config_(NULL), |
70 raw_(NULL) { | 76 raw_(NULL), |
77 input_image_(NULL), | |
78 tl0_pic_idx_(0), | |
79 gof_idx_(0), | |
80 num_temporal_layers_(0), | |
81 num_spatial_layers_(0) { | |
71 memset(&codec_, 0, sizeof(codec_)); | 82 memset(&codec_, 0, sizeof(codec_)); |
72 uint32_t seed = static_cast<uint32_t>(TickTime::MillisecondTimestamp()); | 83 uint32_t seed = static_cast<uint32_t>(TickTime::MillisecondTimestamp()); |
73 srand(seed); | 84 srand(seed); |
74 } | 85 } |
75 | 86 |
76 VP9EncoderImpl::~VP9EncoderImpl() { | 87 VP9EncoderImpl::~VP9EncoderImpl() { |
77 Release(); | 88 Release(); |
78 } | 89 } |
79 | 90 |
80 int VP9EncoderImpl::Release() { | 91 int VP9EncoderImpl::Release() { |
(...skipping 13 matching lines...) Expand all Loading... | |
94 config_ = NULL; | 105 config_ = NULL; |
95 } | 106 } |
96 if (raw_ != NULL) { | 107 if (raw_ != NULL) { |
97 vpx_img_free(raw_); | 108 vpx_img_free(raw_); |
98 raw_ = NULL; | 109 raw_ = NULL; |
99 } | 110 } |
100 inited_ = false; | 111 inited_ = false; |
101 return WEBRTC_VIDEO_CODEC_OK; | 112 return WEBRTC_VIDEO_CODEC_OK; |
102 } | 113 } |
103 | 114 |
115 bool VP9EncoderImpl::SetSvcRates() { | |
116 float rate_ratio[VPX_MAX_LAYERS] = {0}; | |
117 float total = 0; | |
118 uint8_t i = 0; | |
119 | |
120 for (i = 0; i < num_spatial_layers_; ++i) { | |
121 if (svc_internal_.svc_params.scaling_factor_num[i] <= 0 || | |
122 svc_internal_.svc_params.scaling_factor_den[i] <= 0) { | |
123 return false; | |
124 } | |
125 rate_ratio[i] = static_cast<float>( | |
126 svc_internal_.svc_params.scaling_factor_num[i]) / | |
127 svc_internal_.svc_params.scaling_factor_den[i]; | |
128 total += rate_ratio[i]; | |
129 } | |
130 | |
131 for (i = 0; i < num_spatial_layers_; ++i) { | |
132 config_->ss_target_bitrate[i] = static_cast<unsigned int>( | |
133 config_->rc_target_bitrate * rate_ratio[i] / total); | |
134 if (num_temporal_layers_ == 1) { | |
135 config_->layer_target_bitrate[i] = config_->ss_target_bitrate[i]; | |
136 } else if (num_temporal_layers_ == 2) { | |
137 config_->layer_target_bitrate[i * num_temporal_layers_] = | |
138 config_->ss_target_bitrate[i] * 2 / 3; | |
139 config_->layer_target_bitrate[i * num_temporal_layers_ + 1] = | |
140 config_->ss_target_bitrate[i]; | |
141 } else if (num_temporal_layers_ == 3) { | |
142 config_->layer_target_bitrate[i * num_temporal_layers_] = | |
143 config_->ss_target_bitrate[i] / 2; | |
144 config_->layer_target_bitrate[i * num_temporal_layers_ + 1] = | |
145 config_->layer_target_bitrate[i * num_temporal_layers_] + | |
146 (config_->ss_target_bitrate[i] / 4); | |
147 config_->layer_target_bitrate[i * num_temporal_layers_ + 2] = | |
148 config_->ss_target_bitrate[i]; | |
149 } else { | |
150 return false; | |
151 } | |
152 } | |
153 | |
154 // For now, temporal layers only supported when having one spatial layer. | |
155 if (num_spatial_layers_ == 1) { | |
156 for (i = 0; i < num_temporal_layers_; ++i) { | |
157 config_->ts_target_bitrate[i] = config_->layer_target_bitrate[i]; | |
158 } | |
159 } | |
160 | |
161 return true; | |
162 } | |
163 | |
104 int VP9EncoderImpl::SetRates(uint32_t new_bitrate_kbit, | 164 int VP9EncoderImpl::SetRates(uint32_t new_bitrate_kbit, |
105 uint32_t new_framerate) { | 165 uint32_t new_framerate) { |
106 if (!inited_) { | 166 if (!inited_) { |
107 return WEBRTC_VIDEO_CODEC_UNINITIALIZED; | 167 return WEBRTC_VIDEO_CODEC_UNINITIALIZED; |
108 } | 168 } |
109 if (encoder_->err) { | 169 if (encoder_->err) { |
110 return WEBRTC_VIDEO_CODEC_ERROR; | 170 return WEBRTC_VIDEO_CODEC_ERROR; |
111 } | 171 } |
112 if (new_framerate < 1) { | 172 if (new_framerate < 1) { |
113 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; | 173 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; |
114 } | 174 } |
115 // Update bit rate | 175 // Update bit rate |
116 if (codec_.maxBitrate > 0 && new_bitrate_kbit > codec_.maxBitrate) { | 176 if (codec_.maxBitrate > 0 && new_bitrate_kbit > codec_.maxBitrate) { |
117 new_bitrate_kbit = codec_.maxBitrate; | 177 new_bitrate_kbit = codec_.maxBitrate; |
118 } | 178 } |
119 config_->rc_target_bitrate = new_bitrate_kbit; | 179 config_->rc_target_bitrate = new_bitrate_kbit; |
120 codec_.maxFramerate = new_framerate; | 180 codec_.maxFramerate = new_framerate; |
181 | |
182 if (!SetSvcRates()) { | |
183 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; | |
184 } | |
185 | |
121 // Update encoder context | 186 // Update encoder context |
122 if (vpx_codec_enc_config_set(encoder_, config_)) { | 187 if (vpx_codec_enc_config_set(encoder_, config_)) { |
123 return WEBRTC_VIDEO_CODEC_ERROR; | 188 return WEBRTC_VIDEO_CODEC_ERROR; |
124 } | 189 } |
125 return WEBRTC_VIDEO_CODEC_OK; | 190 return WEBRTC_VIDEO_CODEC_OK; |
126 } | 191 } |
127 | 192 |
128 int VP9EncoderImpl::InitEncode(const VideoCodec* inst, | 193 int VP9EncoderImpl::InitEncode(const VideoCodec* inst, |
129 int number_of_cores, | 194 int number_of_cores, |
130 size_t /*max_payload_size*/) { | 195 size_t /*max_payload_size*/) { |
131 if (inst == NULL) { | 196 if (inst == NULL) { |
132 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; | 197 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; |
133 } | 198 } |
134 if (inst->maxFramerate < 1) { | 199 if (inst->maxFramerate < 1) { |
135 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; | 200 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; |
136 } | 201 } |
137 // Allow zero to represent an unspecified maxBitRate | 202 // Allow zero to represent an unspecified maxBitRate |
138 if (inst->maxBitrate > 0 && inst->startBitrate > inst->maxBitrate) { | 203 if (inst->maxBitrate > 0 && inst->startBitrate > inst->maxBitrate) { |
139 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; | 204 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; |
140 } | 205 } |
141 if (inst->width < 1 || inst->height < 1) { | 206 if (inst->width < 1 || inst->height < 1) { |
142 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; | 207 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; |
143 } | 208 } |
144 if (number_of_cores < 1) { | 209 if (number_of_cores < 1) { |
145 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; | 210 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; |
146 } | 211 } |
212 if (inst->codecSpecific.VP9.numberOfTemporalLayers < 1 || | |
holmer
2015/07/31 09:05:26
Should 0 be allowed as "no temporal layers"?
åsapersson
2015/07/31 10:09:37
See change.
| |
213 inst->codecSpecific.VP9.numberOfTemporalLayers > 3) { | |
214 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; | |
215 } | |
216 // For now, only support one spatial layer. | |
217 if (inst->codecSpecific.VP9.numberOfSpatialLayers != 1) { | |
218 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; | |
219 } | |
147 int retVal = Release(); | 220 int retVal = Release(); |
148 if (retVal < 0) { | 221 if (retVal < 0) { |
149 return retVal; | 222 return retVal; |
150 } | 223 } |
151 if (encoder_ == NULL) { | 224 if (encoder_ == NULL) { |
152 encoder_ = new vpx_codec_ctx_t; | 225 encoder_ = new vpx_codec_ctx_t; |
153 } | 226 } |
154 if (config_ == NULL) { | 227 if (config_ == NULL) { |
155 config_ = new vpx_codec_enc_cfg_t; | 228 config_ = new vpx_codec_enc_cfg_t; |
156 } | 229 } |
157 timestamp_ = 0; | 230 timestamp_ = 0; |
158 if (&codec_ != inst) { | 231 if (&codec_ != inst) { |
159 codec_ = *inst; | 232 codec_ = *inst; |
160 } | 233 } |
234 | |
235 num_temporal_layers_ = inst->codecSpecific.VP9.numberOfTemporalLayers; | |
236 num_spatial_layers_ = inst->codecSpecific.VP9.numberOfSpatialLayers; | |
237 | |
161 // Random start 16 bits is enough. | 238 // Random start 16 bits is enough. |
162 picture_id_ = static_cast<uint16_t>(rand()) & 0x7FFF; | 239 picture_id_ = static_cast<uint16_t>(rand()) & 0x7FFF; |
163 // Allocate memory for encoded image | 240 // Allocate memory for encoded image |
164 if (encoded_image_._buffer != NULL) { | 241 if (encoded_image_._buffer != NULL) { |
165 delete [] encoded_image_._buffer; | 242 delete [] encoded_image_._buffer; |
166 } | 243 } |
167 encoded_image_._size = CalcBufferSize(kI420, codec_.width, codec_.height); | 244 encoded_image_._size = CalcBufferSize(kI420, codec_.width, codec_.height); |
168 encoded_image_._buffer = new uint8_t[encoded_image_._size]; | 245 encoded_image_._buffer = new uint8_t[encoded_image_._size]; |
169 encoded_image_._completeFrame = true; | 246 encoded_image_._completeFrame = true; |
170 // Creating a wrapper to the image - setting image data to NULL. Actual | 247 // Creating a wrapper to the image - setting image data to NULL. Actual |
(...skipping 31 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
202 if (inst->codecSpecific.VP9.keyFrameInterval > 0) { | 279 if (inst->codecSpecific.VP9.keyFrameInterval > 0) { |
203 config_->kf_mode = VPX_KF_AUTO; | 280 config_->kf_mode = VPX_KF_AUTO; |
204 config_->kf_max_dist = inst->codecSpecific.VP9.keyFrameInterval; | 281 config_->kf_max_dist = inst->codecSpecific.VP9.keyFrameInterval; |
205 } else { | 282 } else { |
206 config_->kf_mode = VPX_KF_DISABLED; | 283 config_->kf_mode = VPX_KF_DISABLED; |
207 } | 284 } |
208 // Determine number of threads based on the image size and #cores. | 285 // Determine number of threads based on the image size and #cores. |
209 config_->g_threads = NumberOfThreads(config_->g_w, | 286 config_->g_threads = NumberOfThreads(config_->g_w, |
210 config_->g_h, | 287 config_->g_h, |
211 number_of_cores); | 288 number_of_cores); |
289 | |
212 cpu_speed_ = GetCpuSpeed(config_->g_w, config_->g_h); | 290 cpu_speed_ = GetCpuSpeed(config_->g_w, config_->g_h); |
291 | |
292 // TODO(asapersson): Check configuration of temporal switch up and increase | |
293 // pattern length. | |
294 if (num_temporal_layers_ == 1) { | |
295 gof_.SetGofInfoVP9(kTemporalStructureMode1); | |
296 config_->temporal_layering_mode = VP9E_TEMPORAL_LAYERING_MODE_NOLAYERING; | |
297 config_->ts_number_layers = 1; | |
298 config_->ts_rate_decimator[0] = 1; | |
299 config_->ts_periodicity = 1; | |
300 config_->ts_layer_id[0] = 0; | |
301 } else if (num_temporal_layers_ == 2) { | |
302 gof_.SetGofInfoVP9(kTemporalStructureMode2); | |
303 config_->temporal_layering_mode = VP9E_TEMPORAL_LAYERING_MODE_0101; | |
304 config_->ts_number_layers = 2; | |
305 config_->ts_rate_decimator[0] = 2; | |
306 config_->ts_rate_decimator[1] = 1; | |
307 config_->ts_periodicity = 2; | |
308 config_->ts_layer_id[0] = 0; | |
309 config_->ts_layer_id[1] = 1; | |
310 } else if (num_temporal_layers_ == 3) { | |
311 gof_.SetGofInfoVP9(kTemporalStructureMode3); | |
312 config_->temporal_layering_mode = VP9E_TEMPORAL_LAYERING_MODE_0212; | |
313 config_->ts_number_layers = 3; | |
314 config_->ts_rate_decimator[0] = 4; | |
315 config_->ts_rate_decimator[1] = 2; | |
316 config_->ts_rate_decimator[2] = 1; | |
317 config_->ts_periodicity = 4; | |
318 config_->ts_layer_id[0] = 0; | |
319 config_->ts_layer_id[1] = 2; | |
320 config_->ts_layer_id[2] = 1; | |
321 config_->ts_layer_id[3] = 2; | |
322 } else { | |
323 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; | |
324 } | |
325 | |
326 tl0_pic_idx_ = static_cast<uint8_t>(rand()); | |
327 | |
213 return InitAndSetControlSettings(inst); | 328 return InitAndSetControlSettings(inst); |
214 } | 329 } |
215 | 330 |
216 int VP9EncoderImpl::NumberOfThreads(int width, | 331 int VP9EncoderImpl::NumberOfThreads(int width, |
217 int height, | 332 int height, |
218 int number_of_cores) { | 333 int number_of_cores) { |
334 // For the current libvpx library, only 1 thread is supported when SVC is | |
335 // turned on. | |
336 if (num_temporal_layers_ > 1 || num_spatial_layers_ > 1) { | |
337 return 1; | |
338 } | |
339 | |
219 // Keep the number of encoder threads equal to the possible number of column | 340 // Keep the number of encoder threads equal to the possible number of column |
220 // tiles, which is (1, 2, 4, 8). See comments below for VP9E_SET_TILE_COLUMNS. | 341 // tiles, which is (1, 2, 4, 8). See comments below for VP9E_SET_TILE_COLUMNS. |
221 if (width * height >= 1280 * 720 && number_of_cores > 4) { | 342 if (width * height >= 1280 * 720 && number_of_cores > 4) { |
222 return 4; | 343 return 4; |
223 } else if (width * height >= 640 * 480 && number_of_cores > 2) { | 344 } else if (width * height >= 640 * 480 && number_of_cores > 2) { |
224 return 2; | 345 return 2; |
225 } else { | 346 } else { |
226 // 1 thread less than VGA. | 347 // 1 thread less than VGA. |
227 return 1; | 348 return 1; |
228 } | 349 } |
229 } | 350 } |
230 | 351 |
231 int VP9EncoderImpl::InitAndSetControlSettings(const VideoCodec* inst) { | 352 int VP9EncoderImpl::InitAndSetControlSettings(const VideoCodec* inst) { |
353 | |
354 config_->ss_number_layers = num_spatial_layers_; | |
355 | |
356 if (num_spatial_layers_ > 1) { | |
357 config_->rc_min_quantizer = 0; | |
358 config_->rc_max_quantizer = 63; | |
359 } | |
360 int scaling_factor_num = 256; | |
361 for (int i = num_spatial_layers_ - 1; i >= 0; --i) { | |
362 svc_internal_.svc_params.max_quantizers[i] = config_->rc_max_quantizer; | |
363 svc_internal_.svc_params.min_quantizers[i] = config_->rc_min_quantizer; | |
364 // 1:2 scaling in each dimension. | |
365 svc_internal_.svc_params.scaling_factor_num[i] = scaling_factor_num; | |
366 svc_internal_.svc_params.scaling_factor_den[i] = 256; | |
367 scaling_factor_num /= 2; | |
368 } | |
369 | |
370 if (!SetSvcRates()) { | |
371 return WEBRTC_VIDEO_CODEC_ERR_PARAMETER; | |
372 } | |
373 | |
232 if (vpx_codec_enc_init(encoder_, vpx_codec_vp9_cx(), config_, 0)) { | 374 if (vpx_codec_enc_init(encoder_, vpx_codec_vp9_cx(), config_, 0)) { |
233 return WEBRTC_VIDEO_CODEC_UNINITIALIZED; | 375 return WEBRTC_VIDEO_CODEC_UNINITIALIZED; |
234 } | 376 } |
235 vpx_codec_control(encoder_, VP8E_SET_CPUUSED, cpu_speed_); | 377 vpx_codec_control(encoder_, VP8E_SET_CPUUSED, cpu_speed_); |
236 vpx_codec_control(encoder_, VP8E_SET_MAX_INTRA_BITRATE_PCT, | 378 vpx_codec_control(encoder_, VP8E_SET_MAX_INTRA_BITRATE_PCT, |
237 rc_max_intra_target_); | 379 rc_max_intra_target_); |
238 vpx_codec_control(encoder_, VP9E_SET_AQ_MODE, | 380 vpx_codec_control(encoder_, VP9E_SET_AQ_MODE, |
239 inst->codecSpecific.VP9.adaptiveQpMode ? 3 : 0); | 381 inst->codecSpecific.VP9.adaptiveQpMode ? 3 : 0); |
382 | |
383 vpx_codec_control( | |
384 encoder_, VP9E_SET_SVC, | |
385 (num_temporal_layers_ > 1 || num_spatial_layers_ > 1) ? 1 : 0); | |
386 if (num_temporal_layers_ > 1 || num_spatial_layers_ > 1) { | |
387 vpx_codec_control(encoder_, VP9E_SET_SVC_PARAMETERS, | |
388 &svc_internal_.svc_params); | |
389 } | |
390 // Register callback for getting each spatial layer. | |
391 vpx_codec_priv_output_cx_pkt_cb_pair_t cbp = { | |
392 VP9EncoderImpl::EncoderOutputCodedPacketCallback, (void*)(this)}; | |
393 vpx_codec_control(encoder_, VP9E_REGISTER_CX_CALLBACK, (void*)(&cbp)); | |
394 | |
240 // Control function to set the number of column tiles in encoding a frame, in | 395 // Control function to set the number of column tiles in encoding a frame, in |
241 // log2 unit: e.g., 0 = 1 tile column, 1 = 2 tile columns, 2 = 4 tile columns. | 396 // log2 unit: e.g., 0 = 1 tile column, 1 = 2 tile columns, 2 = 4 tile columns. |
242 // The number tile columns will be capped by the encoder based on image size | 397 // The number tile columns will be capped by the encoder based on image size |
243 // (minimum width of tile column is 256 pixels, maximum is 4096). | 398 // (minimum width of tile column is 256 pixels, maximum is 4096). |
244 vpx_codec_control(encoder_, VP9E_SET_TILE_COLUMNS, (config_->g_threads >> 1)); | 399 vpx_codec_control(encoder_, VP9E_SET_TILE_COLUMNS, (config_->g_threads >> 1)); |
245 #if !defined(WEBRTC_ARCH_ARM) && !defined(WEBRTC_ARCH_ARM64) | 400 #if !defined(WEBRTC_ARCH_ARM) && !defined(WEBRTC_ARCH_ARM64) |
246 // Note denoiser is still off by default until further testing/optimization, | 401 // Note denoiser is still off by default until further testing/optimization, |
247 // i.e., codecSpecific.VP9.denoisingOn == 0. | 402 // i.e., codecSpecific.VP9.denoisingOn == 0. |
248 vpx_codec_control(encoder_, VP9E_SET_NOISE_SENSITIVITY, | 403 vpx_codec_control(encoder_, VP9E_SET_NOISE_SENSITIVITY, |
249 inst->codecSpecific.VP9.denoisingOn ? 1 : 0); | 404 inst->codecSpecific.VP9.denoisingOn ? 1 : 0); |
(...skipping 29 matching lines...) Expand all Loading... | |
279 if (encoded_complete_callback_ == NULL) { | 434 if (encoded_complete_callback_ == NULL) { |
280 return WEBRTC_VIDEO_CODEC_UNINITIALIZED; | 435 return WEBRTC_VIDEO_CODEC_UNINITIALIZED; |
281 } | 436 } |
282 VideoFrameType frame_type = kDeltaFrame; | 437 VideoFrameType frame_type = kDeltaFrame; |
283 // We only support one stream at the moment. | 438 // We only support one stream at the moment. |
284 if (frame_types && frame_types->size() > 0) { | 439 if (frame_types && frame_types->size() > 0) { |
285 frame_type = (*frame_types)[0]; | 440 frame_type = (*frame_types)[0]; |
286 } | 441 } |
287 DCHECK_EQ(input_image.width(), static_cast<int>(raw_->d_w)); | 442 DCHECK_EQ(input_image.width(), static_cast<int>(raw_->d_w)); |
288 DCHECK_EQ(input_image.height(), static_cast<int>(raw_->d_h)); | 443 DCHECK_EQ(input_image.height(), static_cast<int>(raw_->d_h)); |
444 | |
445 // Set input image for use in the callback. | |
446 // This was necessary since you need some information from input_image. | |
447 // You can save only the necessary information (such as timestamp) instead of | |
448 // doing this. | |
449 input_image_ = &input_image; | |
450 | |
289 // Image in vpx_image_t format. | 451 // Image in vpx_image_t format. |
290 // Input image is const. VPX's raw image is not defined as const. | 452 // Input image is const. VPX's raw image is not defined as const. |
291 raw_->planes[VPX_PLANE_Y] = const_cast<uint8_t*>(input_image.buffer(kYPlane)); | 453 raw_->planes[VPX_PLANE_Y] = const_cast<uint8_t*>(input_image.buffer(kYPlane)); |
292 raw_->planes[VPX_PLANE_U] = const_cast<uint8_t*>(input_image.buffer(kUPlane)); | 454 raw_->planes[VPX_PLANE_U] = const_cast<uint8_t*>(input_image.buffer(kUPlane)); |
293 raw_->planes[VPX_PLANE_V] = const_cast<uint8_t*>(input_image.buffer(kVPlane)); | 455 raw_->planes[VPX_PLANE_V] = const_cast<uint8_t*>(input_image.buffer(kVPlane)); |
294 raw_->stride[VPX_PLANE_Y] = input_image.stride(kYPlane); | 456 raw_->stride[VPX_PLANE_Y] = input_image.stride(kYPlane); |
295 raw_->stride[VPX_PLANE_U] = input_image.stride(kUPlane); | 457 raw_->stride[VPX_PLANE_U] = input_image.stride(kUPlane); |
296 raw_->stride[VPX_PLANE_V] = input_image.stride(kVPlane); | 458 raw_->stride[VPX_PLANE_V] = input_image.stride(kVPlane); |
297 | 459 |
298 int flags = 0; | 460 int flags = 0; |
299 bool send_keyframe = (frame_type == kKeyFrame); | 461 bool send_keyframe = (frame_type == kKeyFrame); |
300 if (send_keyframe) { | 462 if (send_keyframe) { |
301 // Key frame request from caller. | 463 // Key frame request from caller. |
302 flags = VPX_EFLAG_FORCE_KF; | 464 flags = VPX_EFLAG_FORCE_KF; |
303 } | 465 } |
304 assert(codec_.maxFramerate > 0); | 466 assert(codec_.maxFramerate > 0); |
305 uint32_t duration = 90000 / codec_.maxFramerate; | 467 uint32_t duration = 90000 / codec_.maxFramerate; |
306 if (vpx_codec_encode(encoder_, raw_, timestamp_, duration, flags, | 468 if (vpx_codec_encode(encoder_, raw_, timestamp_, duration, flags, |
307 VPX_DL_REALTIME)) { | 469 VPX_DL_REALTIME)) { |
308 return WEBRTC_VIDEO_CODEC_ERROR; | 470 return WEBRTC_VIDEO_CODEC_ERROR; |
309 } | 471 } |
310 timestamp_ += duration; | 472 timestamp_ += duration; |
311 return GetEncodedPartitions(input_image); | 473 |
474 return WEBRTC_VIDEO_CODEC_OK; | |
312 } | 475 } |
313 | 476 |
314 void VP9EncoderImpl::PopulateCodecSpecific(CodecSpecificInfo* codec_specific, | 477 void VP9EncoderImpl::PopulateCodecSpecific(CodecSpecificInfo* codec_specific, |
315 const vpx_codec_cx_pkt& pkt, | 478 const vpx_codec_cx_pkt& pkt, |
316 uint32_t timestamp) { | 479 uint32_t timestamp) { |
317 assert(codec_specific != NULL); | 480 assert(codec_specific != NULL); |
318 codec_specific->codecType = kVideoCodecVP9; | 481 codec_specific->codecType = kVideoCodecVP9; |
319 CodecSpecificInfoVP9 *vp9_info = &(codec_specific->codecSpecific.VP9); | 482 CodecSpecificInfoVP9 *vp9_info = &(codec_specific->codecSpecific.VP9); |
320 vp9_info->pictureId = picture_id_; | 483 // TODO(asapersson): Set correct values. |
321 vp9_info->keyIdx = kNoKeyIdx; | 484 vp9_info->inter_pic_predicted = |
322 vp9_info->nonReference = (pkt.data.frame.flags & VPX_FRAME_IS_DROPPABLE) != 0; | 485 (pkt.data.frame.flags & VPX_FRAME_IS_KEY) ? false : true; |
323 // TODO(marpan): Temporal layers are supported in the current VP9 version, | 486 vp9_info->flexible_mode = codec_.codecSpecific.VP9.flexibleMode; |
324 // but for now use 1 temporal layer encoding. Will update this when temporal | 487 vp9_info->ss_data_available = |
325 // layer support for VP9 is added in webrtc. | 488 (pkt.data.frame.flags & VPX_FRAME_IS_KEY) ? true : false; |
326 vp9_info->temporalIdx = kNoTemporalIdx; | 489 if (pkt.data.frame.flags & VPX_FRAME_IS_KEY) { |
327 vp9_info->layerSync = false; | 490 gof_idx_ = 0; |
328 vp9_info->tl0PicIdx = kNoTl0PicIdx; | 491 } |
329 picture_id_ = (picture_id_ + 1) & 0x7FFF; | 492 |
493 vpx_svc_layer_id_t layer_id = {0}; | |
494 vpx_codec_control(encoder_, VP9E_GET_SVC_LAYER_ID, &layer_id); | |
495 | |
496 assert(num_temporal_layers_ > 0); | |
497 assert(num_spatial_layers_ > 0); | |
498 if (num_temporal_layers_ == 1) { | |
499 assert(layer_id.temporal_layer_id == 0); | |
500 vp9_info->temporal_idx = kNoTemporalIdx; | |
501 } else { | |
502 vp9_info->temporal_idx = layer_id.temporal_layer_id; | |
503 } | |
504 if (num_spatial_layers_ == 1) { | |
505 assert(layer_id.spatial_layer_id == 0); | |
506 vp9_info->spatial_idx = kNoSpatialIdx; | |
507 } else { | |
508 vp9_info->spatial_idx = layer_id.spatial_layer_id; | |
509 } | |
510 if (layer_id.spatial_layer_id != 0) { | |
511 vp9_info->ss_data_available = false; | |
512 } | |
513 | |
514 if (vp9_info->flexible_mode) { | |
515 vp9_info->gof_idx = kNoGofIdx; | |
516 } else { | |
517 vp9_info->gof_idx = | |
518 static_cast<uint8_t>(gof_idx_++ % gof_.num_frames_in_gof); | |
519 } | |
520 | |
521 // TODO(asapersson): this info has to be obtained from the encoder. | |
522 vp9_info->temporal_up_switch = true; | |
523 | |
524 if (layer_id.spatial_layer_id == 0) { | |
525 picture_id_ = (picture_id_ + 1) & 0x7FFF; | |
526 // TODO(asapersson): this info has to be obtained from the encoder. | |
527 vp9_info->inter_layer_predicted = false; | |
528 } else { | |
529 // TODO(asapersson): this info has to be obtained from the encoder. | |
530 vp9_info->inter_layer_predicted = true; | |
531 } | |
532 | |
533 vp9_info->picture_id = picture_id_; | |
534 | |
535 if (!vp9_info->flexible_mode) { | |
536 if (layer_id.temporal_layer_id == 0 && layer_id.spatial_layer_id == 0) { | |
537 tl0_pic_idx_++; | |
538 } | |
539 vp9_info->tl0_pic_idx = tl0_pic_idx_; | |
540 } | |
541 | |
542 if (vp9_info->ss_data_available) { | |
543 vp9_info->num_spatial_layers = num_spatial_layers_; | |
544 vp9_info->spatial_layer_resolution_present = true; | |
545 for (size_t i = 0; i < vp9_info->num_spatial_layers; ++i) { | |
546 vp9_info->width[i] = codec_.width * | |
547 svc_internal_.svc_params.scaling_factor_num[i] / | |
548 svc_internal_.svc_params.scaling_factor_den[i]; | |
549 vp9_info->height[i] = codec_.height * | |
550 svc_internal_.svc_params.scaling_factor_num[i] / | |
551 svc_internal_.svc_params.scaling_factor_den[i]; | |
552 } | |
553 if (!vp9_info->flexible_mode) { | |
554 vp9_info->gof.CopyGofInfoVP9(gof_); | |
555 } | |
556 } | |
330 } | 557 } |
331 | 558 |
332 int VP9EncoderImpl::GetEncodedPartitions(const VideoFrame& input_image) { | 559 int VP9EncoderImpl::GetEncodedLayerFrame(const vpx_codec_cx_pkt* pkt) { |
333 vpx_codec_iter_t iter = NULL; | |
334 encoded_image_._length = 0; | 560 encoded_image_._length = 0; |
335 encoded_image_._frameType = kDeltaFrame; | 561 encoded_image_._frameType = kDeltaFrame; |
336 RTPFragmentationHeader frag_info; | 562 RTPFragmentationHeader frag_info; |
337 // Note: no data partitioning in VP9, so 1 partition only. We keep this | 563 // Note: no data partitioning in VP9, so 1 partition only. We keep this |
338 // fragmentation data for now, until VP9 packetizer is implemented. | 564 // fragmentation data for now, until VP9 packetizer is implemented. |
339 frag_info.VerifyAndAllocateFragmentationHeader(1); | 565 frag_info.VerifyAndAllocateFragmentationHeader(1); |
340 int part_idx = 0; | 566 int part_idx = 0; |
341 CodecSpecificInfo codec_specific; | 567 CodecSpecificInfo codec_specific; |
342 const vpx_codec_cx_pkt_t *pkt = NULL; | 568 |
343 while ((pkt = vpx_codec_get_cx_data(encoder_, &iter)) != NULL) { | 569 assert(pkt->kind == VPX_CODEC_CX_FRAME_PKT); |
344 switch (pkt->kind) { | 570 memcpy(&encoded_image_._buffer[encoded_image_._length], pkt->data.frame.buf, |
345 case VPX_CODEC_CX_FRAME_PKT: { | 571 pkt->data.frame.sz); |
346 memcpy(&encoded_image_._buffer[encoded_image_._length], | 572 frag_info.fragmentationOffset[part_idx] = encoded_image_._length; |
347 pkt->data.frame.buf, | 573 frag_info.fragmentationLength[part_idx] = |
348 pkt->data.frame.sz); | 574 static_cast<uint32_t>(pkt->data.frame.sz); |
349 frag_info.fragmentationOffset[part_idx] = encoded_image_._length; | 575 frag_info.fragmentationPlType[part_idx] = 0; |
350 frag_info.fragmentationLength[part_idx] = | 576 frag_info.fragmentationTimeDiff[part_idx] = 0; |
351 static_cast<uint32_t>(pkt->data.frame.sz); | 577 encoded_image_._length += static_cast<uint32_t>(pkt->data.frame.sz); |
352 frag_info.fragmentationPlType[part_idx] = 0; | 578 assert(encoded_image_._length <= encoded_image_._size); |
353 frag_info.fragmentationTimeDiff[part_idx] = 0; | 579 |
354 encoded_image_._length += static_cast<uint32_t>(pkt->data.frame.sz); | 580 // End of frame. |
355 assert(encoded_image_._length <= encoded_image_._size); | 581 // Check if encoded frame is a key frame. |
356 break; | 582 if (pkt->data.frame.flags & VPX_FRAME_IS_KEY) { |
357 } | 583 encoded_image_._frameType = kKeyFrame; |
358 default: { | |
359 break; | |
360 } | |
361 } | |
362 // End of frame. | |
363 if ((pkt->data.frame.flags & VPX_FRAME_IS_FRAGMENT) == 0) { | |
364 // Check if encoded frame is a key frame. | |
365 if (pkt->data.frame.flags & VPX_FRAME_IS_KEY) { | |
366 encoded_image_._frameType = kKeyFrame; | |
367 } | |
368 PopulateCodecSpecific(&codec_specific, *pkt, input_image.timestamp()); | |
369 break; | |
370 } | |
371 } | 584 } |
585 PopulateCodecSpecific(&codec_specific, *pkt, input_image_->timestamp()); | |
586 | |
372 if (encoded_image_._length > 0) { | 587 if (encoded_image_._length > 0) { |
373 TRACE_COUNTER1("webrtc", "EncodedFrameSize", encoded_image_._length); | 588 TRACE_COUNTER1("webrtc", "EncodedFrameSize", encoded_image_._length); |
374 encoded_image_._timeStamp = input_image.timestamp(); | 589 encoded_image_._timeStamp = input_image_->timestamp(); |
375 encoded_image_.capture_time_ms_ = input_image.render_time_ms(); | 590 encoded_image_.capture_time_ms_ = input_image_->render_time_ms(); |
376 encoded_image_._encodedHeight = raw_->d_h; | 591 encoded_image_._encodedHeight = raw_->d_h; |
377 encoded_image_._encodedWidth = raw_->d_w; | 592 encoded_image_._encodedWidth = raw_->d_w; |
378 encoded_complete_callback_->Encoded(encoded_image_, &codec_specific, | 593 encoded_complete_callback_->Encoded(encoded_image_, &codec_specific, |
379 &frag_info); | 594 &frag_info); |
380 } | 595 } |
381 return WEBRTC_VIDEO_CODEC_OK; | 596 return WEBRTC_VIDEO_CODEC_OK; |
382 } | 597 } |
383 | 598 |
384 int VP9EncoderImpl::SetChannelParameters(uint32_t packet_loss, int64_t rtt) { | 599 int VP9EncoderImpl::SetChannelParameters(uint32_t packet_loss, int64_t rtt) { |
385 return WEBRTC_VIDEO_CODEC_OK; | 600 return WEBRTC_VIDEO_CODEC_OK; |
386 } | 601 } |
387 | 602 |
388 int VP9EncoderImpl::RegisterEncodeCompleteCallback( | 603 int VP9EncoderImpl::RegisterEncodeCompleteCallback( |
389 EncodedImageCallback* callback) { | 604 EncodedImageCallback* callback) { |
(...skipping 169 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
559 decoder_ = NULL; | 774 decoder_ = NULL; |
560 } | 775 } |
561 // Releases buffers from the pool. Any buffers not in use are deleted. Buffers | 776 // Releases buffers from the pool. Any buffers not in use are deleted. Buffers |
562 // still referenced externally are deleted once fully released, not returning | 777 // still referenced externally are deleted once fully released, not returning |
563 // to the pool. | 778 // to the pool. |
564 frame_buffer_pool_.ClearPool(); | 779 frame_buffer_pool_.ClearPool(); |
565 inited_ = false; | 780 inited_ = false; |
566 return WEBRTC_VIDEO_CODEC_OK; | 781 return WEBRTC_VIDEO_CODEC_OK; |
567 } | 782 } |
568 } // namespace webrtc | 783 } // namespace webrtc |
OLD | NEW |