Index: talk/media/base/videocapturer.cc |
diff --git a/talk/media/base/videocapturer.cc b/talk/media/base/videocapturer.cc |
index 33cab7189398f71c9b3361c82fbe883d9606bf1a..ca4b9069f11a02f594991e6e2c93d63f39fece99 100644 |
--- a/talk/media/base/videocapturer.cc |
+++ b/talk/media/base/videocapturer.cc |
@@ -58,7 +58,7 @@ enum { |
MSG_STATE_CHANGE |
}; |
-static const int64 kMaxDistance = ~(static_cast<int64>(1) << 63); |
+static const int64_t kMaxDistance = ~(static_cast<int64_t>(1) << 63); |
#ifdef LINUX |
static const int kYU12Penalty = 16; // Needs to be higher than MJPG index. |
#endif |
@@ -86,7 +86,7 @@ CapturedFrame::CapturedFrame() |
data(NULL) {} |
// TODO(fbarchard): Remove this function once lmimediaengine stops using it. |
-bool CapturedFrame::GetDataSize(uint32* size) const { |
+bool CapturedFrame::GetDataSize(uint32_t* size) const { |
if (!size || data_size == CapturedFrame::kUnknownDataSize) { |
return false; |
} |
@@ -275,11 +275,11 @@ bool VideoCapturer::GetBestCaptureFormat(const VideoFormat& format, |
return false; |
} |
LOG(LS_INFO) << " Capture Requested " << format.ToString(); |
- int64 best_distance = kMaxDistance; |
+ int64_t best_distance = kMaxDistance; |
std::vector<VideoFormat>::const_iterator best = supported_formats->end(); |
std::vector<VideoFormat>::const_iterator i; |
for (i = supported_formats->begin(); i != supported_formats->end(); ++i) { |
- int64 distance = GetFormatDistance(format, *i); |
+ int64_t distance = GetFormatDistance(format, *i); |
// TODO(fbarchard): Reduce to LS_VERBOSE if/when camera capture is |
// relatively bug free. |
LOG(LS_INFO) << " Supported " << i->ToString() << " distance " << distance; |
@@ -361,7 +361,7 @@ void VideoCapturer::OnFrameCaptured(VideoCapturer*, |
} |
// Use a temporary buffer to scale |
- rtc::scoped_ptr<uint8[]> scale_buffer; |
+ rtc::scoped_ptr<uint8_t[]> scale_buffer; |
if (IsScreencast()) { |
int scaled_width, scaled_height; |
@@ -390,16 +390,15 @@ void VideoCapturer::OnFrameCaptured(VideoCapturer*, |
CapturedFrame* modified_frame = |
const_cast<CapturedFrame*>(captured_frame); |
const int modified_frame_size = scaled_width * scaled_height * 4; |
- scale_buffer.reset(new uint8[modified_frame_size]); |
+ scale_buffer.reset(new uint8_t[modified_frame_size]); |
// Compute new width such that width * height is less than maximum but |
// maintains original captured frame aspect ratio. |
// Round down width to multiple of 4 so odd width won't round up beyond |
// maximum, and so chroma channel is even width to simplify spatial |
// resampling. |
- libyuv::ARGBScale(reinterpret_cast<const uint8*>(captured_frame->data), |
+ libyuv::ARGBScale(reinterpret_cast<const uint8_t*>(captured_frame->data), |
captured_frame->width * 4, captured_frame->width, |
- captured_frame->height, |
- scale_buffer.get(), |
+ captured_frame->height, scale_buffer.get(), |
scaled_width * 4, scaled_width, scaled_height, |
libyuv::kFilterBilinear); |
modified_frame->width = scaled_width; |
@@ -416,7 +415,7 @@ void VideoCapturer::OnFrameCaptured(VideoCapturer*, |
// TODO(fbarchard): Avoid scale and convert if muted. |
// Temporary buffer is scoped here so it will persist until i420_frame.Init() |
// makes a copy of the frame, converting to I420. |
- rtc::scoped_ptr<uint8[]> temp_buffer; |
+ rtc::scoped_ptr<uint8_t[]> temp_buffer; |
// YUY2 can be scaled vertically using an ARGB scaler. Aspect ratio is only |
// a problem on OSX. OSX always converts webcams to YUY2 or UYVY. |
bool can_scale = |
@@ -450,26 +449,26 @@ void VideoCapturer::OnFrameCaptured(VideoCapturer*, |
scaled_height_ = scaled_height; |
} |
const int modified_frame_size = scaled_width * scaled_height * kYuy2Bpp; |
- uint8* temp_buffer_data; |
+ uint8_t* temp_buffer_data; |
// Pixels are wide and short; Increasing height. Requires temporary buffer. |
if (scaled_height > captured_frame->height) { |
- temp_buffer.reset(new uint8[modified_frame_size]); |
+ temp_buffer.reset(new uint8_t[modified_frame_size]); |
temp_buffer_data = temp_buffer.get(); |
} else { |
// Pixels are narrow and tall; Decreasing height. Scale will be done |
// in place. |
- temp_buffer_data = reinterpret_cast<uint8*>(captured_frame->data); |
+ temp_buffer_data = reinterpret_cast<uint8_t*>(captured_frame->data); |
} |
// Use ARGBScaler to vertically scale the YUY2 image, adjusting for 16 bpp. |
- libyuv::ARGBScale(reinterpret_cast<const uint8*>(captured_frame->data), |
+ libyuv::ARGBScale(reinterpret_cast<const uint8_t*>(captured_frame->data), |
captured_frame->width * kYuy2Bpp, // Stride for YUY2. |
captured_frame->width * kYuy2Bpp / kArgbBpp, // Width. |
- abs(captured_frame->height), // Height. |
+ abs(captured_frame->height), // Height. |
temp_buffer_data, |
- scaled_width * kYuy2Bpp, // Stride for YUY2. |
+ scaled_width * kYuy2Bpp, // Stride for YUY2. |
scaled_width * kYuy2Bpp / kArgbBpp, // Width. |
- abs(scaled_height), // New height. |
+ abs(scaled_height), // New height. |
libyuv::kFilterBilinear); |
modified_frame->width = scaled_width; |
modified_frame->height = scaled_height; |
@@ -589,16 +588,16 @@ void VideoCapturer::OnMessage(rtc::Message* message) { |
// 3) Framerate closeness. If not same, we prefer faster. |
// 4) Compression. If desired format has a specific fourcc, we need exact match; |
// otherwise, we use preference. |
-int64 VideoCapturer::GetFormatDistance(const VideoFormat& desired, |
- const VideoFormat& supported) { |
- int64 distance = kMaxDistance; |
+int64_t VideoCapturer::GetFormatDistance(const VideoFormat& desired, |
+ const VideoFormat& supported) { |
+ int64_t distance = kMaxDistance; |
// Check fourcc. |
- uint32 supported_fourcc = CanonicalFourCC(supported.fourcc); |
- int64 delta_fourcc = kMaxDistance; |
+ uint32_t supported_fourcc = CanonicalFourCC(supported.fourcc); |
+ int64_t delta_fourcc = kMaxDistance; |
if (FOURCC_ANY == desired.fourcc) { |
// Any fourcc is OK for the desired. Use preference to find best fourcc. |
- std::vector<uint32> preferred_fourccs; |
+ std::vector<uint32_t> preferred_fourccs; |
if (!GetPreferredFourccs(&preferred_fourccs)) { |
return distance; |
} |
@@ -629,15 +628,15 @@ int64 VideoCapturer::GetFormatDistance(const VideoFormat& desired, |
// Check resolution and fps. |
int desired_width = desired.width; |
int desired_height = desired.height; |
- int64 delta_w = supported.width - desired_width; |
+ int64_t delta_w = supported.width - desired_width; |
float supported_fps = VideoFormat::IntervalToFpsFloat(supported.interval); |
float delta_fps = |
supported_fps - VideoFormat::IntervalToFpsFloat(desired.interval); |
// Check height of supported height compared to height we would like it to be. |
- int64 aspect_h = |
- desired_width ? supported.width * desired_height / desired_width |
- : desired_height; |
- int64 delta_h = supported.height - aspect_h; |
+ int64_t aspect_h = desired_width |
+ ? supported.width * desired_height / desired_width |
+ : desired_height; |
+ int64_t delta_h = supported.height - aspect_h; |
distance = 0; |
// Set high penalty if the supported format is lower than the desired format. |
@@ -664,12 +663,12 @@ int64 VideoCapturer::GetFormatDistance(const VideoFormat& desired, |
VideoFormat::IntervalToFpsFloat(desired.interval) * 23.f / 30.f; |
delta_fps = -delta_fps; |
if (supported_fps < min_desirable_fps) { |
- distance |= static_cast<int64>(1) << 62; |
+ distance |= static_cast<int64_t>(1) << 62; |
} else { |
- distance |= static_cast<int64>(1) << 15; |
+ distance |= static_cast<int64_t>(1) << 15; |
} |
} |
- int64 idelta_fps = static_cast<int>(delta_fps); |
+ int64_t idelta_fps = static_cast<int>(delta_fps); |
// 12 bits for width and height and 8 bits for fps and fourcc. |
distance |= |