Index: remoting/codec/video_encoder_vpx.cc |
diff --git a/remoting/codec/video_encoder_vpx.cc b/remoting/codec/video_encoder_vpx.cc |
index 0dcd456ac32980e86227ede4a2e90942ccfa2dc9..bea9a8a71e4be729cc709b1d469c02a55a5bcf23 100644 |
--- a/remoting/codec/video_encoder_vpx.cc |
+++ b/remoting/codec/video_encoder_vpx.cc |
@@ -149,7 +149,7 @@ void SetVp9CodecOptions(vpx_codec_ctx_t* codec, bool lossless_encode) { |
void FreeImageIfMismatched(bool use_i444, |
const webrtc::DesktopSize& size, |
scoped_ptr<vpx_image_t>* out_image, |
- scoped_ptr<uint8[]>* out_image_buffer) { |
+ scoped_ptr<uint8_t[]>* out_image_buffer) { |
if (*out_image) { |
const vpx_img_fmt_t desired_fmt = |
use_i444 ? VPX_IMG_FMT_I444 : VPX_IMG_FMT_I420; |
@@ -164,7 +164,7 @@ void FreeImageIfMismatched(bool use_i444, |
void CreateImage(bool use_i444, |
const webrtc::DesktopSize& size, |
scoped_ptr<vpx_image_t>* out_image, |
- scoped_ptr<uint8[]>* out_image_buffer) { |
+ scoped_ptr<uint8_t[]>* out_image_buffer) { |
DCHECK(!size.is_empty()); |
DCHECK(!*out_image_buffer); |
DCHECK(!*out_image); |
@@ -207,7 +207,7 @@ void CreateImage(bool use_i444, |
// Allocate a YUV buffer large enough for the aligned data & padding. |
const int buffer_size = y_stride * y_rows + 2*uv_stride * uv_rows; |
- scoped_ptr<uint8[]> image_buffer(new uint8[buffer_size]); |
+ scoped_ptr<uint8_t[]> image_buffer(new uint8_t[buffer_size]); |
// Reset image value to 128 so we just need to fill in the y plane. |
memset(image_buffer.get(), 128, buffer_size); |
@@ -360,7 +360,7 @@ void VideoEncoderVpx::Configure(const webrtc::DesktopSize& size) { |
(size.width() + kMacroBlockSize - 1) / kMacroBlockSize, |
(size.height() + kMacroBlockSize - 1) / kMacroBlockSize); |
active_map_.reset( |
- new uint8[active_map_size_.width() * active_map_size_.height()]); |
+ new uint8_t[active_map_size_.width() * active_map_size_.height()]); |
// TODO(wez): Remove this hack once VPX can handle frame size reconfiguration. |
// See https://code.google.com/p/webm/issues/detail?id=912. |
@@ -447,14 +447,14 @@ void VideoEncoderVpx::PrepareImage(const webrtc::DesktopFrame& frame, |
} |
// Convert the updated region to YUV ready for encoding. |
- const uint8* rgb_data = frame.data(); |
+ const uint8_t* rgb_data = frame.data(); |
const int rgb_stride = frame.stride(); |
const int y_stride = image_->stride[0]; |
DCHECK_EQ(image_->stride[1], image_->stride[2]); |
const int uv_stride = image_->stride[1]; |
- uint8* y_data = image_->planes[0]; |
- uint8* u_data = image_->planes[1]; |
- uint8* v_data = image_->planes[2]; |
+ uint8_t* y_data = image_->planes[0]; |
+ uint8_t* u_data = image_->planes[1]; |
+ uint8_t* v_data = image_->planes[2]; |
switch (image_->fmt) { |
case VPX_IMG_FMT_I444: |
@@ -509,7 +509,7 @@ void VideoEncoderVpx::SetActiveMapFromRegion( |
DCHECK_LT(right, active_map_size_.width()); |
DCHECK_LT(bottom, active_map_size_.height()); |
- uint8* map = active_map_.get() + top * active_map_size_.width(); |
+ uint8_t* map = active_map_.get() + top * active_map_size_.width(); |
for (int y = top; y <= bottom; ++y) { |
for (int x = left; x <= right; ++x) |
map[x] = 1; |
@@ -520,7 +520,7 @@ void VideoEncoderVpx::SetActiveMapFromRegion( |
void VideoEncoderVpx::UpdateRegionFromActiveMap( |
webrtc::DesktopRegion* updated_region) { |
- const uint8* map = active_map_.get(); |
+ const uint8_t* map = active_map_.get(); |
for (int y = 0; y < active_map_size_.height(); ++y) { |
for (int x0 = 0; x0 < active_map_size_.width();) { |
int x1 = x0; |