Chromium Code Reviews| Index: content/common/gpu/media/video_encode_accelerator_unittest.cc |
| diff --git a/content/common/gpu/media/video_encode_accelerator_unittest.cc b/content/common/gpu/media/video_encode_accelerator_unittest.cc |
| index 9d1e3b69dc7c9edccf56b2fab6021bc0fb20f23b..f54c3cf610682b4fa146b0c9977d7ee5623e8a75 100644 |
| --- a/content/common/gpu/media/video_encode_accelerator_unittest.cc |
| +++ b/content/common/gpu/media/video_encode_accelerator_unittest.cc |
| @@ -33,6 +33,8 @@ |
| #error The VideoEncodeAcceleratorUnittest is not supported on this platform. |
| #endif |
| +#define ALIGN_64_BYTES(x) (((x) + 63) & ~63) |
| + |
| using media::VideoEncodeAccelerator; |
| namespace content { |
| @@ -103,6 +105,7 @@ struct TestStream { |
| ~TestStream() {} |
| gfx::Size size; |
| + std::string in_filename; |
| base::MemoryMappedFile input_file; |
| media::VideoCodecProfile requested_profile; |
| std::string out_filename; |
| @@ -112,6 +115,96 @@ struct TestStream { |
| unsigned int requested_subsequent_framerate; |
| }; |
| +// ARM performs CPU cache management with CPU cache line granularity. We thus |
| +// need to ensure our buffers are CPU cache line-aligned (64 byte-aligned). |
| +// Otherwise newer kernels will refuse to accept them, and on older kernels |
| +// we'll be treating ourselves to random corruption. |
| +// Since we are just mmapping and passing chunks of the input file, to ensure |
| +// alignment, if the starting virtual addresses of YUV planes of the frames |
| +// in it were not 64 byte-aligned, we'd have to prepare a memory with 64 |
| +// byte-aligned starting address and make sure the addresses of YUV planes of |
| +// each frame are 64 byte-aligned before sending to the encoder. |
| +// Now we test resolutions different from coded size and prepare chunks before |
| +// encoding to avoid performance impact. |
| +// Use |visible_size| and |coded_size| to copy YUV data into memory from |
| +// |in_filename|. The copied result will be saved in |input_file|. Also |
| +// calculate the byte size of an input frame and set it to |coded_buffer_size|. |
| +static void PrepareInputBuffers(const gfx::Size& visible_size, |
| + const gfx::Size& coded_size, |
| + const std::string in_filename, |
| + base::MemoryMappedFile* input_file, |
| + size_t* coded_buffer_size) { |
| + base::FilePath temp_file; |
| + size_t input_num_planes = media::VideoFrame::NumPlanes(kInputFormat); |
| + std::vector<size_t> padding_sizes(input_num_planes); |
| + std::vector<size_t> coded_bpl(input_num_planes); |
| + std::vector<size_t> visible_bpl(input_num_planes); |
| + std::vector<size_t> visible_plane_rows(input_num_planes); |
| + |
| + // YUV plane starting address should be 64 bytes alignment. |
| + // Calculate padding size for each plane, and frame allocation size for |
|
wuchengli
2014/08/28 10:17:48
80 char aligned. Move this with the previous line.
henryhsu
2014/08/29 06:36:41
Done.
|
| + // coded size. And also store bytes per line information of coded size and |
|
wuchengli
2014/08/28 10:17:48
s/And also/Also/
henryhsu
2014/08/29 06:36:41
Done.
|
| + // visible size. |
| + *coded_buffer_size = 0; |
| + for (off_t i = 0; i < input_num_planes; i++) { |
| + size_t size = |
| + media::VideoFrame::PlaneAllocationSize(kInputFormat, i, coded_size); |
| + size_t padding_bytes = ALIGN_64_BYTES(size) - size; |
| + |
| + coded_bpl[i] = |
| + media::VideoFrame::RowBytes(i, kInputFormat, coded_size.width()); |
| + visible_bpl[i] = |
| + media::VideoFrame::RowBytes(i, kInputFormat, visible_size.width()); |
| + visible_plane_rows[i] = |
| + media::VideoFrame::Rows(i, kInputFormat, visible_size.height()); |
| + size_t padding_rows = |
| + media::VideoFrame::Rows(i, kInputFormat, coded_size.height()) - |
| + visible_plane_rows[i]; |
| + padding_sizes[i] = padding_rows * coded_bpl[i] + padding_bytes; |
| + *coded_buffer_size += ALIGN_64_BYTES(size); |
|
wuchengli
2014/08/28 10:17:48
Move this after padding_bytes so the use of |size|
henryhsu
2014/08/29 06:36:41
Done.
|
| + } |
| + |
| + // Test case may have many encoders and memory should be prepared once. |
| + if (input_file->IsValid()) |
| + return; |
| + |
| + base::MemoryMappedFile src_file; |
| + CHECK(base::CreateTemporaryFile(&temp_file)); |
| + CHECK(src_file.Initialize(base::FilePath(in_filename))); |
| + |
| + size_t visible_buffer_size = |
| + media::VideoFrame::AllocationSize(kInputFormat, visible_size); |
| + size_t num_frames = src_file.length() / visible_buffer_size; |
| + uint32 flags = base::File::FLAG_CREATE_ALWAYS | base::File::FLAG_WRITE | |
| + base::File::FLAG_READ; |
| + |
| + // Create a temporary file with coded_size length. |
| + base::File file(base::FilePath(temp_file), flags); |
| + file.Write(*coded_buffer_size * num_frames - 1, ".", 1); |
|
wuchengli
2014/08/28 10:17:48
Can we use SetLength for this? Is SetLength fast?
henryhsu
2014/08/29 06:36:41
Done.
|
| + CHECK(input_file->Initialize(file.Pass(), true)); |
| + |
| + off_t src_offset = 0, dest_offset = 0; |
| + while (src_offset < static_cast<off_t>(src_file.length())) { |
| + for (off_t i = 0; i < input_num_planes; i++) { |
| +#if defined(ARCH_CPU_ARMEL) |
| + // Assert that each plane of frame starts at 64-byte boundary. |
| + const uint8* ptr = input_file->data() + dest_offset; |
| + ASSERT_EQ(reinterpret_cast<off_t>(ptr) & 63, 0) |
| + << "Planes of frame should be mapped at a 64 byte boundary"; |
| +#endif |
| + for (off_t j = 0; j < visible_plane_rows[i]; j++) { |
| + const uint8* src = src_file.data() + src_offset; |
| + uint8* dest = const_cast<uint8*>(input_file->data() + dest_offset); |
|
wuchengli
2014/08/28 10:17:48
We should not return a const and access it. Maybe
henryhsu
2014/08/29 06:36:41
Done.
|
| + memcpy(dest, src, visible_bpl[i]); |
| + src_offset += visible_bpl[i]; |
| + dest_offset += coded_bpl[i]; |
| + } |
| + dest_offset += padding_sizes[i]; |
| + } |
| + } |
| + base::DeleteFile(temp_file, false); |
| +} |
| + |
| // Parse |data| into its constituent parts, set the various output fields |
| // accordingly, read in video stream, and store them to |test_streams|. |
| static void ParseAndReadTestStreamData(const base::FilePath::StringType& data, |
| @@ -129,7 +222,7 @@ static void ParseAndReadTestStreamData(const base::FilePath::StringType& data, |
| CHECK_LE(fields.size(), 9U) << data; |
| TestStream* test_stream = new TestStream(); |
| - base::FilePath::StringType filename = fields[0]; |
| + test_stream->in_filename = fields[0]; |
| int width, height; |
| CHECK(base::StringToInt(fields[1], &width)); |
| CHECK(base::StringToInt(fields[2], &height)); |
| @@ -161,7 +254,6 @@ static void ParseAndReadTestStreamData(const base::FilePath::StringType& data, |
| &test_stream->requested_subsequent_framerate)); |
| } |
| - CHECK(test_stream->input_file.Initialize(base::FilePath(filename))); |
| test_streams->push_back(test_stream); |
| } |
| } |
| @@ -552,29 +644,6 @@ VEAClient::VEAClient(const TestStream& test_stream, |
| EXPECT_EQ(0, base::WriteFile(out_filename, NULL, 0)); |
| } |
| - input_buffer_size_ = |
| - media::VideoFrame::AllocationSize(kInputFormat, test_stream.size); |
| - CHECK_GT(input_buffer_size_, 0UL); |
| - |
| - // Calculate the number of frames in the input stream by dividing its length |
| - // in bytes by frame size in bytes. |
| - CHECK_EQ(test_stream_.input_file.length() % input_buffer_size_, 0U) |
| - << "Stream byte size is not a product of calculated frame byte size"; |
| - num_frames_in_stream_ = test_stream_.input_file.length() / input_buffer_size_; |
| - CHECK_GT(num_frames_in_stream_, 0UL); |
| - CHECK_LE(num_frames_in_stream_, kMaxFrameNum); |
| - |
| - // We may need to loop over the stream more than once if more frames than |
| - // provided is required for bitrate tests. |
| - if (force_bitrate_ && num_frames_in_stream_ < kMinFramesForBitrateTests) { |
| - DVLOG(1) << "Stream too short for bitrate test (" << num_frames_in_stream_ |
| - << " frames), will loop it to reach " << kMinFramesForBitrateTests |
| - << " frames"; |
| - num_frames_to_encode_ = kMinFramesForBitrateTests; |
| - } else { |
| - num_frames_to_encode_ = num_frames_in_stream_; |
| - } |
| - |
| thread_checker_.DetachFromThread(); |
| } |
| @@ -629,35 +698,34 @@ void VEAClient::RequireBitstreamBuffers(unsigned int input_count, |
| ASSERT_EQ(state_, CS_INITIALIZED); |
| SetState(CS_ENCODING); |
| - // TODO(posciak): For now we only support input streams that meet encoder |
| - // size requirements exactly (i.e. coded size == visible size), so that we |
| - // can simply mmap the stream file and feed the encoder directly with chunks |
| - // of that, instead of memcpying from mmapped file into a separate set of |
| - // input buffers that would meet the coded size and alignment requirements. |
| - // If/when this is changed, the ARM-specific alignment check below should be |
| - // redone as well. |
| - input_coded_size_ = input_coded_size; |
| - ASSERT_EQ(input_coded_size_, test_stream_.size); |
| -#if defined(ARCH_CPU_ARMEL) |
| - // ARM performs CPU cache management with CPU cache line granularity. We thus |
| - // need to ensure our buffers are CPU cache line-aligned (64 byte-aligned). |
| - // Otherwise newer kernels will refuse to accept them, and on older kernels |
| - // we'll be treating ourselves to random corruption. |
| - // Since we are just mmapping and passing chunks of the input file, to ensure |
| - // alignment, if the starting virtual addresses of the frames in it were not |
| - // 64 byte-aligned, we'd have to use a separate set of input buffers and copy |
| - // the frames into them before sending to the encoder. It would have been an |
| - // overkill here though, because, for now at least, we only test resolutions |
| - // that result in proper alignment, and it would have also interfered with |
| - // performance testing. So just assert that the frame size is a multiple of |
| - // 64 bytes. This ensures all frames start at 64-byte boundary, because |
| - // MemoryMappedFile should be mmapp()ed at virtual page start as well. |
| - ASSERT_EQ(input_buffer_size_ & 63, 0u) |
| - << "Frame size has to be a multiple of 64 bytes"; |
| - ASSERT_EQ(reinterpret_cast<off_t>(test_stream_.input_file.data()) & 63, 0) |
| - << "Mapped file should be mapped at a 64 byte boundary"; |
| -#endif |
| + PrepareInputBuffers( |
| + test_stream_.size, |
| + input_coded_size, |
| + test_stream_.in_filename, |
| + const_cast<base::MemoryMappedFile*>(&test_stream_.input_file), |
|
wuchengli
2014/08/28 10:17:48
We should remove the const of test_stream_ because
henryhsu
2014/08/29 06:36:41
Done.
|
| + &input_buffer_size_); |
| + CHECK_GT(input_buffer_size_, 0UL); |
| + |
| + // Calculate the number of frames in the input stream by dividing its length |
| + // in bytes by frame size in bytes. |
| + CHECK_EQ(test_stream_.input_file.length() % input_buffer_size_, 0U) |
| + << "Stream byte size is not a product of calculated frame byte size"; |
| + num_frames_in_stream_ = test_stream_.input_file.length() / input_buffer_size_; |
| + CHECK_GT(num_frames_in_stream_, 0UL); |
| + CHECK_LE(num_frames_in_stream_, kMaxFrameNum); |
| + // We may need to loop over the stream more than once if more frames than |
| + // provided is required for bitrate tests. |
| + if (force_bitrate_ && num_frames_in_stream_ < kMinFramesForBitrateTests) { |
| + DVLOG(1) << "Stream too short for bitrate test (" << num_frames_in_stream_ |
| + << " frames), will loop it to reach " << kMinFramesForBitrateTests |
| + << " frames"; |
| + num_frames_to_encode_ = kMinFramesForBitrateTests; |
| + } else { |
| + num_frames_to_encode_ = num_frames_in_stream_; |
| + } |
| + |
| + input_coded_size_ = input_coded_size; |
| num_required_input_buffers_ = input_count; |
| ASSERT_GT(num_required_input_buffers_, 0UL); |
| @@ -743,8 +811,14 @@ void VEAClient::InputNoLongerNeededCallback(int32 input_id) { |
| scoped_refptr<media::VideoFrame> VEAClient::PrepareInputFrame(off_t position) { |
| CHECK_LE(position + input_buffer_size_, test_stream_.input_file.length()); |
| - uint8* frame_data = |
| + uint8* frame_data_y = |
| const_cast<uint8*>(test_stream_.input_file.data() + position); |
| + uint8* frame_data_u = |
| + frame_data_y + ALIGN_64_BYTES(media::VideoFrame::PlaneAllocationSize( |
| + kInputFormat, 0, input_coded_size_)); |
| + uint8* frame_data_v = |
| + frame_data_u + ALIGN_64_BYTES(media::VideoFrame::PlaneAllocationSize( |
| + kInputFormat, 1, input_coded_size_)); |
| CHECK_GT(current_framerate_, 0U); |
| scoped_refptr<media::VideoFrame> frame = |
| @@ -756,9 +830,9 @@ scoped_refptr<media::VideoFrame> VEAClient::PrepareInputFrame(off_t position) { |
| input_coded_size_.width(), |
| input_coded_size_.width() / 2, |
| input_coded_size_.width() / 2, |
| - frame_data, |
| - frame_data + input_coded_size_.GetArea(), |
| - frame_data + (input_coded_size_.GetArea() * 5 / 4), |
| + frame_data_y, |
| + frame_data_u, |
| + frame_data_v, |
| base::TimeDelta().FromMilliseconds( |
| next_input_id_ * base::Time::kMillisecondsPerSecond / |
| current_framerate_), |