| Index: media/ffmpeg/ffmpeg_unittest.cc
|
| diff --git a/media/ffmpeg/ffmpeg_unittest.cc b/media/ffmpeg/ffmpeg_unittest.cc
|
| deleted file mode 100644
|
| index 0bb7fa7ca4f5681aee2d01d1fc7421202a78a010..0000000000000000000000000000000000000000
|
| --- a/media/ffmpeg/ffmpeg_unittest.cc
|
| +++ /dev/null
|
| @@ -1,589 +0,0 @@
|
| -// Copyright (c) 2012 The Chromium Authors. All rights reserved.
|
| -// Use of this source code is governed by a BSD-style license that can be
|
| -// found in the LICENSE file.
|
| -
|
| -// ffmpeg_unittests verify that the parts of the FFmpeg API that Chromium uses
|
| -// function as advertised for each media format that Chromium supports. This
|
| -// mostly includes stuff like reporting proper timestamps, seeking to
|
| -// keyframes, and supporting certain features like reordered_opaque.
|
| -//
|
| -
|
| -#include <limits>
|
| -#include <queue>
|
| -
|
| -#include "base/base_paths.h"
|
| -#include "base/files/file_path.h"
|
| -#include "base/files/file_util.h"
|
| -#include "base/files/memory_mapped_file.h"
|
| -#include "base/memory/scoped_ptr.h"
|
| -#include "base/path_service.h"
|
| -#include "base/strings/string_util.h"
|
| -#include "base/test/perf_test_suite.h"
|
| -#include "base/test/perf_time_logger.h"
|
| -#include "media/base/media.h"
|
| -#include "media/ffmpeg/ffmpeg_common.h"
|
| -#include "media/filters/ffmpeg_glue.h"
|
| -#include "media/filters/in_memory_url_protocol.h"
|
| -#include "testing/gtest/include/gtest/gtest.h"
|
| -
|
| -int main(int argc, char** argv) {
|
| - return base::PerfTestSuite(argc, argv).Run();
|
| -}
|
| -
|
| -namespace media {
|
| -
|
| -// Mirror setting in ffmpeg_video_decoder.
|
| -static const int kDecodeThreads = 2;
|
| -
|
| -class AVPacketQueue {
|
| - public:
|
| - AVPacketQueue() {
|
| - }
|
| -
|
| - ~AVPacketQueue() {
|
| - flush();
|
| - }
|
| -
|
| - bool empty() {
|
| - return packets_.empty();
|
| - }
|
| -
|
| - AVPacket* peek() {
|
| - return packets_.front();
|
| - }
|
| -
|
| - void pop() {
|
| - AVPacket* packet = packets_.front();
|
| - packets_.pop();
|
| - av_free_packet(packet);
|
| - delete packet;
|
| - }
|
| -
|
| - void push(AVPacket* packet) {
|
| - av_dup_packet(packet);
|
| - packets_.push(packet);
|
| - }
|
| -
|
| - void flush() {
|
| - while (!empty()) {
|
| - pop();
|
| - }
|
| - }
|
| -
|
| - private:
|
| - std::queue<AVPacket*> packets_;
|
| -
|
| - DISALLOW_COPY_AND_ASSIGN(AVPacketQueue);
|
| -};
|
| -
|
| -// TODO(dalecurtis): We should really just use PipelineIntegrationTests instead
|
| -// of a one-off step decoder so we're exercising the real pipeline.
|
| -class FFmpegTest : public testing::TestWithParam<const char*> {
|
| - protected:
|
| - FFmpegTest()
|
| - : av_format_context_(NULL),
|
| - audio_stream_index_(-1),
|
| - video_stream_index_(-1),
|
| - decoded_audio_time_(AV_NOPTS_VALUE),
|
| - decoded_audio_duration_(AV_NOPTS_VALUE),
|
| - decoded_video_time_(AV_NOPTS_VALUE),
|
| - decoded_video_duration_(AV_NOPTS_VALUE),
|
| - duration_(AV_NOPTS_VALUE) {
|
| - InitializeFFmpeg();
|
| -
|
| - audio_buffer_.reset(av_frame_alloc());
|
| - video_buffer_.reset(av_frame_alloc());
|
| - }
|
| -
|
| - virtual ~FFmpegTest() {
|
| - }
|
| -
|
| - void OpenAndReadFile(const std::string& name) {
|
| - OpenFile(name);
|
| - OpenCodecs();
|
| - ReadRemainingFile();
|
| - }
|
| -
|
| - void OpenFile(const std::string& name) {
|
| - base::FilePath path;
|
| - PathService::Get(base::DIR_SOURCE_ROOT, &path);
|
| - path = path.AppendASCII("media")
|
| - .AppendASCII("test")
|
| - .AppendASCII("data")
|
| - .AppendASCII("content")
|
| - .AppendASCII(name.c_str());
|
| - EXPECT_TRUE(base::PathExists(path));
|
| -
|
| - CHECK(file_data_.Initialize(path));
|
| - protocol_.reset(new InMemoryUrlProtocol(
|
| - file_data_.data(), file_data_.length(), false));
|
| - glue_.reset(new FFmpegGlue(protocol_.get()));
|
| -
|
| - ASSERT_TRUE(glue_->OpenContext()) << "Could not open " << path.value();
|
| - av_format_context_ = glue_->format_context();
|
| - ASSERT_LE(0, avformat_find_stream_info(av_format_context_, NULL))
|
| - << "Could not find stream information for " << path.value();
|
| -
|
| - // Determine duration by picking max stream duration.
|
| - for (unsigned int i = 0; i < av_format_context_->nb_streams; ++i) {
|
| - AVStream* av_stream = av_format_context_->streams[i];
|
| - int64 duration = ConvertFromTimeBase(
|
| - av_stream->time_base, av_stream->duration).InMicroseconds();
|
| - duration_ = std::max(duration_, duration);
|
| - }
|
| -
|
| - // Final check to see if the container itself specifies a duration.
|
| - AVRational av_time_base = {1, AV_TIME_BASE};
|
| - int64 duration =
|
| - ConvertFromTimeBase(av_time_base,
|
| - av_format_context_->duration).InMicroseconds();
|
| - duration_ = std::max(duration_, duration);
|
| - }
|
| -
|
| - void OpenCodecs() {
|
| - for (unsigned int i = 0; i < av_format_context_->nb_streams; ++i) {
|
| - AVStream* av_stream = av_format_context_->streams[i];
|
| - AVCodecContext* av_codec_context = av_stream->codec;
|
| - AVCodec* av_codec = avcodec_find_decoder(av_codec_context->codec_id);
|
| -
|
| - EXPECT_TRUE(av_codec)
|
| - << "Could not find AVCodec with CodecID "
|
| - << av_codec_context->codec_id;
|
| -
|
| - av_codec_context->error_concealment = FF_EC_GUESS_MVS | FF_EC_DEBLOCK;
|
| - av_codec_context->thread_count = kDecodeThreads;
|
| -
|
| - EXPECT_EQ(0, avcodec_open2(av_codec_context, av_codec, NULL))
|
| - << "Could not open AVCodecContext with CodecID "
|
| - << av_codec_context->codec_id;
|
| -
|
| - if (av_codec->type == AVMEDIA_TYPE_AUDIO) {
|
| - EXPECT_EQ(-1, audio_stream_index_) << "Found multiple audio streams.";
|
| - audio_stream_index_ = static_cast<int>(i);
|
| - } else if (av_codec->type == AVMEDIA_TYPE_VIDEO) {
|
| - EXPECT_EQ(-1, video_stream_index_) << "Found multiple video streams.";
|
| - video_stream_index_ = static_cast<int>(i);
|
| - } else {
|
| - ADD_FAILURE() << "Found unknown stream type.";
|
| - }
|
| - }
|
| - }
|
| -
|
| - void Flush() {
|
| - if (has_audio()) {
|
| - audio_packets_.flush();
|
| - avcodec_flush_buffers(av_audio_context());
|
| - }
|
| - if (has_video()) {
|
| - video_packets_.flush();
|
| - avcodec_flush_buffers(av_video_context());
|
| - }
|
| - }
|
| -
|
| - void ReadUntil(int64 time) {
|
| - while (true) {
|
| - scoped_ptr<AVPacket> packet(new AVPacket());
|
| - if (av_read_frame(av_format_context_, packet.get()) < 0) {
|
| - break;
|
| - }
|
| -
|
| - int stream_index = static_cast<int>(packet->stream_index);
|
| - int64 packet_time = AV_NOPTS_VALUE;
|
| - if (stream_index == audio_stream_index_) {
|
| - packet_time =
|
| - ConvertFromTimeBase(av_audio_stream()->time_base, packet->pts)
|
| - .InMicroseconds();
|
| - audio_packets_.push(packet.release());
|
| - } else if (stream_index == video_stream_index_) {
|
| - packet_time =
|
| - ConvertFromTimeBase(av_video_stream()->time_base, packet->pts)
|
| - .InMicroseconds();
|
| - video_packets_.push(packet.release());
|
| - } else {
|
| - ADD_FAILURE() << "Found packet that belongs to unknown stream.";
|
| - }
|
| -
|
| - if (packet_time > time) {
|
| - break;
|
| - }
|
| - }
|
| - }
|
| -
|
| - void ReadRemainingFile() {
|
| - ReadUntil(std::numeric_limits<int64>::max());
|
| - }
|
| -
|
| - bool StepDecodeAudio() {
|
| - EXPECT_TRUE(has_audio());
|
| - if (!has_audio() || audio_packets_.empty()) {
|
| - return false;
|
| - }
|
| -
|
| - // Decode until output is produced, end of stream, or error.
|
| - while (true) {
|
| - int result = 0;
|
| - int got_audio = 0;
|
| - bool end_of_stream = false;
|
| -
|
| - AVPacket packet;
|
| - if (audio_packets_.empty()) {
|
| - av_init_packet(&packet);
|
| - end_of_stream = true;
|
| - } else {
|
| - memcpy(&packet, audio_packets_.peek(), sizeof(packet));
|
| - }
|
| -
|
| - av_frame_unref(audio_buffer_.get());
|
| - result = avcodec_decode_audio4(av_audio_context(), audio_buffer_.get(),
|
| - &got_audio, &packet);
|
| - if (!audio_packets_.empty()) {
|
| - audio_packets_.pop();
|
| - }
|
| -
|
| - EXPECT_GE(result, 0) << "Audio decode error.";
|
| - if (result < 0 || (got_audio == 0 && end_of_stream)) {
|
| - return false;
|
| - }
|
| -
|
| - if (result > 0) {
|
| - double microseconds = 1.0L * audio_buffer_->nb_samples /
|
| - av_audio_context()->sample_rate *
|
| - base::Time::kMicrosecondsPerSecond;
|
| - decoded_audio_duration_ = static_cast<int64>(microseconds);
|
| -
|
| - if (packet.pts == static_cast<int64>(AV_NOPTS_VALUE)) {
|
| - EXPECT_NE(decoded_audio_time_, static_cast<int64>(AV_NOPTS_VALUE))
|
| - << "We never received an initial timestamped audio packet! "
|
| - << "Looks like there's a seeking/parsing bug in FFmpeg.";
|
| - decoded_audio_time_ += decoded_audio_duration_;
|
| - } else {
|
| - decoded_audio_time_ =
|
| - ConvertFromTimeBase(av_audio_stream()->time_base, packet.pts)
|
| - .InMicroseconds();
|
| - }
|
| - return true;
|
| - }
|
| - }
|
| - return true;
|
| - }
|
| -
|
| - bool StepDecodeVideo() {
|
| - EXPECT_TRUE(has_video());
|
| - if (!has_video() || video_packets_.empty()) {
|
| - return false;
|
| - }
|
| -
|
| - // Decode until output is produced, end of stream, or error.
|
| - while (true) {
|
| - int result = 0;
|
| - int got_picture = 0;
|
| - bool end_of_stream = false;
|
| -
|
| - AVPacket packet;
|
| - if (video_packets_.empty()) {
|
| - av_init_packet(&packet);
|
| - end_of_stream = true;
|
| - } else {
|
| - memcpy(&packet, video_packets_.peek(), sizeof(packet));
|
| - }
|
| -
|
| - av_frame_unref(video_buffer_.get());
|
| - av_video_context()->reordered_opaque = packet.pts;
|
| - result = avcodec_decode_video2(av_video_context(), video_buffer_.get(),
|
| - &got_picture, &packet);
|
| - if (!video_packets_.empty()) {
|
| - video_packets_.pop();
|
| - }
|
| -
|
| - EXPECT_GE(result, 0) << "Video decode error.";
|
| - if (result < 0 || (got_picture == 0 && end_of_stream)) {
|
| - return false;
|
| - }
|
| -
|
| - if (got_picture) {
|
| - AVRational doubled_time_base;
|
| - doubled_time_base.den = av_video_stream()->r_frame_rate.num;
|
| - doubled_time_base.num = av_video_stream()->r_frame_rate.den;
|
| - doubled_time_base.den *= 2;
|
| -
|
| - decoded_video_time_ =
|
| - ConvertFromTimeBase(av_video_stream()->time_base,
|
| - video_buffer_->reordered_opaque)
|
| - .InMicroseconds();
|
| - decoded_video_duration_ =
|
| - ConvertFromTimeBase(doubled_time_base,
|
| - 2 + video_buffer_->repeat_pict)
|
| - .InMicroseconds();
|
| - return true;
|
| - }
|
| - }
|
| - }
|
| -
|
| - void DecodeRemainingAudio() {
|
| - while (StepDecodeAudio()) {}
|
| - }
|
| -
|
| - void DecodeRemainingVideo() {
|
| - while (StepDecodeVideo()) {}
|
| - }
|
| -
|
| - void SeekTo(double position) {
|
| - int64 seek_time =
|
| - static_cast<int64>(position * base::Time::kMicrosecondsPerSecond);
|
| - int flags = AVSEEK_FLAG_BACKWARD;
|
| -
|
| - // Passing -1 as our stream index lets FFmpeg pick a default stream.
|
| - // FFmpeg will attempt to use the lowest-index video stream, if present,
|
| - // followed by the lowest-index audio stream.
|
| - EXPECT_GE(0, av_seek_frame(av_format_context_, -1, seek_time, flags))
|
| - << "Failed to seek to position " << position;
|
| - Flush();
|
| - }
|
| -
|
| - bool has_audio() { return audio_stream_index_ >= 0; }
|
| - bool has_video() { return video_stream_index_ >= 0; }
|
| - int64 decoded_audio_time() { return decoded_audio_time_; }
|
| - int64 decoded_audio_duration() { return decoded_audio_duration_; }
|
| - int64 decoded_video_time() { return decoded_video_time_; }
|
| - int64 decoded_video_duration() { return decoded_video_duration_; }
|
| - int64 duration() { return duration_; }
|
| -
|
| - AVStream* av_audio_stream() {
|
| - return av_format_context_->streams[audio_stream_index_];
|
| - }
|
| - AVStream* av_video_stream() {
|
| - return av_format_context_->streams[video_stream_index_];
|
| - }
|
| - AVCodecContext* av_audio_context() {
|
| - return av_audio_stream()->codec;
|
| - }
|
| - AVCodecContext* av_video_context() {
|
| - return av_video_stream()->codec;
|
| - }
|
| -
|
| - private:
|
| - void InitializeFFmpeg() {
|
| - static bool initialized = false;
|
| - if (initialized) {
|
| - return;
|
| - }
|
| -
|
| - base::FilePath path;
|
| - PathService::Get(base::DIR_MODULE, &path);
|
| - EXPECT_TRUE(InitializeMediaLibrary(path))
|
| - << "Could not initialize media library.";
|
| -
|
| - initialized = true;
|
| - }
|
| -
|
| - AVFormatContext* av_format_context_;
|
| - int audio_stream_index_;
|
| - int video_stream_index_;
|
| - AVPacketQueue audio_packets_;
|
| - AVPacketQueue video_packets_;
|
| -
|
| - scoped_ptr<AVFrame, media::ScopedPtrAVFreeFrame> audio_buffer_;
|
| - scoped_ptr<AVFrame, media::ScopedPtrAVFreeFrame> video_buffer_;
|
| -
|
| - int64 decoded_audio_time_;
|
| - int64 decoded_audio_duration_;
|
| - int64 decoded_video_time_;
|
| - int64 decoded_video_duration_;
|
| - int64 duration_;
|
| -
|
| - base::MemoryMappedFile file_data_;
|
| - scoped_ptr<InMemoryUrlProtocol> protocol_;
|
| - scoped_ptr<FFmpegGlue> glue_;
|
| -
|
| - DISALLOW_COPY_AND_ASSIGN(FFmpegTest);
|
| -};
|
| -
|
| -#define FFMPEG_TEST_CASE(name, extension) \
|
| - INSTANTIATE_TEST_CASE_P(name##_##extension, FFmpegTest, \
|
| - testing::Values(#name "." #extension));
|
| -
|
| -// Covers all our basic formats.
|
| -FFMPEG_TEST_CASE(sync0, mp4);
|
| -FFMPEG_TEST_CASE(sync0, ogv);
|
| -FFMPEG_TEST_CASE(sync0, webm);
|
| -FFMPEG_TEST_CASE(sync1, m4a);
|
| -FFMPEG_TEST_CASE(sync1, mp3);
|
| -FFMPEG_TEST_CASE(sync1, mp4);
|
| -FFMPEG_TEST_CASE(sync1, ogg);
|
| -FFMPEG_TEST_CASE(sync1, ogv);
|
| -FFMPEG_TEST_CASE(sync1, webm);
|
| -FFMPEG_TEST_CASE(sync2, m4a);
|
| -FFMPEG_TEST_CASE(sync2, mp3);
|
| -FFMPEG_TEST_CASE(sync2, mp4);
|
| -FFMPEG_TEST_CASE(sync2, ogg);
|
| -FFMPEG_TEST_CASE(sync2, ogv);
|
| -FFMPEG_TEST_CASE(sync2, webm);
|
| -
|
| -// Covers our LayoutTest file.
|
| -FFMPEG_TEST_CASE(counting, ogv);
|
| -
|
| -TEST_P(FFmpegTest, Perf) {
|
| - {
|
| - base::PerfTimeLogger timer("Opening file");
|
| - OpenFile(GetParam());
|
| - }
|
| - {
|
| - base::PerfTimeLogger timer("Opening codecs");
|
| - OpenCodecs();
|
| - }
|
| - {
|
| - base::PerfTimeLogger timer("Reading file");
|
| - ReadRemainingFile();
|
| - }
|
| - if (has_audio()) {
|
| - base::PerfTimeLogger timer("Decoding audio");
|
| - DecodeRemainingAudio();
|
| - }
|
| - if (has_video()) {
|
| - base::PerfTimeLogger timer("Decoding video");
|
| - DecodeRemainingVideo();
|
| - }
|
| - {
|
| - base::PerfTimeLogger timer("Seeking to zero");
|
| - SeekTo(0);
|
| - }
|
| -}
|
| -
|
| -TEST_P(FFmpegTest, Loop_Audio) {
|
| - OpenAndReadFile(GetParam());
|
| - if (!has_audio()) {
|
| - return;
|
| - }
|
| -
|
| - const int kSteps = 4;
|
| - std::vector<int64> expected_timestamps_;
|
| - for (int i = 0; i < kSteps; ++i) {
|
| - EXPECT_TRUE(StepDecodeAudio());
|
| - expected_timestamps_.push_back(decoded_audio_time());
|
| - }
|
| -
|
| - SeekTo(0);
|
| - ReadRemainingFile();
|
| -
|
| - for (int i = 0; i < kSteps; ++i) {
|
| - EXPECT_TRUE(StepDecodeAudio());
|
| - EXPECT_EQ(expected_timestamps_[i], decoded_audio_time())
|
| - << "Frame " << i << " had a mismatched timestamp.";
|
| - }
|
| -}
|
| -
|
| -TEST_P(FFmpegTest, Loop_Video) {
|
| - OpenAndReadFile(GetParam());
|
| - if (!has_video()) {
|
| - return;
|
| - }
|
| -
|
| - const int kSteps = 4;
|
| - std::vector<int64> expected_timestamps_;
|
| - for (int i = 0; i < kSteps; ++i) {
|
| - EXPECT_TRUE(StepDecodeVideo());
|
| - expected_timestamps_.push_back(decoded_video_time());
|
| - }
|
| -
|
| - SeekTo(0);
|
| - ReadRemainingFile();
|
| -
|
| - for (int i = 0; i < kSteps; ++i) {
|
| - EXPECT_TRUE(StepDecodeVideo());
|
| - EXPECT_EQ(expected_timestamps_[i], decoded_video_time())
|
| - << "Frame " << i << " had a mismatched timestamp.";
|
| - }
|
| -}
|
| -
|
| -TEST_P(FFmpegTest, Seek_Audio) {
|
| - OpenAndReadFile(GetParam());
|
| - if (!has_audio() && duration() >= 0.5) {
|
| - return;
|
| - }
|
| -
|
| - SeekTo(duration() - 0.5);
|
| - ReadRemainingFile();
|
| -
|
| - EXPECT_TRUE(StepDecodeAudio());
|
| - EXPECT_NE(static_cast<int64>(AV_NOPTS_VALUE), decoded_audio_time());
|
| -}
|
| -
|
| -TEST_P(FFmpegTest, Seek_Video) {
|
| - OpenAndReadFile(GetParam());
|
| - if (!has_video() && duration() >= 0.5) {
|
| - return;
|
| - }
|
| -
|
| - SeekTo(duration() - 0.5);
|
| - ReadRemainingFile();
|
| -
|
| - EXPECT_TRUE(StepDecodeVideo());
|
| - EXPECT_NE(static_cast<int64>(AV_NOPTS_VALUE), decoded_video_time());
|
| -}
|
| -
|
| -TEST_P(FFmpegTest, Decode_Audio) {
|
| - OpenAndReadFile(GetParam());
|
| - if (!has_audio()) {
|
| - return;
|
| - }
|
| -
|
| - int64 last_audio_time = AV_NOPTS_VALUE;
|
| - while (StepDecodeAudio()) {
|
| - ASSERT_GT(decoded_audio_time(), last_audio_time);
|
| - last_audio_time = decoded_audio_time();
|
| - }
|
| -}
|
| -
|
| -TEST_P(FFmpegTest, Decode_Video) {
|
| - OpenAndReadFile(GetParam());
|
| - if (!has_video()) {
|
| - return;
|
| - }
|
| -
|
| - int64 last_video_time = AV_NOPTS_VALUE;
|
| - while (StepDecodeVideo()) {
|
| - ASSERT_GT(decoded_video_time(), last_video_time);
|
| - last_video_time = decoded_video_time();
|
| - }
|
| -}
|
| -
|
| -TEST_P(FFmpegTest, Duration) {
|
| - OpenAndReadFile(GetParam());
|
| -
|
| - if (has_audio()) {
|
| - DecodeRemainingAudio();
|
| - }
|
| -
|
| - if (has_video()) {
|
| - DecodeRemainingVideo();
|
| - }
|
| -
|
| - double expected = static_cast<double>(duration());
|
| - double actual = static_cast<double>(
|
| - std::max(decoded_audio_time() + decoded_audio_duration(),
|
| - decoded_video_time() + decoded_video_duration()));
|
| - EXPECT_NEAR(expected, actual, 500000)
|
| - << "Duration is off by more than 0.5 seconds.";
|
| -}
|
| -
|
| -TEST_F(FFmpegTest, VideoPlayedCollapse) {
|
| - OpenFile("test.ogv");
|
| - OpenCodecs();
|
| -
|
| - SeekTo(0.5);
|
| - ReadRemainingFile();
|
| - EXPECT_TRUE(StepDecodeVideo());
|
| - VLOG(1) << decoded_video_time();
|
| -
|
| - SeekTo(2.83);
|
| - ReadRemainingFile();
|
| - EXPECT_TRUE(StepDecodeVideo());
|
| - VLOG(1) << decoded_video_time();
|
| -
|
| - SeekTo(0.4);
|
| - ReadRemainingFile();
|
| - EXPECT_TRUE(StepDecodeVideo());
|
| - VLOG(1) << decoded_video_time();
|
| -}
|
| -
|
| -} // namespace media
|
|
|