Index: media/base/android/media_codec_bridge.cc |
diff --git a/media/base/android/media_codec_bridge.cc b/media/base/android/media_codec_bridge.cc |
index 0231c00777d6197d778c13bf451f7e547169fe07..bd285310761d438b898efef37391ce54a4a5142b 100644 |
--- a/media/base/android/media_codec_bridge.cc |
+++ b/media/base/android/media_codec_bridge.cc |
@@ -6,16 +6,174 @@ |
#include <algorithm> |
#include <limits> |
+#include <memory> |
+#include <utility> |
+#include "base/android/build_info.h" |
+#include "base/android/jni_android.h" |
+#include "base/android/jni_array.h" |
+#include "base/android/jni_string.h" |
#include "base/logging.h" |
+#include "base/numerics/safe_conversions.h" |
#include "base/strings/string_util.h" |
+#include "jni/MediaCodecBridge_jni.h" |
+#include "media/base/android/media_codec_util.h" |
+#include "media/base/bit_reader.h" |
#include "media/base/subsample_entry.h" |
+using base::android::AttachCurrentThread; |
+using base::android::ConvertJavaStringToUTF8; |
+using base::android::ConvertUTF8ToJavaString; |
+using base::android::JavaIntArrayToIntVector; |
+using base::android::ScopedJavaLocalRef; |
+ |
+#define RETURN_ON_ERROR(condition) \ |
+ do { \ |
+ if (!(condition)) { \ |
+ LOG(ERROR) << "Unable to parse AAC header: " #condition; \ |
+ return false; \ |
+ } \ |
+ } while (0) |
+ |
namespace media { |
+namespace { |
-MediaCodecBridge::MediaCodecBridge() {} |
+enum { |
+ kBufferFlagSyncFrame = 1, // BUFFER_FLAG_SYNC_FRAME |
+ kBufferFlagEndOfStream = 4, // BUFFER_FLAG_END_OF_STREAM |
+ kConfigureFlagEncode = 1, // CONFIGURE_FLAG_ENCODE |
+}; |
-MediaCodecBridge::~MediaCodecBridge() {} |
+const std::string AudioCodecToAndroidMimeType(const AudioCodec& codec) { |
+ switch (codec) { |
+ case kCodecMP3: |
+ return "audio/mpeg"; |
+ case kCodecVorbis: |
+ return "audio/vorbis"; |
+ case kCodecOpus: |
+ return "audio/opus"; |
+ case kCodecAAC: |
+ return "audio/mp4a-latm"; |
+ case kCodecAC3: |
+ return "audio/ac3"; |
+ case kCodecEAC3: |
+ return "audio/eac3"; |
+ default: |
+ return std::string(); |
+ } |
+} |
+ |
+const std::string VideoCodecToAndroidMimeType(const VideoCodec& codec) { |
+ switch (codec) { |
+ case kCodecH264: |
+ return "video/avc"; |
+ case kCodecHEVC: |
+ return "video/hevc"; |
+ case kCodecVP8: |
+ return "video/x-vnd.on2.vp8"; |
+ case kCodecVP9: |
+ return "video/x-vnd.on2.vp9"; |
+ default: |
+ return std::string(); |
+ } |
+} |
+ |
+static ScopedJavaLocalRef<jintArray> |
+ToJavaIntArray(JNIEnv* env, std::unique_ptr<jint[]> native_array, int size) { |
+ ScopedJavaLocalRef<jintArray> j_array(env, env->NewIntArray(size)); |
+ env->SetIntArrayRegion(j_array.obj(), 0, size, native_array.get()); |
+ return j_array; |
+} |
+ |
+} // namespace |
+ |
+MediaCodecBridge::MediaCodecBridge() = default; |
+ |
+MediaCodecBridge::MediaCodecBridge(const std::string& mime, |
+ bool is_secure, |
+ MediaCodecDirection direction, |
+ bool require_software_codec) { |
+ JNIEnv* env = AttachCurrentThread(); |
+ DCHECK(!mime.empty()); |
+ ScopedJavaLocalRef<jstring> j_mime = ConvertUTF8ToJavaString(env, mime); |
+ j_media_codec_.Reset(Java_MediaCodecBridge_create( |
+ env, j_mime, is_secure, direction, require_software_codec)); |
+} |
+ |
+MediaCodecBridge::~MediaCodecBridge() { |
+ JNIEnv* env = AttachCurrentThread(); |
+ if (j_media_codec_.obj()) |
+ Java_MediaCodecBridge_release(env, j_media_codec_); |
+} |
+ |
+bool MediaCodecBridge::Start() { |
+ JNIEnv* env = AttachCurrentThread(); |
+ return Java_MediaCodecBridge_start(env, j_media_codec_); |
+} |
+ |
+void MediaCodecBridge::Stop() { |
+ JNIEnv* env = AttachCurrentThread(); |
+ Java_MediaCodecBridge_stop(env, j_media_codec_); |
+} |
+ |
+MediaCodecStatus MediaCodecBridge::Flush() { |
+ JNIEnv* env = AttachCurrentThread(); |
+ return static_cast<MediaCodecStatus>( |
+ Java_MediaCodecBridge_flush(env, j_media_codec_)); |
+} |
+ |
+MediaCodecStatus MediaCodecBridge::GetOutputSize(gfx::Size* size) { |
+ JNIEnv* env = AttachCurrentThread(); |
+ ScopedJavaLocalRef<jobject> result = |
+ Java_MediaCodecBridge_getOutputFormat(env, j_media_codec_); |
+ MediaCodecStatus status = static_cast<MediaCodecStatus>( |
+ Java_GetOutputFormatResult_status(env, result)); |
+ if (status == MEDIA_CODEC_OK) { |
+ size->SetSize(Java_GetOutputFormatResult_width(env, result), |
+ Java_GetOutputFormatResult_height(env, result)); |
+ } |
+ return status; |
+} |
+ |
+MediaCodecStatus MediaCodecBridge::GetOutputSamplingRate(int* sampling_rate) { |
+ JNIEnv* env = AttachCurrentThread(); |
+ ScopedJavaLocalRef<jobject> result = |
+ Java_MediaCodecBridge_getOutputFormat(env, j_media_codec_); |
+ MediaCodecStatus status = static_cast<MediaCodecStatus>( |
+ Java_GetOutputFormatResult_status(env, result)); |
+ if (status == MEDIA_CODEC_OK) |
+ *sampling_rate = Java_GetOutputFormatResult_sampleRate(env, result); |
+ return status; |
+} |
+ |
+MediaCodecStatus MediaCodecBridge::GetOutputChannelCount(int* channel_count) { |
+ JNIEnv* env = AttachCurrentThread(); |
+ ScopedJavaLocalRef<jobject> result = |
+ Java_MediaCodecBridge_getOutputFormat(env, j_media_codec_); |
+ MediaCodecStatus status = static_cast<MediaCodecStatus>( |
+ Java_GetOutputFormatResult_status(env, result)); |
+ if (status == MEDIA_CODEC_OK) |
+ *channel_count = Java_GetOutputFormatResult_channelCount(env, result); |
+ return status; |
+} |
+ |
+MediaCodecStatus MediaCodecBridge::QueueInputBuffer( |
+ int index, |
+ const uint8_t* data, |
+ size_t data_size, |
+ base::TimeDelta presentation_time) { |
+ DVLOG(3) << __func__ << " " << index << ": " << data_size; |
+ if (data_size > |
+ base::checked_cast<size_t>(std::numeric_limits<int32_t>::max())) { |
+ return MEDIA_CODEC_ERROR; |
+ } |
+ if (data && !FillInputBuffer(index, data, data_size)) |
+ return MEDIA_CODEC_ERROR; |
+ JNIEnv* env = AttachCurrentThread(); |
+ return static_cast<MediaCodecStatus>(Java_MediaCodecBridge_queueInputBuffer( |
+ env, j_media_codec_, index, 0, data_size, |
+ presentation_time.InMicroseconds(), 0)); |
+} |
MediaCodecStatus MediaCodecBridge::QueueSecureInputBuffer( |
int index, |
@@ -26,12 +184,132 @@ MediaCodecStatus MediaCodecBridge::QueueSecureInputBuffer( |
const std::vector<SubsampleEntry>& subsamples, |
const EncryptionScheme& encryption_scheme, |
base::TimeDelta presentation_time) { |
- const std::vector<char> key_vec(key_id.begin(), key_id.end()); |
- const std::vector<char> iv_vec(iv.begin(), iv.end()); |
- return QueueSecureInputBuffer(index, data, data_size, key_vec, iv_vec, |
- subsamples.empty() ? nullptr : &subsamples[0], |
- (int)subsamples.size(), encryption_scheme, |
- presentation_time); |
+ DVLOG(3) << __func__ << " " << index << ": " << data_size; |
+ if (data_size > |
+ base::checked_cast<size_t>(std::numeric_limits<int32_t>::max())) { |
+ return MEDIA_CODEC_ERROR; |
+ } |
+ if (data && !FillInputBuffer(index, data, data_size)) |
+ return MEDIA_CODEC_ERROR; |
+ |
+ JNIEnv* env = AttachCurrentThread(); |
+ ScopedJavaLocalRef<jbyteArray> j_key_id = base::android::ToJavaByteArray( |
+ env, reinterpret_cast<const uint8_t*>(key_id.data()), key_id.size()); |
+ ScopedJavaLocalRef<jbyteArray> j_iv = base::android::ToJavaByteArray( |
+ env, reinterpret_cast<const uint8_t*>(iv.data()), iv.size()); |
+ |
+ // The MediaCodec.CryptoInfo documentation says to pass NULL for |clear_array| |
+ // to indicate that all data is encrypted. But it doesn't specify what |
+ // |cypher_array| and |subsamples_size| should be in that case. We pass |
+ // one subsample here just to be on the safe side. |
+ int num_subsamples = std::max(1u, subsamples.size()); |
+ |
+ std::unique_ptr<jint[]> native_clear_array(new jint[num_subsamples]); |
+ std::unique_ptr<jint[]> native_cypher_array(new jint[num_subsamples]); |
+ |
+ if (subsamples.empty()) { |
+ native_clear_array[0] = 0; |
+ native_cypher_array[0] = data_size; |
+ } else { |
+ for (size_t i = 0; i < subsamples.size(); ++i) { |
+ DCHECK(subsamples[i].clear_bytes <= std::numeric_limits<uint16_t>::max()); |
+ if (subsamples[i].cypher_bytes > |
+ static_cast<uint32_t>(std::numeric_limits<jint>::max())) { |
+ return MEDIA_CODEC_ERROR; |
+ } |
+ |
+ native_clear_array[i] = subsamples[i].clear_bytes; |
+ native_cypher_array[i] = subsamples[i].cypher_bytes; |
+ } |
+ } |
+ |
+ ScopedJavaLocalRef<jintArray> clear_array = |
+ ToJavaIntArray(env, std::move(native_clear_array), num_subsamples); |
+ ScopedJavaLocalRef<jintArray> cypher_array = |
+ ToJavaIntArray(env, std::move(native_cypher_array), num_subsamples); |
+ |
+ return static_cast<MediaCodecStatus>( |
+ Java_MediaCodecBridge_queueSecureInputBuffer( |
+ env, j_media_codec_.obj(), index, 0, j_iv.obj(), j_key_id.obj(), |
+ clear_array, cypher_array, num_subsamples, |
+ static_cast<int>(encryption_scheme.mode()), |
+ static_cast<int>(encryption_scheme.pattern().encrypt_blocks()), |
+ static_cast<int>(encryption_scheme.pattern().skip_blocks()), |
+ presentation_time.InMicroseconds())); |
+} |
+ |
+void MediaCodecBridge::QueueEOS(int input_buffer_index) { |
+ DVLOG(3) << __func__ << ": " << input_buffer_index; |
+ JNIEnv* env = AttachCurrentThread(); |
+ Java_MediaCodecBridge_queueInputBuffer( |
+ env, j_media_codec_, input_buffer_index, 0, 0, 0, kBufferFlagEndOfStream); |
+} |
+ |
+MediaCodecStatus MediaCodecBridge::DequeueInputBuffer(base::TimeDelta timeout, |
+ int* index) { |
+ JNIEnv* env = AttachCurrentThread(); |
+ ScopedJavaLocalRef<jobject> result = Java_MediaCodecBridge_dequeueInputBuffer( |
+ env, j_media_codec_, timeout.InMicroseconds()); |
+ *index = Java_DequeueInputResult_index(env, result); |
+ MediaCodecStatus status = static_cast<MediaCodecStatus>( |
+ Java_DequeueInputResult_status(env, result)); |
+ DVLOG(3) << __func__ << ": status: " << status << ", index: " << *index; |
+ return status; |
+} |
+ |
+MediaCodecStatus MediaCodecBridge::DequeueOutputBuffer( |
+ base::TimeDelta timeout, |
+ int* index, |
+ size_t* offset, |
+ size_t* size, |
+ base::TimeDelta* presentation_time, |
+ bool* end_of_stream, |
+ bool* key_frame) { |
+ JNIEnv* env = AttachCurrentThread(); |
+ ScopedJavaLocalRef<jobject> result = |
+ Java_MediaCodecBridge_dequeueOutputBuffer(env, j_media_codec_, |
+ timeout.InMicroseconds()); |
+ *index = Java_DequeueOutputResult_index(env, result); |
+ *offset = |
+ base::checked_cast<size_t>(Java_DequeueOutputResult_offset(env, result)); |
+ *size = base::checked_cast<size_t>( |
+ Java_DequeueOutputResult_numBytes(env, result)); |
+ if (presentation_time) { |
+ *presentation_time = base::TimeDelta::FromMicroseconds( |
+ Java_DequeueOutputResult_presentationTimeMicroseconds(env, result)); |
+ } |
+ int flags = Java_DequeueOutputResult_flags(env, result); |
+ if (end_of_stream) |
+ *end_of_stream = flags & kBufferFlagEndOfStream; |
+ if (key_frame) |
+ *key_frame = flags & kBufferFlagSyncFrame; |
+ MediaCodecStatus status = static_cast<MediaCodecStatus>( |
+ Java_DequeueOutputResult_status(env, result)); |
+ DVLOG(3) << __func__ << ": status: " << status << ", index: " << *index |
+ << ", offset: " << *offset << ", size: " << *size |
+ << ", flags: " << flags; |
+ return status; |
+} |
+ |
+void MediaCodecBridge::ReleaseOutputBuffer(int index, bool render) { |
+ DVLOG(3) << __func__ << ": " << index; |
+ JNIEnv* env = AttachCurrentThread(); |
+ Java_MediaCodecBridge_releaseOutputBuffer(env, j_media_codec_, index, render); |
+} |
+ |
+MediaCodecStatus MediaCodecBridge::GetInputBuffer(int input_buffer_index, |
+ uint8_t** data, |
+ size_t* capacity) { |
+ JNIEnv* env = AttachCurrentThread(); |
+ ScopedJavaLocalRef<jobject> j_buffer(Java_MediaCodecBridge_getInputBuffer( |
+ env, j_media_codec_, input_buffer_index)); |
+ if (j_buffer.is_null()) |
+ return MEDIA_CODEC_ERROR; |
+ |
+ *data = static_cast<uint8_t*>(env->GetDirectBufferAddress(j_buffer.obj())); |
+ *capacity = |
+ base::checked_cast<size_t>(env->GetDirectBufferCapacity(j_buffer.obj())); |
+ return MEDIA_CODEC_OK; |
} |
MediaCodecStatus MediaCodecBridge::CopyFromOutputBuffer(int index, |
@@ -49,6 +327,33 @@ MediaCodecStatus MediaCodecBridge::CopyFromOutputBuffer(int index, |
return status; |
} |
+MediaCodecStatus MediaCodecBridge::GetOutputBufferAddress(int index, |
+ size_t offset, |
+ const uint8_t** addr, |
+ size_t* capacity) { |
+ JNIEnv* env = AttachCurrentThread(); |
+ ScopedJavaLocalRef<jobject> j_buffer( |
+ Java_MediaCodecBridge_getOutputBuffer(env, j_media_codec_, index)); |
+ if (j_buffer.is_null()) |
+ return MEDIA_CODEC_ERROR; |
+ const size_t total_capacity = env->GetDirectBufferCapacity(j_buffer.obj()); |
+ CHECK_GE(total_capacity, offset); |
+ *addr = reinterpret_cast<const uint8_t*>( |
+ env->GetDirectBufferAddress(j_buffer.obj())) + |
+ offset; |
+ *capacity = total_capacity - offset; |
+ return MEDIA_CODEC_OK; |
+} |
+ |
+std::string MediaCodecBridge::GetName() { |
+ if (base::android::BuildInfo::GetInstance()->sdk_int() < 18) |
+ return ""; |
+ JNIEnv* env = AttachCurrentThread(); |
+ ScopedJavaLocalRef<jstring> j_name = |
+ Java_MediaCodecBridge_getName(env, j_media_codec_); |
+ return ConvertJavaStringToUTF8(env, j_name); |
+} |
+ |
bool MediaCodecBridge::FillInputBuffer(int index, |
const uint8_t* data, |
size_t size) { |
@@ -70,4 +375,346 @@ bool MediaCodecBridge::FillInputBuffer(int index, |
return true; |
} |
+// static |
+AudioCodecBridge* AudioCodecBridge::Create(const AudioCodec& codec) { |
+ if (!MediaCodecUtil::IsMediaCodecAvailable()) |
+ return nullptr; |
+ |
+ const std::string mime = AudioCodecToAndroidMimeType(codec); |
+ if (mime.empty()) |
+ return nullptr; |
+ |
+ std::unique_ptr<AudioCodecBridge> bridge(new AudioCodecBridge(mime)); |
+ if (!bridge->media_codec()) |
+ return nullptr; |
+ |
+ return bridge.release(); |
+} |
+ |
+// static |
+bool AudioCodecBridge::IsKnownUnaccelerated(const AudioCodec& codec) { |
+ return MediaCodecUtil::IsKnownUnaccelerated( |
+ AudioCodecToAndroidMimeType(codec), MEDIA_CODEC_DECODER); |
+} |
+ |
+AudioCodecBridge::AudioCodecBridge(const std::string& mime) |
+ // Audio codec doesn't care about security level and there is no need for |
+ // audio encoding yet. |
+ : MediaCodecBridge(mime, false, MEDIA_CODEC_DECODER, false) {} |
+ |
+bool AudioCodecBridge::ConfigureAndStart(const AudioDecoderConfig& config, |
+ jobject media_crypto) { |
+ const int channel_count = |
+ ChannelLayoutToChannelCount(config.channel_layout()); |
+ const int64_t codec_delay_ns = base::Time::kNanosecondsPerSecond * |
+ config.codec_delay() / |
+ config.samples_per_second(); |
+ const int64_t seek_preroll_ns = |
+ 1000LL * config.seek_preroll().InMicroseconds(); |
+ |
+ return ConfigureAndStart(config.codec(), config.samples_per_second(), |
+ channel_count, config.extra_data().data(), |
+ config.extra_data().size(), codec_delay_ns, |
+ seek_preroll_ns, media_crypto); |
+} |
+ |
+bool AudioCodecBridge::ConfigureAndStart(const AudioCodec& codec, |
+ int sample_rate, |
+ int channel_count, |
+ const uint8_t* extra_data, |
+ size_t extra_data_size, |
+ int64_t codec_delay_ns, |
+ int64_t seek_preroll_ns, |
+ jobject media_crypto) { |
+ DVLOG(2) << __func__ << ": " |
+ << " codec:" << GetCodecName(codec) |
+ << " samples_per_second:" << sample_rate |
+ << " channel_count:" << channel_count |
+ << " codec_delay_ns:" << codec_delay_ns |
+ << " seek_preroll_ns:" << seek_preroll_ns |
+ << " extra data size:" << extra_data_size |
+ << " media_crypto:" << media_crypto; |
+ DCHECK(media_codec()); |
+ |
+ std::string codec_string = AudioCodecToAndroidMimeType(codec); |
+ if (codec_string.empty()) |
+ return false; |
+ |
+ JNIEnv* env = AttachCurrentThread(); |
+ |
+ ScopedJavaLocalRef<jstring> j_mime = |
+ ConvertUTF8ToJavaString(env, codec_string); |
+ ScopedJavaLocalRef<jobject> j_format(Java_MediaCodecBridge_createAudioFormat( |
+ env, j_mime, sample_rate, channel_count)); |
+ DCHECK(!j_format.is_null()); |
+ |
+ if (!ConfigureMediaFormat(j_format.obj(), codec, extra_data, extra_data_size, |
+ codec_delay_ns, seek_preroll_ns)) { |
+ return false; |
+ } |
+ |
+ if (!Java_MediaCodecBridge_configureAudio(env, media_codec(), j_format, |
+ media_crypto, 0)) { |
+ return false; |
+ } |
+ |
+ return Start(); |
+} |
+ |
+bool AudioCodecBridge::ConfigureMediaFormat(jobject j_format, |
+ const AudioCodec& codec, |
+ const uint8_t* extra_data, |
+ size_t extra_data_size, |
+ int64_t codec_delay_ns, |
+ int64_t seek_preroll_ns) { |
+ if (extra_data_size == 0 && codec != kCodecOpus) |
+ return true; |
+ |
+ JNIEnv* env = AttachCurrentThread(); |
+ switch (codec) { |
+ case kCodecVorbis: { |
+ if (extra_data[0] != 2) { |
+ LOG(ERROR) << "Invalid number of vorbis headers before the codec " |
+ << "header: " << extra_data[0]; |
+ return false; |
+ } |
+ |
+ size_t header_length[2]; |
+ // |total_length| keeps track of the total number of bytes before the last |
+ // header. |
+ size_t total_length = 1; |
+ const uint8_t* current_pos = extra_data; |
+ // Calculate the length of the first 2 headers. |
+ for (int i = 0; i < 2; ++i) { |
+ header_length[i] = 0; |
+ while (total_length < extra_data_size) { |
+ size_t size = *(++current_pos); |
+ total_length += 1 + size; |
+ if (total_length > 0x80000000) { |
+ LOG(ERROR) << "Vorbis header size too large"; |
+ return false; |
+ } |
+ header_length[i] += size; |
+ if (size < 0xFF) |
+ break; |
+ } |
+ if (total_length >= extra_data_size) { |
+ LOG(ERROR) << "Invalid vorbis header size in the extra data"; |
+ return false; |
+ } |
+ } |
+ current_pos++; |
+ // The first header is identification header. |
+ ScopedJavaLocalRef<jbyteArray> first_header = |
+ base::android::ToJavaByteArray(env, current_pos, header_length[0]); |
+ Java_MediaCodecBridge_setCodecSpecificData(env, j_format, 0, |
+ first_header); |
+ // The last header is codec header. |
+ ScopedJavaLocalRef<jbyteArray> last_header = |
+ base::android::ToJavaByteArray(env, extra_data + total_length, |
+ extra_data_size - total_length); |
+ Java_MediaCodecBridge_setCodecSpecificData(env, j_format, 1, last_header); |
+ break; |
+ } |
+ case kCodecAAC: { |
+ media::BitReader reader(extra_data, extra_data_size); |
+ |
+ // The following code is copied from aac.cc |
+ // TODO(qinmin): refactor the code in aac.cc to make it more reusable. |
+ uint8_t profile = 0; |
+ uint8_t frequency_index = 0; |
+ uint8_t channel_config = 0; |
+ RETURN_ON_ERROR(reader.ReadBits(5, &profile)); |
+ RETURN_ON_ERROR(reader.ReadBits(4, &frequency_index)); |
+ |
+ if (0xf == frequency_index) |
+ RETURN_ON_ERROR(reader.SkipBits(24)); |
+ RETURN_ON_ERROR(reader.ReadBits(4, &channel_config)); |
+ |
+ if (profile == 5 || profile == 29) { |
+ // Read extension config. |
+ RETURN_ON_ERROR(reader.ReadBits(4, &frequency_index)); |
+ if (frequency_index == 0xf) |
+ RETURN_ON_ERROR(reader.SkipBits(24)); |
+ RETURN_ON_ERROR(reader.ReadBits(5, &profile)); |
+ } |
+ |
+ if (profile < 1 || profile > 4 || frequency_index == 0xf || |
+ channel_config > 7) { |
+ LOG(ERROR) << "Invalid AAC header"; |
+ return false; |
+ } |
+ |
+ const size_t kCsdLength = 2; |
+ uint8_t csd[kCsdLength]; |
+ csd[0] = profile << 3 | frequency_index >> 1; |
+ csd[1] = (frequency_index & 0x01) << 7 | channel_config << 3; |
+ ScopedJavaLocalRef<jbyteArray> byte_array = |
+ base::android::ToJavaByteArray(env, csd, kCsdLength); |
+ Java_MediaCodecBridge_setCodecSpecificData(env, j_format, 0, byte_array); |
+ |
+ // TODO(qinmin): pass an extra variable to this function to determine |
+ // whether we need to call this. |
+ Java_MediaCodecBridge_setFrameHasADTSHeader(env, j_format); |
+ break; |
+ } |
+ case kCodecOpus: { |
+ if (!extra_data || extra_data_size == 0 || codec_delay_ns < 0 || |
+ seek_preroll_ns < 0) { |
+ LOG(ERROR) << "Invalid Opus Header"; |
+ return false; |
+ } |
+ |
+ // csd0 - Opus Header |
+ ScopedJavaLocalRef<jbyteArray> csd0 = |
+ base::android::ToJavaByteArray(env, extra_data, extra_data_size); |
+ Java_MediaCodecBridge_setCodecSpecificData(env, j_format, 0, csd0); |
+ |
+ // csd1 - Codec Delay |
+ ScopedJavaLocalRef<jbyteArray> csd1 = base::android::ToJavaByteArray( |
+ env, reinterpret_cast<const uint8_t*>(&codec_delay_ns), |
+ sizeof(int64_t)); |
+ Java_MediaCodecBridge_setCodecSpecificData(env, j_format, 1, csd1); |
+ |
+ // csd2 - Seek Preroll |
+ ScopedJavaLocalRef<jbyteArray> csd2 = base::android::ToJavaByteArray( |
+ env, reinterpret_cast<const uint8_t*>(&seek_preroll_ns), |
+ sizeof(int64_t)); |
+ Java_MediaCodecBridge_setCodecSpecificData(env, j_format, 2, csd2); |
+ break; |
+ } |
+ default: |
+ LOG(ERROR) << "Invalid header encountered for codec: " |
+ << AudioCodecToAndroidMimeType(codec); |
+ return false; |
+ } |
+ return true; |
+} |
+ |
+// static |
+bool VideoCodecBridge::IsKnownUnaccelerated(const VideoCodec& codec, |
+ MediaCodecDirection direction) { |
+ return MediaCodecUtil::IsKnownUnaccelerated( |
+ VideoCodecToAndroidMimeType(codec), direction); |
+} |
+ |
+// static |
+VideoCodecBridge* VideoCodecBridge::CreateDecoder( |
+ const VideoCodec& codec, |
+ bool is_secure, |
+ const gfx::Size& size, |
+ jobject surface, |
+ jobject media_crypto, |
+ const std::vector<uint8_t>& csd0, |
+ const std::vector<uint8_t>& csd1, |
+ bool allow_adaptive_playback, |
+ bool require_software_codec) { |
+ if (!MediaCodecUtil::IsMediaCodecAvailable()) |
+ return nullptr; |
+ |
+ const std::string mime = VideoCodecToAndroidMimeType(codec); |
+ if (mime.empty()) |
+ return nullptr; |
+ |
+ std::unique_ptr<VideoCodecBridge> bridge(new VideoCodecBridge( |
+ mime, is_secure, MEDIA_CODEC_DECODER, require_software_codec)); |
+ if (!bridge->media_codec()) |
+ return nullptr; |
+ |
+ JNIEnv* env = AttachCurrentThread(); |
+ ScopedJavaLocalRef<jstring> j_mime = ConvertUTF8ToJavaString(env, mime); |
+ ScopedJavaLocalRef<jobject> j_format( |
+ Java_MediaCodecBridge_createVideoDecoderFormat(env, j_mime, size.width(), |
+ size.height())); |
+ DCHECK(!j_format.is_null()); |
+ |
+ if (!csd0.empty()) { |
+ ScopedJavaLocalRef<jbyteArray> j_csd0 = |
+ base::android::ToJavaByteArray(env, csd0.data(), csd0.size()); |
+ Java_MediaCodecBridge_setCodecSpecificData(env, j_format, 0, j_csd0); |
+ } |
+ |
+ if (!csd1.empty()) { |
+ ScopedJavaLocalRef<jbyteArray> j_csd1 = |
+ base::android::ToJavaByteArray(env, csd1.data(), csd1.size()); |
+ Java_MediaCodecBridge_setCodecSpecificData(env, j_format, 1, j_csd1); |
+ } |
+ |
+ if (!Java_MediaCodecBridge_configureVideo(env, bridge->media_codec(), |
+ j_format, surface, media_crypto, 0, |
+ allow_adaptive_playback)) { |
+ return nullptr; |
+ } |
+ |
+ return bridge->Start() ? bridge.release() : nullptr; |
+} |
+ |
+// static |
+VideoCodecBridge* VideoCodecBridge::CreateEncoder(const VideoCodec& codec, |
+ const gfx::Size& size, |
+ int bit_rate, |
+ int frame_rate, |
+ int i_frame_interval, |
+ int color_format) { |
+ if (!MediaCodecUtil::IsMediaCodecAvailable()) |
+ return nullptr; |
+ |
+ const std::string mime = VideoCodecToAndroidMimeType(codec); |
+ if (mime.empty()) |
+ return nullptr; |
+ |
+ std::unique_ptr<VideoCodecBridge> bridge( |
+ new VideoCodecBridge(mime, false, MEDIA_CODEC_ENCODER, false)); |
+ if (!bridge->media_codec()) |
+ return nullptr; |
+ |
+ JNIEnv* env = AttachCurrentThread(); |
+ ScopedJavaLocalRef<jstring> j_mime = ConvertUTF8ToJavaString(env, mime); |
+ ScopedJavaLocalRef<jobject> j_format( |
+ Java_MediaCodecBridge_createVideoEncoderFormat( |
+ env, bridge->media_codec(), j_mime, size.width(), size.height(), |
+ bit_rate, frame_rate, i_frame_interval, color_format)); |
+ DCHECK(!j_format.is_null()); |
+ if (!Java_MediaCodecBridge_configureVideo(env, bridge->media_codec(), |
+ j_format, nullptr, nullptr, |
+ kConfigureFlagEncode, true)) { |
+ return nullptr; |
+ } |
+ |
+ return bridge->Start() ? bridge.release() : nullptr; |
+} |
+ |
+VideoCodecBridge::VideoCodecBridge(const std::string& mime, |
+ bool is_secure, |
+ MediaCodecDirection direction, |
+ bool require_software_codec) |
+ : MediaCodecBridge(mime, is_secure, direction, require_software_codec), |
+ adaptive_playback_supported_for_testing_(-1) {} |
+ |
+bool VideoCodecBridge::SetSurface(jobject surface) { |
+ DCHECK_GE(base::android::BuildInfo::GetInstance()->sdk_int(), 23); |
+ JNIEnv* env = AttachCurrentThread(); |
+ return Java_MediaCodecBridge_setSurface(env, media_codec(), surface); |
+} |
+ |
+void VideoCodecBridge::SetVideoBitrate(int bps, int frame_rate) { |
+ JNIEnv* env = AttachCurrentThread(); |
+ Java_MediaCodecBridge_setVideoBitrate(env, media_codec(), bps, frame_rate); |
+} |
+ |
+void VideoCodecBridge::RequestKeyFrameSoon() { |
+ JNIEnv* env = AttachCurrentThread(); |
+ Java_MediaCodecBridge_requestKeyFrameSoon(env, media_codec()); |
+} |
+ |
+bool VideoCodecBridge::IsAdaptivePlaybackSupported(int width, int height) { |
+ if (adaptive_playback_supported_for_testing_ == 0) |
+ return false; |
+ else if (adaptive_playback_supported_for_testing_ > 0) |
+ return true; |
+ JNIEnv* env = AttachCurrentThread(); |
+ return Java_MediaCodecBridge_isAdaptivePlaybackSupported(env, media_codec(), |
+ width, height); |
+} |
+ |
} // namespace media |