Index: media/gpu/dxva_video_decode_accelerator_win.cc |
diff --git a/media/gpu/dxva_video_decode_accelerator_win.cc b/media/gpu/dxva_video_decode_accelerator_win.cc |
index e046b3582d47c2f95c2255c8a41fafbef21d9b40..a3ebb2d14593aeff7f5d82a6133ac04c39d8664f 100644 |
--- a/media/gpu/dxva_video_decode_accelerator_win.cc |
+++ b/media/gpu/dxva_video_decode_accelerator_win.cc |
@@ -500,7 +500,6 @@ DXVAVideoDecodeAccelerator::DXVAVideoDecodeAccelerator( |
!workarounds.disable_nv12_dxgi_video), |
use_dx11_(false), |
use_keyed_mutex_(false), |
- dx11_video_format_converter_media_type_needs_init_(true), |
using_angle_device_(false), |
enable_accelerated_vpx_decode_( |
gpu_preferences.enable_accelerated_vpx_decode), |
@@ -830,6 +829,12 @@ bool DXVAVideoDecodeAccelerator::CreateDX11DevManager() { |
&feature_level_out, d3d11_device_context_.Receive()); |
RETURN_ON_HR_FAILURE(hr, "Failed to create DX11 device", false); |
} |
+ |
+ hr = d3d11_device_.QueryInterface(video_device_.Receive()); |
+ RETURN_ON_HR_FAILURE(hr, "Failed to get video device", false); |
+ |
+ hr = d3d11_device_context_.QueryInterface(video_context_.Receive()); |
+ RETURN_ON_HR_FAILURE(hr, "Failed to get video context", false); |
} |
D3D11_FEATURE_DATA_D3D11_OPTIONS options; |
@@ -868,27 +873,6 @@ bool DXVAVideoDecodeAccelerator::CreateDX11DevManager() { |
hr = d3d11_device_->CreateQuery(&query_desc, d3d11_query_.Receive()); |
RETURN_ON_HR_FAILURE(hr, "Failed to create DX11 device query", false); |
- HMODULE video_processor_dll = ::GetModuleHandle(L"msvproc.dll"); |
- RETURN_ON_FAILURE(video_processor_dll, "Failed to load video processor", |
- false); |
- |
- hr = CreateCOMObjectFromDll(video_processor_dll, CLSID_VideoProcessorMFT, |
- __uuidof(IMFTransform), |
- video_format_converter_mft_.ReceiveVoid()); |
- RETURN_ON_HR_FAILURE(hr, "Failed to create video format converter", false); |
- |
- base::win::ScopedComPtr<IMFAttributes> converter_attributes; |
- hr = video_format_converter_mft_->GetAttributes( |
- converter_attributes.Receive()); |
- RETURN_ON_HR_FAILURE(hr, "Failed to get converter attributes", false); |
- |
- hr = converter_attributes->SetUINT32(MF_XVP_PLAYBACK_MODE, TRUE); |
- RETURN_ON_HR_FAILURE( |
- hr, "Failed to set MF_XVP_PLAYBACK_MODE attribute on converter", false); |
- |
- hr = converter_attributes->SetUINT32(MF_LOW_LATENCY, FALSE); |
- RETURN_ON_HR_FAILURE( |
- hr, "Failed to set MF_LOW_LATENCY attribute on converter", false); |
return true; |
} |
@@ -1967,17 +1951,16 @@ void DXVAVideoDecodeAccelerator::Invalidate() { |
pending_input_buffers_.clear(); |
pictures_requested_ = false; |
if (use_dx11_) { |
- if (video_format_converter_mft_.get()) { |
- video_format_converter_mft_->ProcessMessage( |
- MFT_MESSAGE_NOTIFY_END_STREAMING, 0); |
- video_format_converter_mft_.Release(); |
- } |
+ d3d11_processor_.Release(); |
+ enumerator_.Release(); |
+ video_context_.Release(); |
+ video_device_.Release(); |
d3d11_device_context_.Release(); |
d3d11_device_.Release(); |
d3d11_device_manager_.Release(); |
d3d11_query_.Release(); |
multi_threaded_.Release(); |
- dx11_video_format_converter_media_type_needs_init_ = true; |
+ processor_width_ = processor_height_ = 0; |
} else { |
d3d9_.Release(); |
d3d9_device_ex_.Release(); |
@@ -2278,8 +2261,6 @@ void DXVAVideoDecodeAccelerator::DecodeInternal( |
void DXVAVideoDecodeAccelerator::HandleResolutionChanged(int width, |
int height) { |
- dx11_video_format_converter_media_type_needs_init_ = true; |
- |
main_thread_task_runner_->PostTask( |
FROM_HERE, base::Bind(&DXVAVideoDecodeAccelerator::DismissStaleBuffers, |
weak_ptr_, false)); |
@@ -2569,21 +2550,14 @@ void DXVAVideoDecodeAccelerator::CopyTexture( |
// the StretchRect API in the IDirect3DDevice9Ex interface did the color |
// space conversion for us. Sadly in DX11 land the API does not provide |
// a straightforward way to do this. |
- // We use the video processor MFT. |
- // https://msdn.microsoft.com/en-us/library/hh162913(v=vs.85).aspx |
- // This object implements a media foundation transform (IMFTransform) |
- // which follows the same contract as the decoder. The color space |
- // conversion as per msdn is done in the GPU. |
D3D11_TEXTURE2D_DESC source_desc; |
src_texture->GetDesc(&source_desc); |
- |
- // Set up the input and output types for the video processor MFT. |
- if (!InitializeDX11VideoFormatConverterMediaType( |
- source_desc.Width, source_desc.Height, color_space)) { |
- RETURN_AND_NOTIFY_ON_FAILURE( |
- false, "Failed to initialize media types for convesion.", |
- PLATFORM_FAILURE, ); |
+ if (!InitializeID3D11VideoProcessor(source_desc.Width, source_desc.Height, |
+ color_space)) { |
+ RETURN_AND_NOTIFY_ON_FAILURE(false, |
+ "Failed to initialize D3D11 video processor.", |
+ PLATFORM_FAILURE, ); |
} |
// The input to the video processor is the output sample. |
@@ -2616,7 +2590,7 @@ void DXVAVideoDecodeAccelerator::CopyTextureOnDecoderThread( |
DCHECK(use_dx11_); |
DCHECK(!!input_sample); |
- DCHECK(video_format_converter_mft_.get()); |
+ DCHECK(d3d11_processor_.get()); |
if (dest_keyed_mutex) { |
HRESULT hr = |
@@ -2625,50 +2599,56 @@ void DXVAVideoDecodeAccelerator::CopyTextureOnDecoderThread( |
hr == S_OK, "D3D11 failed to acquire keyed mutex for texture.", |
PLATFORM_FAILURE, ); |
} |
- // The video processor MFT requires output samples to be allocated by the |
- // caller. We create a sample with a buffer backed with the ID3D11Texture2D |
- // interface exposed by ANGLE. This works nicely as this ensures that the |
- // video processor coverts the color space of the output frame and copies |
- // the result into the ANGLE texture. |
- base::win::ScopedComPtr<IMFSample> output_sample; |
- hr = MFCreateSample(output_sample.Receive()); |
- if (FAILED(hr)) { |
- RETURN_AND_NOTIFY_ON_HR_FAILURE(hr, "Failed to create output sample.", |
- PLATFORM_FAILURE, ); |
- } |
base::win::ScopedComPtr<IMFMediaBuffer> output_buffer; |
- hr = MFCreateDXGISurfaceBuffer(__uuidof(ID3D11Texture2D), dest_texture, 0, |
- FALSE, output_buffer.Receive()); |
- if (FAILED(hr)) { |
- RETURN_AND_NOTIFY_ON_HR_FAILURE(hr, "Failed to create output sample.", |
- PLATFORM_FAILURE, ); |
- } |
+ hr = input_sample->GetBufferByIndex(0, output_buffer.Receive()); |
+ RETURN_AND_NOTIFY_ON_HR_FAILURE(hr, "Failed to get buffer from output sample", |
+ PLATFORM_FAILURE, ); |
- output_sample->AddBuffer(output_buffer.get()); |
+ base::win::ScopedComPtr<IMFDXGIBuffer> dxgi_buffer; |
+ hr = dxgi_buffer.QueryFrom(output_buffer.get()); |
+ RETURN_AND_NOTIFY_ON_HR_FAILURE( |
+ hr, "Failed to get DXGIBuffer from output sample", PLATFORM_FAILURE, ); |
+ UINT index = 0; |
+ hr = dxgi_buffer->GetSubresourceIndex(&index); |
+ RETURN_AND_NOTIFY_ON_HR_FAILURE(hr, "Failed to get resource index", |
+ PLATFORM_FAILURE, ); |
- hr = video_format_converter_mft_->ProcessInput(0, input_sample.get(), 0); |
- if (FAILED(hr)) { |
- DCHECK(false); |
- RETURN_AND_NOTIFY_ON_HR_FAILURE( |
- hr, "Failed to convert output sample format.", PLATFORM_FAILURE, ); |
- } |
+ base::win::ScopedComPtr<ID3D11Texture2D> dx11_decoding_texture; |
+ hr = dxgi_buffer->GetResource(IID_PPV_ARGS(dx11_decoding_texture.Receive())); |
+ RETURN_AND_NOTIFY_ON_HR_FAILURE( |
+ hr, "Failed to get resource from output sample", PLATFORM_FAILURE, ); |
+ |
+ D3D11_VIDEO_PROCESSOR_OUTPUT_VIEW_DESC output_view_desc = { |
+ D3D11_VPOV_DIMENSION_TEXTURE2D}; |
+ output_view_desc.Texture2D.MipSlice = 0; |
+ base::win::ScopedComPtr<ID3D11VideoProcessorOutputView> output_view; |
+ hr = video_device_->CreateVideoProcessorOutputView( |
+ dest_texture, enumerator_.get(), &output_view_desc, |
+ output_view.Receive()); |
+ RETURN_AND_NOTIFY_ON_HR_FAILURE(hr, "Failed to get output view", |
+ PLATFORM_FAILURE, ); |
- input_sample.Release(); |
+ D3D11_VIDEO_PROCESSOR_INPUT_VIEW_DESC input_view_desc = {0}; |
+ input_view_desc.ViewDimension = D3D11_VPIV_DIMENSION_TEXTURE2D; |
+ input_view_desc.Texture2D.ArraySlice = index; |
+ input_view_desc.Texture2D.MipSlice = 0; |
+ base::win::ScopedComPtr<ID3D11VideoProcessorInputView> input_view; |
+ hr = video_device_->CreateVideoProcessorInputView( |
+ dx11_decoding_texture.get(), enumerator_.get(), &input_view_desc, |
+ input_view.Receive()); |
+ RETURN_AND_NOTIFY_ON_HR_FAILURE(hr, "Failed to get input view", |
+ PLATFORM_FAILURE, ); |
- DWORD status = 0; |
- MFT_OUTPUT_DATA_BUFFER format_converter_output = {}; |
- format_converter_output.pSample = output_sample.get(); |
- hr = video_format_converter_mft_->ProcessOutput( |
- 0, // No flags |
- 1, // # of out streams to pull from |
- &format_converter_output, &status); |
+ D3D11_VIDEO_PROCESSOR_STREAM streams = {0}; |
+ streams.Enable = TRUE; |
+ streams.pInputSurface = input_view.get(); |
- if (FAILED(hr)) { |
- DCHECK(false); |
- RETURN_AND_NOTIFY_ON_HR_FAILURE( |
- hr, "Failed to convert output sample format.", PLATFORM_FAILURE, ); |
- } |
+ hr = video_context_->VideoProcessorBlt(d3d11_processor_.get(), |
+ output_view.get(), 0, 1, &streams); |
+ |
+ RETURN_AND_NOTIFY_ON_HR_FAILURE(hr, "VideoProcessBlit failed", |
+ PLATFORM_FAILURE, ); |
if (dest_keyed_mutex) { |
HRESULT hr = dest_keyed_mutex->ReleaseSync(keyed_mutex_value + 1); |
@@ -2741,94 +2721,82 @@ void DXVAVideoDecodeAccelerator::FlushDecoder(int iterations, |
picture_buffer_id, input_buffer_id)); |
} |
-bool DXVAVideoDecodeAccelerator::InitializeDX11VideoFormatConverterMediaType( |
+bool DXVAVideoDecodeAccelerator::InitializeID3D11VideoProcessor( |
int width, |
int height, |
const gfx::ColorSpace& color_space) { |
- if (!dx11_video_format_converter_media_type_needs_init_ && |
- (!use_color_info_ || color_space == dx11_converter_color_space_)) { |
- return true; |
+ if (width < processor_width_ || height != processor_height_) { |
+ d3d11_processor_.Release(); |
+ enumerator_.Release(); |
+ processor_width_ = 0; |
+ processor_height_ = 0; |
+ |
+ D3D11_VIDEO_PROCESSOR_CONTENT_DESC desc; |
+ desc.InputFrameFormat = D3D11_VIDEO_FRAME_FORMAT_PROGRESSIVE; |
+ desc.InputFrameRate.Numerator = 60; |
+ desc.InputFrameRate.Denominator = 1; |
+ desc.InputWidth = width; |
+ desc.InputHeight = height; |
+ desc.OutputFrameRate.Numerator = 60; |
+ desc.OutputFrameRate.Denominator = 1; |
+ desc.OutputWidth = width; |
+ desc.OutputHeight = height; |
+ desc.Usage = D3D11_VIDEO_USAGE_PLAYBACK_NORMAL; |
+ |
+ HRESULT hr = video_device_->CreateVideoProcessorEnumerator( |
+ &desc, enumerator_.Receive()); |
+ RETURN_ON_HR_FAILURE(hr, "Failed to enumerate video processors", false); |
+ |
+ // TODO(Hubbe): Find correct index |
+ hr = video_device_->CreateVideoProcessor(enumerator_.get(), 0, |
+ d3d11_processor_.Receive()); |
+ RETURN_ON_HR_FAILURE(hr, "Failed to create video processor.", false); |
+ processor_width_ = width; |
+ processor_height_ = height; |
+ |
+ video_context_->VideoProcessorSetStreamAutoProcessingMode( |
+ d3d11_processor_.get(), 0, false); |
} |
- CHECK(video_format_converter_mft_.get()); |
- |
- HRESULT hr = video_format_converter_mft_->ProcessMessage( |
- MFT_MESSAGE_SET_D3D_MANAGER, |
- reinterpret_cast<ULONG_PTR>(d3d11_device_manager_.get())); |
- |
- if (FAILED(hr)) |
- DCHECK(false); |
- |
- RETURN_AND_NOTIFY_ON_HR_FAILURE(hr, |
- "Failed to initialize video format converter", |
- PLATFORM_FAILURE, false); |
- |
- video_format_converter_mft_->ProcessMessage(MFT_MESSAGE_NOTIFY_END_STREAMING, |
- 0); |
- |
- base::win::ScopedComPtr<IMFMediaType> media_type; |
- hr = MFCreateMediaType(media_type.Receive()); |
- RETURN_AND_NOTIFY_ON_HR_FAILURE(hr, "MFCreateMediaType failed", |
- PLATFORM_FAILURE, false); |
- |
- hr = media_type->SetGUID(MF_MT_MAJOR_TYPE, MFMediaType_Video); |
- RETURN_AND_NOTIFY_ON_HR_FAILURE(hr, "Failed to set major input type", |
- PLATFORM_FAILURE, false); |
- |
- hr = media_type->SetGUID(MF_MT_SUBTYPE, MFVideoFormat_NV12); |
- RETURN_AND_NOTIFY_ON_HR_FAILURE(hr, "Failed to set input sub type", |
- PLATFORM_FAILURE, false); |
- |
- hr = MFSetAttributeSize(media_type.get(), MF_MT_FRAME_SIZE, width, height); |
- RETURN_AND_NOTIFY_ON_HR_FAILURE(hr, "Failed to set media type attributes", |
- PLATFORM_FAILURE, false); |
- |
- if (use_color_info_) { |
- DXVA2_ExtendedFormat format = |
- gfx::ColorSpaceWin::GetExtendedFormat(color_space); |
- media_type->SetUINT32(MF_MT_YUV_MATRIX, format.VideoTransferMatrix); |
- media_type->SetUINT32(MF_MT_VIDEO_NOMINAL_RANGE, format.NominalRange); |
- media_type->SetUINT32(MF_MT_VIDEO_PRIMARIES, format.VideoPrimaries); |
- media_type->SetUINT32(MF_MT_TRANSFER_FUNCTION, |
- format.VideoTransferFunction); |
- dx11_converter_color_space_ = color_space; |
- } |
- |
- hr = video_format_converter_mft_->SetInputType(0, media_type.get(), 0); |
- if (FAILED(hr)) |
- DCHECK(false); |
- |
- RETURN_AND_NOTIFY_ON_HR_FAILURE(hr, "Failed to set converter input type", |
- PLATFORM_FAILURE, false); |
- |
- // It appears that we fail to set MFVideoFormat_ARGB32 as the output media |
- // type in certain configurations. Try to fallback to MFVideoFormat_RGB32 |
- // in such cases. If both fail, then bail. |
- |
- bool media_type_set = false; |
if (copy_nv12_textures_) { |
- media_type_set = SetTransformOutputType(video_format_converter_mft_.get(), |
- MFVideoFormat_NV12, width, height); |
- RETURN_AND_NOTIFY_ON_FAILURE(media_type_set, |
- "Failed to set NV12 converter output type", |
- PLATFORM_FAILURE, false); |
- } |
+ // If we're copying NV12 textures, make sure we set the same |
+ // color space on input and output. |
+ D3D11_VIDEO_PROCESSOR_COLOR_SPACE d3d11_color_space = {0}; |
+ d3d11_color_space.RGB_Range = 1; |
+ d3d11_color_space.Nominal_Range = D3D11_VIDEO_PROCESSOR_NOMINAL_RANGE_0_255; |
- if (!media_type_set) { |
- media_type_set = SetTransformOutputType( |
- video_format_converter_mft_.get(), MFVideoFormat_ARGB32, width, height); |
- } |
- if (!media_type_set) { |
- media_type_set = SetTransformOutputType(video_format_converter_mft_.get(), |
- MFVideoFormat_RGB32, width, height); |
- } |
+ video_context_->VideoProcessorSetOutputColorSpace(d3d11_processor_.get(), |
+ &d3d11_color_space); |
- if (!media_type_set) { |
- LOG(ERROR) << "Failed to find a matching RGB output type in the converter"; |
- return false; |
+ video_context_->VideoProcessorSetStreamColorSpace(d3d11_processor_.get(), 0, |
+ &d3d11_color_space); |
+ } else { |
+ // Not sure if this call is expensive, let's only do it if the color |
+ // space changes. |
+ gfx::ColorSpace output_color_space = gfx::ColorSpace::CreateSRGB(); |
+ if (use_color_info_ && dx11_converter_color_space_ != color_space) { |
+ base::win::ScopedComPtr<ID3D11VideoContext1> video_context1; |
+ HRESULT hr = video_context_.QueryInterface(video_context1.Receive()); |
+ if (SUCCEEDED(hr)) { |
+ video_context1->VideoProcessorSetStreamColorSpace1( |
+ d3d11_processor_.get(), 0, |
+ gfx::ColorSpaceWin::GetDXGIColorSpace(color_space)); |
+ video_context1->VideoProcessorSetOutputColorSpace1( |
+ d3d11_processor_.get(), |
+ gfx::ColorSpaceWin::GetDXGIColorSpace(output_color_space)); |
+ } else { |
+ D3D11_VIDEO_PROCESSOR_COLOR_SPACE d3d11_color_space = |
+ gfx::ColorSpaceWin::GetD3D11ColorSpace(color_space); |
+ video_context_->VideoProcessorSetStreamColorSpace( |
+ d3d11_processor_.get(), 0, &d3d11_color_space); |
+ d3d11_color_space = |
+ gfx::ColorSpaceWin::GetD3D11ColorSpace(output_color_space); |
+ video_context_->VideoProcessorSetOutputColorSpace( |
+ d3d11_processor_.get(), &d3d11_color_space); |
+ } |
+ dx11_converter_color_space_ = color_space; |
+ } |
} |
- |
- dx11_video_format_converter_media_type_needs_init_ = false; |
return true; |
} |