| Index: third_party/libva/va/va_vpp.h
|
| diff --git a/third_party/libva/va/va_vpp.h b/third_party/libva/va/va_vpp.h
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..1fdb53055e46e296a4110d145655262c9c5d6e3f
|
| --- /dev/null
|
| +++ b/third_party/libva/va/va_vpp.h
|
| @@ -0,0 +1,748 @@
|
| +/*
|
| + * Copyright (c) 2007-2011 Intel Corporation. All Rights Reserved.
|
| + *
|
| + * Permission is hereby granted, free of charge, to any person obtaining a
|
| + * copy of this software and associated documentation files (the
|
| + * "Software"), to deal in the Software without restriction, including
|
| + * without limitation the rights to use, copy, modify, merge, publish,
|
| + * distribute, sub license, and/or sell copies of the Software, and to
|
| + * permit persons to whom the Software is furnished to do so, subject to
|
| + * the following conditions:
|
| + *
|
| + * The above copyright notice and this permission notice (including the
|
| + * next paragraph) shall be included in all copies or substantial portions
|
| + * of the Software.
|
| + *
|
| + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
|
| + * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
| + * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
|
| + * IN NO EVENT SHALL INTEL AND/OR ITS SUPPLIERS BE LIABLE FOR
|
| + * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
|
| + * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
|
| + * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
| + */
|
| +
|
| +/**
|
| + * \file va_vpp.h
|
| + * \brief The video processing API
|
| + *
|
| + * This file contains the \ref api_vpp "Video processing API".
|
| + */
|
| +
|
| +#ifndef VA_VPP_H
|
| +#define VA_VPP_H
|
| +
|
| +#ifdef __cplusplus
|
| +extern "C" {
|
| +#endif
|
| +
|
| +/**
|
| + * \defgroup api_vpp Video processing API
|
| + *
|
| + * @{
|
| + *
|
| + * The video processing API uses the same paradigm as for decoding:
|
| + * - Query for supported filters;
|
| + * - Set up a video processing pipeline;
|
| + * - Send video processing parameters through VA buffers.
|
| + *
|
| + * \section api_vpp_caps Query for supported filters
|
| + *
|
| + * Checking whether video processing is supported can be performed
|
| + * with vaQueryConfigEntrypoints() and the profile argument set to
|
| + * #VAProfileNone. If video processing is supported, then the list of
|
| + * returned entry-points will include #VAEntrypointVideoProc.
|
| + *
|
| + * \code
|
| + * VAEntrypoint *entrypoints;
|
| + * int i, num_entrypoints, supportsVideoProcessing = 0;
|
| + *
|
| + * num_entrypoints = vaMaxNumEntrypoints();
|
| + * entrypoints = malloc(num_entrypoints * sizeof(entrypoints[0]);
|
| + * vaQueryConfigEntrypoints(va_dpy, VAProfileNone,
|
| + * entrypoints, &num_entrypoints);
|
| + *
|
| + * for (i = 0; !supportsVideoProcessing && i < num_entrypoints; i++) {
|
| + * if (entrypoints[i] == VAEntrypointVideoProc)
|
| + * supportsVideoProcessing = 1;
|
| + * }
|
| + * \endcode
|
| + *
|
| + * Then, the vaQueryVideoProcFilters() function is used to query the
|
| + * list of video processing filters.
|
| + *
|
| + * \code
|
| + * VAProcFilterType filters[VAProcFilterCount];
|
| + * unsigned int num_filters = VAProcFilterCount;
|
| + *
|
| + * // num_filters shall be initialized to the length of the array
|
| + * vaQueryVideoProcFilters(va_dpy, vpp_ctx, &filters, &num_filters);
|
| + * \endcode
|
| + *
|
| + * Finally, individual filter capabilities can be checked with
|
| + * vaQueryVideoProcFilterCaps().
|
| + *
|
| + * \code
|
| + * VAProcFilterCap denoise_caps;
|
| + * unsigned int num_denoise_caps = 1;
|
| + * vaQueryVideoProcFilterCaps(va_dpy, vpp_ctx,
|
| + * VAProcFilterNoiseReduction,
|
| + * &denoise_caps, &num_denoise_caps
|
| + * );
|
| + *
|
| + * VAProcFilterCapDeinterlacing deinterlacing_caps[VAProcDeinterlacingCount];
|
| + * unsigned int num_deinterlacing_caps = VAProcDeinterlacingCount;
|
| + * vaQueryVideoProcFilterCaps(va_dpy, vpp_ctx,
|
| + * VAProcFilterDeinterlacing,
|
| + * &deinterlacing_caps, &num_deinterlacing_caps
|
| + * );
|
| + * \endcode
|
| + *
|
| + * \section api_vpp_setup Set up a video processing pipeline
|
| + *
|
| + * A video processing pipeline buffer is created for each source
|
| + * surface we want to process. However, buffers holding filter
|
| + * parameters can be created once and for all. Rationale is to avoid
|
| + * multiple creation/destruction chains of filter buffers and also
|
| + * because filter parameters generally won't change frame after
|
| + * frame. e.g. this makes it possible to implement a checkerboard of
|
| + * videos where the same filters are applied to each video source.
|
| + *
|
| + * The general control flow is demonstrated by the following pseudo-code:
|
| + * \code
|
| + * // Create filters
|
| + * VABufferID denoise_filter, deint_filter;
|
| + * VABufferID filter_bufs[VAProcFilterCount];
|
| + * unsigned int num_filter_bufs;
|
| + *
|
| + * for (i = 0; i < num_filters; i++) {
|
| + * switch (filters[i]) {
|
| + * case VAProcFilterNoiseReduction: { // Noise reduction filter
|
| + * VAProcFilterParameterBuffer denoise;
|
| + * denoise.type = VAProcFilterNoiseReduction;
|
| + * denoise.value = 0.5;
|
| + * vaCreateBuffer(va_dpy, vpp_ctx,
|
| + * VAProcFilterParameterBufferType, sizeof(denoise), 1,
|
| + * &denoise, &denoise_filter
|
| + * );
|
| + * filter_bufs[num_filter_bufs++] = denoise_filter;
|
| + * break;
|
| + * }
|
| + *
|
| + * case VAProcFilterDeinterlacing: // Motion-adaptive deinterlacing
|
| + * for (j = 0; j < num_deinterlacing_caps; j++) {
|
| + * VAProcFilterCapDeinterlacing * const cap = &deinterlacing_caps[j];
|
| + * if (cap->type != VAProcDeinterlacingMotionAdaptive)
|
| + * continue;
|
| + *
|
| + * VAProcFilterParameterBufferDeinterlacing deint;
|
| + * deint.type = VAProcFilterDeinterlacing;
|
| + * deint.algorithm = VAProcDeinterlacingMotionAdaptive;
|
| + * vaCreateBuffer(va_dpy, vpp_ctx,
|
| + * VAProcFilterParameterBufferType, sizeof(deint), 1,
|
| + * &deint, &deint_filter
|
| + * );
|
| + * filter_bufs[num_filter_bufs++] = deint_filter;
|
| + * }
|
| + * }
|
| + * }
|
| + * \endcode
|
| + *
|
| + * Once the video processing pipeline is set up, the caller shall check the
|
| + * implied capabilities and requirements with vaQueryVideoProcPipelineCaps().
|
| + * This function can be used to validate the number of reference frames are
|
| + * needed by the specified deinterlacing algorithm, the supported color
|
| + * primaries, etc.
|
| + * \code
|
| + * // Create filters
|
| + * VAProcPipelineCaps pipeline_caps;
|
| + * VASurfaceID *forward_references;
|
| + * unsigned int num_forward_references;
|
| + * VASurfaceID *backward_references;
|
| + * unsigned int num_backward_references;
|
| + * VAProcColorStandardType in_color_standards[VAProcColorStandardCount];
|
| + * VAProcColorStandardType out_color_standards[VAProcColorStandardCount];
|
| + *
|
| + * pipeline_caps.input_color_standards = NULL;
|
| + * pipeline_caps.num_input_color_standards = ARRAY_ELEMS(in_color_standards);
|
| + * pipeline_caps.output_color_standards = NULL;
|
| + * pipeline_caps.num_output_color_standards = ARRAY_ELEMS(out_color_standards);
|
| + * vaQueryVideoProcPipelineCaps(va_dpy, vpp_ctx,
|
| + * filter_bufs, num_filter_bufs,
|
| + * &pipeline_caps
|
| + * );
|
| + *
|
| + * num_forward_references = pipeline_caps.num_forward_references;
|
| + * forward_references =
|
| + * malloc(num__forward_references * sizeof(VASurfaceID));
|
| + * num_backward_references = pipeline_caps.num_backward_references;
|
| + * backward_references =
|
| + * malloc(num_backward_references * sizeof(VASurfaceID));
|
| + * \endcode
|
| + *
|
| + * \section api_vpp_submit Send video processing parameters through VA buffers
|
| + *
|
| + * Video processing pipeline parameters are submitted for each source
|
| + * surface to process. Video filter parameters can also change, per-surface.
|
| + * e.g. the list of reference frames used for deinterlacing.
|
| + *
|
| + * \code
|
| + * foreach (iteration) {
|
| + * vaBeginPicture(va_dpy, vpp_ctx, vpp_surface);
|
| + * foreach (surface) {
|
| + * VARectangle output_region;
|
| + * VABufferID pipeline_buf;
|
| + * VAProcPipelineParameterBuffer *pipeline_param;
|
| + *
|
| + * vaCreateBuffer(va_dpy, vpp_ctx,
|
| + * VAProcPipelineParameterBuffer, sizeof(*pipeline_param), 1,
|
| + * NULL, &pipeline_buf
|
| + * );
|
| + *
|
| + * // Setup output region for this surface
|
| + * // e.g. upper left corner for the first surface
|
| + * output_region.x = BORDER;
|
| + * output_region.y = BORDER;
|
| + * output_region.width =
|
| + * (vpp_surface_width - (Nx_surfaces + 1) * BORDER) / Nx_surfaces;
|
| + * output_region.height =
|
| + * (vpp_surface_height - (Ny_surfaces + 1) * BORDER) / Ny_surfaces;
|
| + *
|
| + * vaMapBuffer(va_dpy, pipeline_buf, &pipeline_param);
|
| + * pipeline_param->surface = surface;
|
| + * pipeline_param->surface_region = NULL;
|
| + * pipeline_param->output_region = &output_region;
|
| + * pipeline_param->output_background_color = 0;
|
| + * if (first surface to render)
|
| + * pipeline_param->output_background_color = 0xff000000; // black
|
| + * pipeline_param->filter_flags = VA_FILTER_SCALING_HQ;
|
| + * pipeline_param->filters = filter_bufs;
|
| + * pipeline_param->num_filters = num_filter_bufs;
|
| + * vaUnmapBuffer(va_dpy, pipeline_buf);
|
| + *
|
| + * // Update reference frames for deinterlacing, if necessary
|
| + * pipeline_param->forward_references = forward_references;
|
| + * pipeline_param->num_forward_references = num_forward_references_used;
|
| + * pipeline_param->backward_references = backward_references;
|
| + * pipeline_param->num_backward_references = num_bacward_references_used;
|
| + *
|
| + * // Apply filters
|
| + * vaRenderPicture(va_dpy, vpp_ctx, &pipeline_buf, 1);
|
| + * }
|
| + * vaEndPicture(va_dpy, vpp_ctx);
|
| + * }
|
| + * \endcode
|
| + */
|
| +
|
| +/** \brief Video filter types. */
|
| +typedef enum _VAProcFilterType {
|
| + VAProcFilterNone = 0,
|
| + /** \brief Noise reduction filter. */
|
| + VAProcFilterNoiseReduction,
|
| + /** \brief Deinterlacing filter. */
|
| + VAProcFilterDeinterlacing,
|
| + /** \brief Sharpening filter. */
|
| + VAProcFilterSharpening,
|
| + /** \brief Color balance parameters. */
|
| + VAProcFilterColorBalance,
|
| + VAProcFilterCount
|
| +} VAProcFilterType;
|
| +
|
| +/** \brief Deinterlacing types. */
|
| +typedef enum _VAProcDeinterlacingType {
|
| + VAProcDeinterlacingNone = 0,
|
| + /** \brief Bob deinterlacing algorithm. */
|
| + VAProcDeinterlacingBob,
|
| + /** \brief Weave deinterlacing algorithm. */
|
| + VAProcDeinterlacingWeave,
|
| + /** \brief Motion adaptive deinterlacing algorithm. */
|
| + VAProcDeinterlacingMotionAdaptive,
|
| + /** \brief Motion compensated deinterlacing algorithm. */
|
| + VAProcDeinterlacingMotionCompensated,
|
| + /** \brief Number of deinterlacing algorithms. */
|
| + VAProcDeinterlacingCount
|
| +} VAProcDeinterlacingType;
|
| +
|
| +/** \brief Color balance types. */
|
| +typedef enum _VAProcColorBalanceType {
|
| + VAProcColorBalanceNone = 0,
|
| + /** \brief Hue. */
|
| + VAProcColorBalanceHue,
|
| + /** \brief Saturation. */
|
| + VAProcColorBalanceSaturation,
|
| + /** \brief Brightness. */
|
| + VAProcColorBalanceBrightness,
|
| + /** \brief Contrast. */
|
| + VAProcColorBalanceContrast,
|
| + /** \brief Automatically adjusted saturation. */
|
| + VAProcColorBalanceAutoSaturation,
|
| + /** \brief Automatically adjusted brightness. */
|
| + VAProcColorBalanceAutoBrightness,
|
| + /** \brief Automatically adjusted contrast. */
|
| + VAProcColorBalanceAutoContrast,
|
| + /** \brief Number of color balance attributes. */
|
| + VAProcColorBalanceCount
|
| +} VAProcColorBalanceType;
|
| +
|
| +/** \brief Color standard types. */
|
| +typedef enum _VAProcColorStandardType {
|
| + VAProcColorStandardNone = 0,
|
| + /** \brief ITU-R BT.601. */
|
| + VAProcColorStandardBT601,
|
| + /** \brief ITU-R BT.709. */
|
| + VAProcColorStandardBT709,
|
| + /** \brief ITU-R BT.470-2 System M. */
|
| + VAProcColorStandardBT470M,
|
| + /** \brief ITU-R BT.470-2 System B, G. */
|
| + VAProcColorStandardBT470BG,
|
| + /** \brief SMPTE-170M. */
|
| + VAProcColorStandardSMPTE170M,
|
| + /** \brief SMPTE-240M. */
|
| + VAProcColorStandardSMPTE240M,
|
| + /** \brief Generic film. */
|
| + VAProcColorStandardGenericFilm,
|
| + /** \brief Number of color standards. */
|
| + VAProcColorStandardCount
|
| +} VAProcColorStandardType;
|
| +
|
| +/** @name Video pipeline flags */
|
| +/**@{*/
|
| +/** \brief Specifies whether to apply subpictures when processing a surface. */
|
| +#define VA_PROC_PIPELINE_SUBPICTURES 0x00000001
|
| +/**
|
| + * \brief Specifies whether to apply power or performance
|
| + * optimizations to a pipeline.
|
| + *
|
| + * When processing several surfaces, it may be necessary to prioritize
|
| + * more certain pipelines than others. This flag is only a hint to the
|
| + * video processor so that it can omit certain filters to save power
|
| + * for example. Typically, this flag could be used with video surfaces
|
| + * decoded from a secondary bitstream.
|
| + */
|
| +#define VA_PROC_PIPELINE_FAST 0x00000002
|
| +/**@}*/
|
| +
|
| +/** @name Video filter flags */
|
| +/**@{*/
|
| +/** \brief Specifies whether the filter shall be present in the pipeline. */
|
| +#define VA_PROC_FILTER_MANDATORY 0x00000001
|
| +/**@}*/
|
| +
|
| +/** @name Pipeline end flags */
|
| +/**@{*/
|
| +/** \brief Specifies the pipeline is the last. */
|
| +#define VA_PIPELINE_FLAG_END 0x00000004
|
| +/**@}*/
|
| +
|
| +/** \brief Video processing pipeline capabilities. */
|
| +typedef struct _VAProcPipelineCaps {
|
| + /** \brief Pipeline flags. See VAProcPipelineParameterBuffer::pipeline_flags. */
|
| + unsigned int pipeline_flags;
|
| + /** \brief Extra filter flags. See VAProcPipelineParameterBuffer::filter_flags. */
|
| + unsigned int filter_flags;
|
| + /** \brief Number of forward reference frames that are needed. */
|
| + unsigned int num_forward_references;
|
| + /** \brief Number of backward reference frames that are needed. */
|
| + unsigned int num_backward_references;
|
| + /** \brief List of color standards supported on input. */
|
| + VAProcColorStandardType *input_color_standards;
|
| + /** \brief Number of elements in \ref input_color_standards array. */
|
| + unsigned int num_input_color_standards;
|
| + /** \brief List of color standards supported on output. */
|
| + VAProcColorStandardType *output_color_standards;
|
| + /** \brief Number of elements in \ref output_color_standards array. */
|
| + unsigned int num_output_color_standards;
|
| +} VAProcPipelineCaps;
|
| +
|
| +/** \brief Specification of values supported by the filter. */
|
| +typedef struct _VAProcFilterValueRange {
|
| + /** \brief Minimum value supported, inclusive. */
|
| + float min_value;
|
| + /** \brief Maximum value supported, inclusive. */
|
| + float max_value;
|
| + /** \brief Default value. */
|
| + float default_value;
|
| + /** \brief Step value that alters the filter behaviour in a sensible way. */
|
| + float step;
|
| +} VAProcFilterValueRange;
|
| +
|
| +/**
|
| + * \brief Video processing pipeline configuration.
|
| + *
|
| + * This buffer defines a video processing pipeline. As for any buffer
|
| + * passed to \c vaRenderPicture(), this is a one-time usage model.
|
| + * However, the actual filters to be applied are provided in the
|
| + * \c filters field, so they can be re-used in other processing
|
| + * pipelines.
|
| + *
|
| + * The target surface is specified by the \c render_target argument of
|
| + * \c vaBeginPicture(). The general usage model is described as follows:
|
| + * - \c vaBeginPicture(): specify the target surface that receives the
|
| + * processed output;
|
| + * - \c vaRenderPicture(): specify a surface to be processed and composed
|
| + * into the \c render_target. Use as many \c vaRenderPicture() calls as
|
| + * necessary surfaces to compose ;
|
| + * - \c vaEndPicture(): tell the driver to start processing the surfaces
|
| + * with the requested filters.
|
| + *
|
| + * If a filter (e.g. noise reduction) needs to be applied with different
|
| + * values for multiple surfaces, the application needs to create as many
|
| + * filter parameter buffers as necessary. i.e. the filter parameters shall
|
| + * not change between two calls to \c vaRenderPicture().
|
| + *
|
| + * For composition usage models, the first surface to process will generally
|
| + * use an opaque background color, i.e. \c output_background_color set with
|
| + * the most significant byte set to \c 0xff. For instance, \c 0xff000000 for
|
| + * a black background. Then, subsequent surfaces would use a transparent
|
| + * background color.
|
| + */
|
| +typedef struct _VAProcPipelineParameterBuffer {
|
| + /**
|
| + * \brief Source surface ID.
|
| + *
|
| + * ID of the source surface to process. If subpictures are associated
|
| + * with the video surfaces then they shall be rendered to the target
|
| + * surface, if the #VA_PROC_PIPELINE_SUBPICTURES pipeline flag is set.
|
| + */
|
| + VASurfaceID surface;
|
| + /**
|
| + * \brief Region within the source surface to be processed.
|
| + *
|
| + * Pointer to a #VARectangle defining the region within the source
|
| + * surface to be processed. If NULL, \c surface_region implies the
|
| + * whole surface.
|
| + */
|
| + const VARectangle *surface_region;
|
| + /**
|
| + * \brief Requested input color primaries.
|
| + *
|
| + * Color primaries are implicitly converted throughout the processing
|
| + * pipeline. The video processor chooses the best moment to apply
|
| + * this conversion. The set of supported color primaries primaries
|
| + * for input shall be queried with vaQueryVideoProcPipelineCaps().
|
| + */
|
| + VAProcColorStandardType surface_color_standard;
|
| + /**
|
| + * \brief Region within the output surface.
|
| + *
|
| + * Pointer to a #VARectangle defining the region within the output
|
| + * surface that receives the processed pixels. If NULL, \c output_region
|
| + * implies the whole surface.
|
| + *
|
| + * Note that any pixels residing outside the specified region will
|
| + * be filled in with the \ref output_background_color.
|
| + */
|
| + const VARectangle *output_region;
|
| + /**
|
| + * \brief Background color.
|
| + *
|
| + * Background color used to fill in pixels that reside outside of the
|
| + * specified \ref output_region. The color is specified in ARGB format:
|
| + * [31:24] alpha, [23:16] red, [15:8] green, [7:0] blue.
|
| + *
|
| + * Unless the alpha value is zero or the \ref output_region represents
|
| + * the whole target surface size, implementations shall not render the
|
| + * source surface to the target surface directly. Rather, in order to
|
| + * maintain the exact semantics of \ref output_background_color, the
|
| + * driver shall use a temporary surface and fill it in with the
|
| + * appropriate background color. Next, the driver will blend this
|
| + * temporary surface into the target surface.
|
| + */
|
| + unsigned int output_background_color;
|
| + /**
|
| + * \brief Requested output color primaries.
|
| + */
|
| + VAProcColorStandardType output_color_standard;
|
| + /**
|
| + * \brief Pipeline filters. See video pipeline flags.
|
| + *
|
| + * Flags to control the pipeline, like whether to apply subpictures
|
| + * or not, notify the driver that it can opt for power optimizations,
|
| + * should this be needed.
|
| + */
|
| + unsigned int pipeline_flags;
|
| + /**
|
| + * \brief Extra filter flags. See vaPutSurface() flags.
|
| + *
|
| + * Filter flags are used as a fast path, wherever possible, to use
|
| + * vaPutSurface() flags instead of explicit filter parameter buffers.
|
| + *
|
| + * Allowed filter flags API-wise. Use vaQueryVideoProcPipelineCaps()
|
| + * to check for implementation details:
|
| + * - Bob-deinterlacing: \c VA_FRAME_PICTURE, \c VA_TOP_FIELD,
|
| + * \c VA_BOTTOM_FIELD. Note that any deinterlacing filter
|
| + * (#VAProcFilterDeinterlacing) will override those flags.
|
| + * - Color space conversion: \c VA_SRC_BT601, \c VA_SRC_BT709,
|
| + * \c VA_SRC_SMPTE_240.
|
| + * - Scaling: \c VA_FILTER_SCALING_DEFAULT, \c VA_FILTER_SCALING_FAST,
|
| + * \c VA_FILTER_SCALING_HQ, \c VA_FILTER_SCALING_NL_ANAMORPHIC.
|
| + */
|
| + unsigned int filter_flags;
|
| + /**
|
| + * \brief Array of filters to apply to the surface.
|
| + *
|
| + * The list of filters shall be ordered in the same way the driver expects
|
| + * them. i.e. as was returned from vaQueryVideoProcFilters().
|
| + * Otherwise, a #VA_STATUS_ERROR_INVALID_FILTER_CHAIN is returned
|
| + * from vaRenderPicture() with this buffer.
|
| + *
|
| + * #VA_STATUS_ERROR_UNSUPPORTED_FILTER is returned if the list
|
| + * contains an unsupported filter.
|
| + *
|
| + * Note: no filter buffer is destroyed after a call to vaRenderPicture(),
|
| + * only this pipeline buffer will be destroyed as per the core API
|
| + * specification. This allows for flexibility in re-using the filter for
|
| + * other surfaces to be processed.
|
| + */
|
| + VABufferID *filters;
|
| + /** \brief Actual number of filters. */
|
| + unsigned int num_filters;
|
| + /** \brief Array of forward reference frames. */
|
| + VASurfaceID *forward_references;
|
| + /** \brief Number of forward reference frames that were supplied. */
|
| + unsigned int num_forward_references;
|
| + /** \brief Array of backward reference frames. */
|
| + VASurfaceID *backward_references;
|
| + /** \brief Number of backward reference frames that were supplied. */
|
| + unsigned int num_backward_references;
|
| +} VAProcPipelineParameterBuffer;
|
| +
|
| +/**
|
| + * \brief Filter parameter buffer base.
|
| + *
|
| + * This is a helper structure used by driver implementations only.
|
| + * Users are not supposed to allocate filter parameter buffers of this
|
| + * type.
|
| + */
|
| +typedef struct _VAProcFilterParameterBufferBase {
|
| + /** \brief Filter type. */
|
| + VAProcFilterType type;
|
| +} VAProcFilterParameterBufferBase;
|
| +
|
| +/**
|
| + * \brief Default filter parametrization.
|
| + *
|
| + * Unless there is a filter-specific parameter buffer,
|
| + * #VAProcFilterParameterBuffer is the default type to use.
|
| + */
|
| +typedef struct _VAProcFilterParameterBuffer {
|
| + /** \brief Filter type. */
|
| + VAProcFilterType type;
|
| + /** \brief Value. */
|
| + float value;
|
| +} VAProcFilterParameterBuffer;
|
| +
|
| +/** @name De-interlacing flags */
|
| +/**@{*/
|
| +/**
|
| + * \brief Bottom field first in the input frame.
|
| + * if this is not set then assumes top field first.
|
| + */
|
| +#define VA_DEINTERLACING_BOTTOM_FIELD_FIRST 0x0001
|
| +/**
|
| + * \brief Bottom field used in deinterlacing.
|
| + * if this is not set then assumes top field is used.
|
| + */
|
| +#define VA_DEINTERLACING_BOTTOM_FIELD 0x0002
|
| +/**
|
| + * \brief A single field is stored in the input frame.
|
| + * if this is not set then assumes the frame contains two interleaved fields.
|
| + */
|
| +#define VA_DEINTERLACING_ONE_FIELD 0x0004
|
| +/**@}*/
|
| +
|
| +/** \brief Deinterlacing filter parametrization. */
|
| +typedef struct _VAProcFilterParameterBufferDeinterlacing {
|
| + /** \brief Filter type. Shall be set to #VAProcFilterDeinterlacing. */
|
| + VAProcFilterType type;
|
| + /** \brief Deinterlacing algorithm. */
|
| + VAProcDeinterlacingType algorithm;
|
| + /** \brief Deinterlacing flags. */
|
| + unsigned int flags;
|
| +} VAProcFilterParameterBufferDeinterlacing;
|
| +
|
| +/**
|
| + * \brief Color balance filter parametrization.
|
| + *
|
| + * This buffer defines color balance attributes. A VA buffer can hold
|
| + * several color balance attributes by creating a VA buffer of desired
|
| + * number of elements. This can be achieved by the following pseudo-code:
|
| + *
|
| + * \code
|
| + * enum { kHue, kSaturation, kBrightness, kContrast };
|
| + *
|
| + * // Initial color balance parameters
|
| + * static const VAProcFilterParameterBufferColorBalance colorBalanceParams[4] =
|
| + * {
|
| + * [kHue] =
|
| + * { VAProcFilterColorBalance, VAProcColorBalanceHue, 0.5 },
|
| + * [kSaturation] =
|
| + * { VAProcFilterColorBalance, VAProcColorBalanceSaturation, 0.5 },
|
| + * [kBrightness] =
|
| + * { VAProcFilterColorBalance, VAProcColorBalanceBrightness, 0.5 },
|
| + * [kSaturation] =
|
| + * { VAProcFilterColorBalance, VAProcColorBalanceSaturation, 0.5 }
|
| + * };
|
| + *
|
| + * // Create buffer
|
| + * VABufferID colorBalanceBuffer;
|
| + * vaCreateBuffer(va_dpy, vpp_ctx,
|
| + * VAProcFilterParameterBufferType, sizeof(*pColorBalanceParam), 4,
|
| + * colorBalanceParams,
|
| + * &colorBalanceBuffer
|
| + * );
|
| + *
|
| + * VAProcFilterParameterBufferColorBalance *pColorBalanceParam;
|
| + * vaMapBuffer(va_dpy, colorBalanceBuffer, &pColorBalanceParam);
|
| + * {
|
| + * // Change brightness only
|
| + * pColorBalanceBuffer[kBrightness].value = 0.75;
|
| + * }
|
| + * vaUnmapBuffer(va_dpy, colorBalanceBuffer);
|
| + * \endcode
|
| + */
|
| +typedef struct _VAProcFilterParameterBufferColorBalance {
|
| + /** \brief Filter type. Shall be set to #VAProcFilterColorBalance. */
|
| + VAProcFilterType type;
|
| + /** \brief Color balance attribute. */
|
| + VAProcColorBalanceType attrib;
|
| + /**
|
| + * \brief Color balance value.
|
| + *
|
| + * Special case for automatically adjusted attributes. e.g.
|
| + * #VAProcColorBalanceAutoSaturation,
|
| + * #VAProcColorBalanceAutoBrightness,
|
| + * #VAProcColorBalanceAutoContrast.
|
| + * - If \ref value is \c 1.0 +/- \c FLT_EPSILON, the attribute is
|
| + * automatically adjusted and overrides any other attribute of
|
| + * the same type that would have been set explicitly;
|
| + * - If \ref value is \c 0.0 +/- \c FLT_EPSILON, the attribute is
|
| + * disabled and other attribute of the same type is used instead.
|
| + */
|
| + float value;
|
| +} VAProcFilterParameterBufferColorBalance;
|
| +
|
| +/**
|
| + * \brief Default filter cap specification (single range value).
|
| + *
|
| + * Unless there is a filter-specific cap structure, #VAProcFilterCap is the
|
| + * default type to use for output caps from vaQueryVideoProcFilterCaps().
|
| + */
|
| +typedef struct _VAProcFilterCap {
|
| + /** \brief Range of supported values for the filter. */
|
| + VAProcFilterValueRange range;
|
| +} VAProcFilterCap;
|
| +
|
| +/** \brief Capabilities specification for the deinterlacing filter. */
|
| +typedef struct _VAProcFilterCapDeinterlacing {
|
| + /** \brief Deinterlacing algorithm. */
|
| + VAProcDeinterlacingType type;
|
| +} VAProcFilterCapDeinterlacing;
|
| +
|
| +/** \brief Capabilities specification for the color balance filter. */
|
| +typedef struct _VAProcFilterCapColorBalance {
|
| + /** \brief Color balance operation. */
|
| + VAProcColorBalanceType type;
|
| + /** \brief Range of supported values for the specified operation. */
|
| + VAProcFilterValueRange range;
|
| +} VAProcFilterCapColorBalance;
|
| +
|
| +/**
|
| + * \brief Queries video processing filters.
|
| + *
|
| + * This function returns the list of video processing filters supported
|
| + * by the driver. The \c filters array is allocated by the user and
|
| + * \c num_filters shall be initialized to the number of allocated
|
| + * elements in that array. Upon successful return, the actual number
|
| + * of filters will be overwritten into \c num_filters. Otherwise,
|
| + * \c VA_STATUS_ERROR_MAX_NUM_EXCEEDED is returned and \c num_filters
|
| + * is adjusted to the number of elements that would be returned if enough
|
| + * space was available.
|
| + *
|
| + * The list of video processing filters supported by the driver shall
|
| + * be ordered in the way they can be iteratively applied. This is needed
|
| + * for both correctness, i.e. some filters would not mean anything if
|
| + * applied at the beginning of the pipeline; but also for performance
|
| + * since some filters can be applied in a single pass (e.g. noise
|
| + * reduction + deinterlacing).
|
| + *
|
| + * @param[in] dpy the VA display
|
| + * @param[in] context the video processing context
|
| + * @param[out] filters the output array of #VAProcFilterType elements
|
| + * @param[in,out] num_filters the number of elements allocated on input,
|
| + * the number of elements actually filled in on output
|
| + */
|
| +VAStatus
|
| +vaQueryVideoProcFilters(
|
| + VADisplay dpy,
|
| + VAContextID context,
|
| + VAProcFilterType *filters,
|
| + unsigned int *num_filters
|
| +);
|
| +
|
| +/**
|
| + * \brief Queries video filter capabilities.
|
| + *
|
| + * This function returns the list of capabilities supported by the driver
|
| + * for a specific video filter. The \c filter_caps array is allocated by
|
| + * the user and \c num_filter_caps shall be initialized to the number
|
| + * of allocated elements in that array. Upon successful return, the
|
| + * actual number of filters will be overwritten into \c num_filter_caps.
|
| + * Otherwise, \c VA_STATUS_ERROR_MAX_NUM_EXCEEDED is returned and
|
| + * \c num_filter_caps is adjusted to the number of elements that would be
|
| + * returned if enough space was available.
|
| + *
|
| + * @param[in] dpy the VA display
|
| + * @param[in] context the video processing context
|
| + * @param[in] type the video filter type
|
| + * @param[out] filter_caps the output array of #VAProcFilterCap elements
|
| + * @param[in,out] num_filter_caps the number of elements allocated on input,
|
| + * the number of elements actually filled in output
|
| + */
|
| +VAStatus
|
| +vaQueryVideoProcFilterCaps(
|
| + VADisplay dpy,
|
| + VAContextID context,
|
| + VAProcFilterType type,
|
| + void *filter_caps,
|
| + unsigned int *num_filter_caps
|
| +);
|
| +
|
| +/**
|
| + * \brief Queries video processing pipeline capabilities.
|
| + *
|
| + * This function returns the video processing pipeline capabilities. The
|
| + * \c filters array defines the video processing pipeline and is an array
|
| + * of buffers holding filter parameters.
|
| + *
|
| + * Note: the #VAProcPipelineCaps structure contains user-provided arrays.
|
| + * If non-NULL, the corresponding \c num_* fields shall be filled in on
|
| + * input with the number of elements allocated. Upon successful return,
|
| + * the actual number of elements will be overwritten into the \c num_*
|
| + * fields. Otherwise, \c VA_STATUS_ERROR_MAX_NUM_EXCEEDED is returned
|
| + * and \c num_* fields are adjusted to the number of elements that would
|
| + * be returned if enough space was available.
|
| + *
|
| + * @param[in] dpy the VA display
|
| + * @param[in] context the video processing context
|
| + * @param[in] filters the array of VA buffers defining the video
|
| + * processing pipeline
|
| + * @param[in] num_filters the number of elements in filters
|
| + * @param[in,out] pipeline_caps the video processing pipeline capabilities
|
| + */
|
| +VAStatus
|
| +vaQueryVideoProcPipelineCaps(
|
| + VADisplay dpy,
|
| + VAContextID context,
|
| + VABufferID *filters,
|
| + unsigned int num_filters,
|
| + VAProcPipelineCaps *pipeline_caps
|
| +);
|
| +
|
| +/**@}*/
|
| +
|
| +#ifdef __cplusplus
|
| +}
|
| +#endif
|
| +
|
| +#endif /* VA_VPP_H */
|
|
|