/* * Copyright (c) 2007-2011 Intel Corporation. All Rights Reserved. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the * "Software"), to deal in the Software without restriction, including * without limitation the rights to use, copy, modify, merge, publish, * distribute, sub license, and/or sell copies of the Software, and to * permit persons to whom the Software is furnished to do so, subject to * the following conditions: * * The above copyright notice and this permission notice (including the * next paragraph) shall be included in all copies or substantial portions * of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. * IN NO EVENT SHALL INTEL AND/OR ITS SUPPLIERS BE LIABLE FOR * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */ /** * \file va_vpp.h * \brief The video processing API * * This file contains the \ref api_vpp "Video processing API". */ #ifndef VA_VPP_H #define VA_VPP_H #ifdef __cplusplus extern "C" { #endif /** * \defgroup api_vpp Video processing API * * @{ * * The video processing API uses the same paradigm as for decoding: * - Query for supported filters; * - Set up a video processing pipeline; * - Send video processing parameters through VA buffers. * * \section api_vpp_caps Query for supported filters * * Checking whether video processing is supported can be performed * with vaQueryConfigEntrypoints() and the profile argument set to * #VAProfileNone. If video processing is supported, then the list of * returned entry-points will include #VAEntrypointVideoProc. * * \code * VAEntrypoint *entrypoints; * int i, num_entrypoints, supportsVideoProcessing = 0; * * num_entrypoints = vaMaxNumEntrypoints(); * entrypoints = malloc(num_entrypoints * sizeof(entrypoints[0]); * vaQueryConfigEntrypoints(va_dpy, VAProfileNone, * entrypoints, &num_entrypoints); * * for (i = 0; !supportsVideoProcessing && i < num_entrypoints; i++) { * if (entrypoints[i] == VAEntrypointVideoProc) * supportsVideoProcessing = 1; * } * \endcode * * Then, the vaQueryVideoProcFilters() function is used to query the * list of video processing filters. * * \code * VAProcFilterType filters[VAProcFilterCount]; * unsigned int num_filters = VAProcFilterCount; * * // num_filters shall be initialized to the length of the array * vaQueryVideoProcFilters(va_dpy, vpp_ctx, &filters, &num_filters); * \endcode * * Finally, individual filter capabilities can be checked with * vaQueryVideoProcFilterCaps(). * * \code * VAProcFilterCap denoise_caps; * unsigned int num_denoise_caps = 1; * vaQueryVideoProcFilterCaps(va_dpy, vpp_ctx, * VAProcFilterNoiseReduction, * &denoise_caps, &num_denoise_caps * ); * * VAProcFilterCapDeinterlacing deinterlacing_caps[VAProcDeinterlacingCount]; * unsigned int num_deinterlacing_caps = VAProcDeinterlacingCount; * vaQueryVideoProcFilterCaps(va_dpy, vpp_ctx, * VAProcFilterDeinterlacing, * &deinterlacing_caps, &num_deinterlacing_caps * ); * \endcode * * \section api_vpp_setup Set up a video processing pipeline * * A video processing pipeline buffer is created for each source * surface we want to process. However, buffers holding filter * parameters can be created once and for all. Rationale is to avoid * multiple creation/destruction chains of filter buffers and also * because filter parameters generally won't change frame after * frame. e.g. this makes it possible to implement a checkerboard of * videos where the same filters are applied to each video source. * * The general control flow is demonstrated by the following pseudo-code: * \code * // Create filters * VABufferID denoise_filter, deint_filter; * VABufferID filter_bufs[VAProcFilterCount]; * unsigned int num_filter_bufs; * * for (i = 0; i < num_filters; i++) { * switch (filters[i]) { * case VAProcFilterNoiseReduction: { // Noise reduction filter * VAProcFilterParameterBuffer denoise; * denoise.type = VAProcFilterNoiseReduction; * denoise.value = 0.5; * vaCreateBuffer(va_dpy, vpp_ctx, * VAProcFilterParameterBufferType, sizeof(denoise), 1, * &denoise, &denoise_filter * ); * filter_bufs[num_filter_bufs++] = denoise_filter; * break; * } * * case VAProcFilterDeinterlacing: // Motion-adaptive deinterlacing * for (j = 0; j < num_deinterlacing_caps; j++) { * VAProcFilterCapDeinterlacing * const cap = &deinterlacing_caps[j]; * if (cap->type != VAProcDeinterlacingMotionAdaptive) * continue; * * VAProcFilterParameterBufferDeinterlacing deint; * deint.type = VAProcFilterDeinterlacing; * deint.algorithm = VAProcDeinterlacingMotionAdaptive; * vaCreateBuffer(va_dpy, vpp_ctx, * VAProcFilterParameterBufferType, sizeof(deint), 1, * &deint, &deint_filter * ); * filter_bufs[num_filter_bufs++] = deint_filter; * } * } * } * \endcode * * Once the video processing pipeline is set up, the caller shall check the * implied capabilities and requirements with vaQueryVideoProcPipelineCaps(). * This function can be used to validate the number of reference frames are * needed by the specified deinterlacing algorithm, the supported color * primaries, etc. * \code * // Create filters * VAProcPipelineCaps pipeline_caps; * VASurfaceID *forward_references; * unsigned int num_forward_references; * VASurfaceID *backward_references; * unsigned int num_backward_references; * VAProcColorStandardType in_color_standards[VAProcColorStandardCount]; * VAProcColorStandardType out_color_standards[VAProcColorStandardCount]; * * pipeline_caps.input_color_standards = NULL; * pipeline_caps.num_input_color_standards = ARRAY_ELEMS(in_color_standards); * pipeline_caps.output_color_standards = NULL; * pipeline_caps.num_output_color_standards = ARRAY_ELEMS(out_color_standards); * vaQueryVideoProcPipelineCaps(va_dpy, vpp_ctx, * filter_bufs, num_filter_bufs, * &pipeline_caps * ); * * num_forward_references = pipeline_caps.num_forward_references; * forward_references = * malloc(num__forward_references * sizeof(VASurfaceID)); * num_backward_references = pipeline_caps.num_backward_references; * backward_references = * malloc(num_backward_references * sizeof(VASurfaceID)); * \endcode * * \section api_vpp_submit Send video processing parameters through VA buffers * * Video processing pipeline parameters are submitted for each source * surface to process. Video filter parameters can also change, per-surface. * e.g. the list of reference frames used for deinterlacing. * * \code * foreach (iteration) { * vaBeginPicture(va_dpy, vpp_ctx, vpp_surface); * foreach (surface) { * VARectangle output_region; * VABufferID pipeline_buf; * VAProcPipelineParameterBuffer *pipeline_param; * * vaCreateBuffer(va_dpy, vpp_ctx, * VAProcPipelineParameterBuffer, sizeof(*pipeline_param), 1, * NULL, &pipeline_buf * ); * * // Setup output region for this surface * // e.g. upper left corner for the first surface * output_region.x = BORDER; * output_region.y = BORDER; * output_region.width = * (vpp_surface_width - (Nx_surfaces + 1) * BORDER) / Nx_surfaces; * output_region.height = * (vpp_surface_height - (Ny_surfaces + 1) * BORDER) / Ny_surfaces; * * vaMapBuffer(va_dpy, pipeline_buf, &pipeline_param); * pipeline_param->surface = surface; * pipeline_param->surface_region = NULL; * pipeline_param->output_region = &output_region; * pipeline_param->output_background_color = 0; * if (first surface to render) * pipeline_param->output_background_color = 0xff000000; // black * pipeline_param->filter_flags = VA_FILTER_SCALING_HQ; * pipeline_param->filters = filter_bufs; * pipeline_param->num_filters = num_filter_bufs; * vaUnmapBuffer(va_dpy, pipeline_buf); * * // Update reference frames for deinterlacing, if necessary * pipeline_param->forward_references = forward_references; * pipeline_param->num_forward_references = num_forward_references_used; * pipeline_param->backward_references = backward_references; * pipeline_param->num_backward_references = num_bacward_references_used; * * // Apply filters * vaRenderPicture(va_dpy, vpp_ctx, &pipeline_buf, 1); * } * vaEndPicture(va_dpy, vpp_ctx); * } * \endcode */ /** \brief Video filter types. */ typedef enum _VAProcFilterType { VAProcFilterNone = 0, /** \brief Noise reduction filter. */ VAProcFilterNoiseReduction, /** \brief Deinterlacing filter. */ VAProcFilterDeinterlacing, /** \brief Sharpening filter. */ VAProcFilterSharpening, /** \brief Color balance parameters. */ VAProcFilterColorBalance, /** \brief Skin Tone Enhancement. */ VAProcFilterSkinToneEnhancement, VAProcFilterCount } VAProcFilterType; /** \brief Deinterlacing types. */ typedef enum _VAProcDeinterlacingType { VAProcDeinterlacingNone = 0, /** \brief Bob deinterlacing algorithm. */ VAProcDeinterlacingBob, /** \brief Weave deinterlacing algorithm. */ VAProcDeinterlacingWeave, /** \brief Motion adaptive deinterlacing algorithm. */ VAProcDeinterlacingMotionAdaptive, /** \brief Motion compensated deinterlacing algorithm. */ VAProcDeinterlacingMotionCompensated, /** \brief Number of deinterlacing algorithms. */ VAProcDeinterlacingCount } VAProcDeinterlacingType; /** \brief Color balance types. */ typedef enum _VAProcColorBalanceType { VAProcColorBalanceNone = 0, /** \brief Hue. */ VAProcColorBalanceHue, /** \brief Saturation. */ VAProcColorBalanceSaturation, /** \brief Brightness. */ VAProcColorBalanceBrightness, /** \brief Contrast. */ VAProcColorBalanceContrast, /** \brief Automatically adjusted saturation. */ VAProcColorBalanceAutoSaturation, /** \brief Automatically adjusted brightness. */ VAProcColorBalanceAutoBrightness, /** \brief Automatically adjusted contrast. */ VAProcColorBalanceAutoContrast, /** \brief Number of color balance attributes. */ VAProcColorBalanceCount } VAProcColorBalanceType; /** \brief Color standard types. */ typedef enum _VAProcColorStandardType { VAProcColorStandardNone = 0, /** \brief ITU-R BT.601. */ VAProcColorStandardBT601, /** \brief ITU-R BT.709. */ VAProcColorStandardBT709, /** \brief ITU-R BT.470-2 System M. */ VAProcColorStandardBT470M, /** \brief ITU-R BT.470-2 System B, G. */ VAProcColorStandardBT470BG, /** \brief SMPTE-170M. */ VAProcColorStandardSMPTE170M, /** \brief SMPTE-240M. */ VAProcColorStandardSMPTE240M, /** \brief Generic film. */ VAProcColorStandardGenericFilm, /** \brief Number of color standards. */ VAProcColorStandardCount } VAProcColorStandardType; /** @name Video pipeline flags */ /**@{*/ /** \brief Specifies whether to apply subpictures when processing a surface. */ #define VA_PROC_PIPELINE_SUBPICTURES 0x00000001 /** * \brief Specifies whether to apply power or performance * optimizations to a pipeline. * * When processing several surfaces, it may be necessary to prioritize * more certain pipelines than others. This flag is only a hint to the * video processor so that it can omit certain filters to save power * for example. Typically, this flag could be used with video surfaces * decoded from a secondary bitstream. */ #define VA_PROC_PIPELINE_FAST 0x00000002 /**@}*/ /** @name Video filter flags */ /**@{*/ /** \brief Specifies whether the filter shall be present in the pipeline. */ #define VA_PROC_FILTER_MANDATORY 0x00000001 /**@}*/ /** @name Pipeline end flags */ /**@{*/ /** \brief Specifies the pipeline is the last. */ #define VA_PIPELINE_FLAG_END 0x00000004 /**@}*/ /** \brief Video processing pipeline capabilities. */ typedef struct _VAProcPipelineCaps { /** \brief Pipeline flags. See VAProcPipelineParameterBuffer::pipeline_flags. */ unsigned int pipeline_flags; /** \brief Extra filter flags. See VAProcPipelineParameterBuffer::filter_flags. */ unsigned int filter_flags; /** \brief Number of forward reference frames that are needed. */ unsigned int num_forward_references; /** \brief Number of backward reference frames that are needed. */ unsigned int num_backward_references; /** \brief List of color standards supported on input. */ VAProcColorStandardType *input_color_standards; /** \brief Number of elements in \ref input_color_standards array. */ unsigned int num_input_color_standards; /** \brief List of color standards supported on output. */ VAProcColorStandardType *output_color_standards; /** \brief Number of elements in \ref output_color_standards array. */ unsigned int num_output_color_standards; } VAProcPipelineCaps; /** \brief Specification of values supported by the filter. */ typedef struct _VAProcFilterValueRange { /** \brief Minimum value supported, inclusive. */ float min_value; /** \brief Maximum value supported, inclusive. */ float max_value; /** \brief Default value. */ float default_value; /** \brief Step value that alters the filter behaviour in a sensible way. */ float step; } VAProcFilterValueRange; /** * \brief Video processing pipeline configuration. * * This buffer defines a video processing pipeline. As for any buffer * passed to \c vaRenderPicture(), this is a one-time usage model. * However, the actual filters to be applied are provided in the * \c filters field, so they can be re-used in other processing * pipelines. * * The target surface is specified by the \c render_target argument of * \c vaBeginPicture(). The general usage model is described as follows: * - \c vaBeginPicture(): specify the target surface that receives the * processed output; * - \c vaRenderPicture(): specify a surface to be processed and composed * into the \c render_target. Use as many \c vaRenderPicture() calls as * necessary surfaces to compose ; * - \c vaEndPicture(): tell the driver to start processing the surfaces * with the requested filters. * * If a filter (e.g. noise reduction) needs to be applied with different * values for multiple surfaces, the application needs to create as many * filter parameter buffers as necessary. i.e. the filter parameters shall * not change between two calls to \c vaRenderPicture(). * * For composition usage models, the first surface to process will generally * use an opaque background color, i.e. \c output_background_color set with * the most significant byte set to \c 0xff. For instance, \c 0xff000000 for * a black background. Then, subsequent surfaces would use a transparent * background color. */ typedef struct _VAProcPipelineParameterBuffer { /** * \brief Source surface ID. * * ID of the source surface to process. If subpictures are associated * with the video surfaces then they shall be rendered to the target * surface, if the #VA_PROC_PIPELINE_SUBPICTURES pipeline flag is set. */ VASurfaceID surface; /** * \brief Region within the source surface to be processed. * * Pointer to a #VARectangle defining the region within the source * surface to be processed. If NULL, \c surface_region implies the * whole surface. */ const VARectangle *surface_region; /** * \brief Requested input color primaries. * * Color primaries are implicitly converted throughout the processing * pipeline. The video processor chooses the best moment to apply * this conversion. The set of supported color primaries primaries * for input shall be queried with vaQueryVideoProcPipelineCaps(). */ VAProcColorStandardType surface_color_standard; /** * \brief Region within the output surface. * * Pointer to a #VARectangle defining the region within the output * surface that receives the processed pixels. If NULL, \c output_region * implies the whole surface. * * Note that any pixels residing outside the specified region will * be filled in with the \ref output_background_color. */ const VARectangle *output_region; /** * \brief Background color. * * Background color used to fill in pixels that reside outside of the * specified \ref output_region. The color is specified in ARGB format: * [31:24] alpha, [23:16] red, [15:8] green, [7:0] blue. * * Unless the alpha value is zero or the \ref output_region represents * the whole target surface size, implementations shall not render the * source surface to the target surface directly. Rather, in order to * maintain the exact semantics of \ref output_background_color, the * driver shall use a temporary surface and fill it in with the * appropriate background color. Next, the driver will blend this * temporary surface into the target surface. */ unsigned int output_background_color; /** * \brief Requested output color primaries. */ VAProcColorStandardType output_color_standard; /** * \brief Pipeline filters. See video pipeline flags. * * Flags to control the pipeline, like whether to apply subpictures * or not, notify the driver that it can opt for power optimizations, * should this be needed. */ unsigned int pipeline_flags; /** * \brief Extra filter flags. See vaPutSurface() flags. * * Filter flags are used as a fast path, wherever possible, to use * vaPutSurface() flags instead of explicit filter parameter buffers. * * Allowed filter flags API-wise. Use vaQueryVideoProcPipelineCaps() * to check for implementation details: * - Bob-deinterlacing: \c VA_FRAME_PICTURE, \c VA_TOP_FIELD, * \c VA_BOTTOM_FIELD. Note that any deinterlacing filter * (#VAProcFilterDeinterlacing) will override those flags. * - Color space conversion: \c VA_SRC_BT601, \c VA_SRC_BT709, * \c VA_SRC_SMPTE_240. * - Scaling: \c VA_FILTER_SCALING_DEFAULT, \c VA_FILTER_SCALING_FAST, * \c VA_FILTER_SCALING_HQ, \c VA_FILTER_SCALING_NL_ANAMORPHIC. */ unsigned int filter_flags; /** * \brief Array of filters to apply to the surface. * * The list of filters shall be ordered in the same way the driver expects * them. i.e. as was returned from vaQueryVideoProcFilters(). * Otherwise, a #VA_STATUS_ERROR_INVALID_FILTER_CHAIN is returned * from vaRenderPicture() with this buffer. * * #VA_STATUS_ERROR_UNSUPPORTED_FILTER is returned if the list * contains an unsupported filter. * * Note: no filter buffer is destroyed after a call to vaRenderPicture(), * only this pipeline buffer will be destroyed as per the core API * specification. This allows for flexibility in re-using the filter for * other surfaces to be processed. */ VABufferID *filters; /** \brief Actual number of filters. */ unsigned int num_filters; /** \brief Array of forward reference frames. */ VASurfaceID *forward_references; /** \brief Number of forward reference frames that were supplied. */ unsigned int num_forward_references; /** \brief Array of backward reference frames. */ VASurfaceID *backward_references; /** \brief Number of backward reference frames that were supplied. */ unsigned int num_backward_references; } VAProcPipelineParameterBuffer; /** * \brief Filter parameter buffer base. * * This is a helper structure used by driver implementations only. * Users are not supposed to allocate filter parameter buffers of this * type. */ typedef struct _VAProcFilterParameterBufferBase { /** \brief Filter type. */ VAProcFilterType type; } VAProcFilterParameterBufferBase; /** * \brief Default filter parametrization. * * Unless there is a filter-specific parameter buffer, * #VAProcFilterParameterBuffer is the default type to use. */ typedef struct _VAProcFilterParameterBuffer { /** \brief Filter type. */ VAProcFilterType type; /** \brief Value. */ float value; } VAProcFilterParameterBuffer; /** @name De-interlacing flags */ /**@{*/ /** * \brief Bottom field first in the input frame. * if this is not set then assumes top field first. */ #define VA_DEINTERLACING_BOTTOM_FIELD_FIRST 0x0001 /** * \brief Bottom field used in deinterlacing. * if this is not set then assumes top field is used. */ #define VA_DEINTERLACING_BOTTOM_FIELD 0x0002 /** * \brief A single field is stored in the input frame. * if this is not set then assumes the frame contains two interleaved fields. */ #define VA_DEINTERLACING_ONE_FIELD 0x0004 /**@}*/ /** \brief Deinterlacing filter parametrization. */ typedef struct _VAProcFilterParameterBufferDeinterlacing { /** \brief Filter type. Shall be set to #VAProcFilterDeinterlacing. */ VAProcFilterType type; /** \brief Deinterlacing algorithm. */ VAProcDeinterlacingType algorithm; /** \brief Deinterlacing flags. */ unsigned int flags; } VAProcFilterParameterBufferDeinterlacing; /** * \brief Color balance filter parametrization. * * This buffer defines color balance attributes. A VA buffer can hold * several color balance attributes by creating a VA buffer of desired * number of elements. This can be achieved by the following pseudo-code: * * \code * enum { kHue, kSaturation, kBrightness, kContrast }; * * // Initial color balance parameters * static const VAProcFilterParameterBufferColorBalance colorBalanceParams[4] = * { * [kHue] = * { VAProcFilterColorBalance, VAProcColorBalanceHue, 0.5 }, * [kSaturation] = * { VAProcFilterColorBalance, VAProcColorBalanceSaturation, 0.5 }, * [kBrightness] = * { VAProcFilterColorBalance, VAProcColorBalanceBrightness, 0.5 }, * [kSaturation] = * { VAProcFilterColorBalance, VAProcColorBalanceSaturation, 0.5 } * }; * * // Create buffer * VABufferID colorBalanceBuffer; * vaCreateBuffer(va_dpy, vpp_ctx, * VAProcFilterParameterBufferType, sizeof(*pColorBalanceParam), 4, * colorBalanceParams, * &colorBalanceBuffer * ); * * VAProcFilterParameterBufferColorBalance *pColorBalanceParam; * vaMapBuffer(va_dpy, colorBalanceBuffer, &pColorBalanceParam); * { * // Change brightness only * pColorBalanceBuffer[kBrightness].value = 0.75; * } * vaUnmapBuffer(va_dpy, colorBalanceBuffer); * \endcode */ typedef struct _VAProcFilterParameterBufferColorBalance { /** \brief Filter type. Shall be set to #VAProcFilterColorBalance. */ VAProcFilterType type; /** \brief Color balance attribute. */ VAProcColorBalanceType attrib; /** * \brief Color balance value. * * Special case for automatically adjusted attributes. e.g. * #VAProcColorBalanceAutoSaturation, * #VAProcColorBalanceAutoBrightness, * #VAProcColorBalanceAutoContrast. * - If \ref value is \c 1.0 +/- \c FLT_EPSILON, the attribute is * automatically adjusted and overrides any other attribute of * the same type that would have been set explicitly; * - If \ref value is \c 0.0 +/- \c FLT_EPSILON, the attribute is * disabled and other attribute of the same type is used instead. */ float value; } VAProcFilterParameterBufferColorBalance; /** * \brief Default filter cap specification (single range value). * * Unless there is a filter-specific cap structure, #VAProcFilterCap is the * default type to use for output caps from vaQueryVideoProcFilterCaps(). */ typedef struct _VAProcFilterCap { /** \brief Range of supported values for the filter. */ VAProcFilterValueRange range; } VAProcFilterCap; /** \brief Capabilities specification for the deinterlacing filter. */ typedef struct _VAProcFilterCapDeinterlacing { /** \brief Deinterlacing algorithm. */ VAProcDeinterlacingType type; } VAProcFilterCapDeinterlacing; /** \brief Capabilities specification for the color balance filter. */ typedef struct _VAProcFilterCapColorBalance { /** \brief Color balance operation. */ VAProcColorBalanceType type; /** \brief Range of supported values for the specified operation. */ VAProcFilterValueRange range; } VAProcFilterCapColorBalance; /** * \brief Queries video processing filters. * * This function returns the list of video processing filters supported * by the driver. The \c filters array is allocated by the user and * \c num_filters shall be initialized to the number of allocated * elements in that array. Upon successful return, the actual number * of filters will be overwritten into \c num_filters. Otherwise, * \c VA_STATUS_ERROR_MAX_NUM_EXCEEDED is returned and \c num_filters * is adjusted to the number of elements that would be returned if enough * space was available. * * The list of video processing filters supported by the driver shall * be ordered in the way they can be iteratively applied. This is needed * for both correctness, i.e. some filters would not mean anything if * applied at the beginning of the pipeline; but also for performance * since some filters can be applied in a single pass (e.g. noise * reduction + deinterlacing). * * @param[in] dpy the VA display * @param[in] context the video processing context * @param[out] filters the output array of #VAProcFilterType elements * @param[in,out] num_filters the number of elements allocated on input, * the number of elements actually filled in on output */ VAStatus vaQueryVideoProcFilters( VADisplay dpy, VAContextID context, VAProcFilterType *filters, unsigned int *num_filters ); /** * \brief Queries video filter capabilities. * * This function returns the list of capabilities supported by the driver * for a specific video filter. The \c filter_caps array is allocated by * the user and \c num_filter_caps shall be initialized to the number * of allocated elements in that array. Upon successful return, the * actual number of filters will be overwritten into \c num_filter_caps. * Otherwise, \c VA_STATUS_ERROR_MAX_NUM_EXCEEDED is returned and * \c num_filter_caps is adjusted to the number of elements that would be * returned if enough space was available. * * @param[in] dpy the VA display * @param[in] context the video processing context * @param[in] type the video filter type * @param[out] filter_caps the output array of #VAProcFilterCap elements * @param[in,out] num_filter_caps the number of elements allocated on input, * the number of elements actually filled in output */ VAStatus vaQueryVideoProcFilterCaps( VADisplay dpy, VAContextID context, VAProcFilterType type, void *filter_caps, unsigned int *num_filter_caps ); /** * \brief Queries video processing pipeline capabilities. * * This function returns the video processing pipeline capabilities. The * \c filters array defines the video processing pipeline and is an array * of buffers holding filter parameters. * * Note: the #VAProcPipelineCaps structure contains user-provided arrays. * If non-NULL, the corresponding \c num_* fields shall be filled in on * input with the number of elements allocated. Upon successful return, * the actual number of elements will be overwritten into the \c num_* * fields. Otherwise, \c VA_STATUS_ERROR_MAX_NUM_EXCEEDED is returned * and \c num_* fields are adjusted to the number of elements that would * be returned if enough space was available. * * @param[in] dpy the VA display * @param[in] context the video processing context * @param[in] filters the array of VA buffers defining the video * processing pipeline * @param[in] num_filters the number of elements in filters * @param[in,out] pipeline_caps the video processing pipeline capabilities */ VAStatus vaQueryVideoProcPipelineCaps( VADisplay dpy, VAContextID context, VABufferID *filters, unsigned int num_filters, VAProcPipelineCaps *pipeline_caps ); /**@}*/ #ifdef __cplusplus } #endif #endif /* VA_VPP_H */