diff mbox

[FFmpeg-devel,v4] avfilter/vaapi: add overlay_vaapi filter

Message ID 20190711092935.8581-1-zachary.zhou@intel.com
State Superseded
Headers show

Commit Message

Zachary Zhou July 11, 2019, 9:29 a.m. UTC
---
 configure                      |   1 +
 libavfilter/Makefile           |   1 +
 libavfilter/allfilters.c       |   1 +
 libavfilter/vf_overlay_vaapi.c | 424 +++++++++++++++++++++++++++++++++
 4 files changed, 427 insertions(+)
 create mode 100644 libavfilter/vf_overlay_vaapi.c

Comments

Zachary Zhou Aug. 19, 2019, 1:17 a.m. UTC | #1
Hi Mark,

Do you have any comments on this patch ?

Thanks,

Zachary


On 7/11/19 5:29 PM, Zachary Zhou wrote:
> ---
>   configure                      |   1 +
>   libavfilter/Makefile           |   1 +
>   libavfilter/allfilters.c       |   1 +
>   libavfilter/vf_overlay_vaapi.c | 424 +++++++++++++++++++++++++++++++++
>   4 files changed, 427 insertions(+)
>   create mode 100644 libavfilter/vf_overlay_vaapi.c
>
> diff --git a/configure b/configure
> index 32fc26356c..f469e6a3b1 100755
> --- a/configure
> +++ b/configure
> @@ -3478,6 +3478,7 @@ openclsrc_filter_deps="opencl"
>   overlay_opencl_filter_deps="opencl"
>   overlay_qsv_filter_deps="libmfx"
>   overlay_qsv_filter_select="qsvvpp"
> +overlay_vaapi_filter_deps="vaapi"
>   owdenoise_filter_deps="gpl"
>   pan_filter_deps="swresample"
>   perspective_filter_deps="gpl"
> diff --git a/libavfilter/Makefile b/libavfilter/Makefile
> index 07ea8d7edc..ccaad0d6a4 100644
> --- a/libavfilter/Makefile
> +++ b/libavfilter/Makefile
> @@ -311,6 +311,7 @@ OBJS-$(CONFIG_OVERLAY_FILTER)                += vf_overlay.o framesync.o
>   OBJS-$(CONFIG_OVERLAY_OPENCL_FILTER)         += vf_overlay_opencl.o opencl.o \
>                                                   opencl/overlay.o framesync.o
>   OBJS-$(CONFIG_OVERLAY_QSV_FILTER)            += vf_overlay_qsv.o framesync.o
> +OBJS-$(CONFIG_OVERLAY_VAAPI_FILTER)          += vf_overlay_vaapi.o framesync.o vaapi_vpp.o
>   OBJS-$(CONFIG_OWDENOISE_FILTER)              += vf_owdenoise.o
>   OBJS-$(CONFIG_PAD_FILTER)                    += vf_pad.o
>   OBJS-$(CONFIG_PALETTEGEN_FILTER)             += vf_palettegen.o
> diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c
> index 9c846b1ddd..27ee1df78b 100644
> --- a/libavfilter/allfilters.c
> +++ b/libavfilter/allfilters.c
> @@ -295,6 +295,7 @@ extern AVFilter ff_vf_oscilloscope;
>   extern AVFilter ff_vf_overlay;
>   extern AVFilter ff_vf_overlay_opencl;
>   extern AVFilter ff_vf_overlay_qsv;
> +extern AVFilter ff_vf_overlay_vaapi;
>   extern AVFilter ff_vf_owdenoise;
>   extern AVFilter ff_vf_pad;
>   extern AVFilter ff_vf_palettegen;
> diff --git a/libavfilter/vf_overlay_vaapi.c b/libavfilter/vf_overlay_vaapi.c
> new file mode 100644
> index 0000000000..9fffa0fcb9
> --- /dev/null
> +++ b/libavfilter/vf_overlay_vaapi.c
> @@ -0,0 +1,424 @@
> +/*
> + * This file is part of FFmpeg.
> + *
> + * FFmpeg is free software; you can redistribute it and/or
> + * modify it under the terms of the GNU Lesser General Public
> + * License as published by the Free Software Foundation; either
> + * version 2.1 of the License, or (at your option) any later version.
> + *
> + * FFmpeg is distributed in the hope that it will be useful,
> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> + * Lesser General Public License for more details.
> + *
> + * You should have received a copy of the GNU Lesser General Public
> + * License along with FFmpeg; if not, write to the Free Software
> + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
> + */
> +#include <string.h>
> +
> +#include "libavutil/avassert.h"
> +#include "libavutil/mem.h"
> +#include "libavutil/opt.h"
> +#include "libavutil/pixdesc.h"
> +
> +#include "avfilter.h"
> +#include "framesync.h"
> +#include "formats.h"
> +#include "internal.h"
> +#include "vaapi_vpp.h"
> +
> +typedef struct OverlayVAAPIContext {
> +    VAAPIVPPContext  vpp_ctx; // must be the first field
> +    FFFrameSync      fs;
> +    int              overlay_x;
> +    int              overlay_y;
> +    int              overlay_w;
> +    int              overlay_h;
> +    float            overlay_alpha;
> +} OverlayVAAPIContext;
> +
> +static int overlay_vaapi_query_formats(AVFilterContext *ctx)
> +{
> +    int i;
> +    int ret;
> +
> +    static const enum AVPixelFormat main_in_fmts[] = {
> +        AV_PIX_FMT_VAAPI,
> +        AV_PIX_FMT_NONE
> +    };
> +    static const enum AVPixelFormat out_pix_fmts[] = {
> +        AV_PIX_FMT_VAAPI,
> +        AV_PIX_FMT_NONE
> +    };
> +
> +    for (i = 0; i < ctx->nb_inputs; i++) {
> +        ret = ff_formats_ref(ff_make_format_list(main_in_fmts), &ctx->inputs[i]->out_formats);
> +        if (ret < 0)
> +            return ret;
> +    }
> +
> +    ret = ff_formats_ref(ff_make_format_list(out_pix_fmts), &ctx->outputs[0]->in_formats);
> +    if (ret < 0)
> +        return ret;
> +
> +    return 0;
> +}
> +
> +static int overlay_vaapi_render(AVFilterContext *avctx,
> +                                VAProcPipelineParameterBuffer *params,
> +                                VAProcPipelineParameterBuffer *subpic_params,
> +                                VASurfaceID output_surface)
> +{
> +    VABufferID params_id;
> +    VABufferID subpic_params_id;
> +    VAStatus vas;
> +    int err = 0;
> +    VAAPIVPPContext *ctx   = avctx->priv;
> +
> +    vas = vaBeginPicture(ctx->hwctx->display,
> +                         ctx->va_context, output_surface);
> +    if (vas != VA_STATUS_SUCCESS) {
> +        av_log(avctx, AV_LOG_ERROR, "Failed to attach new picture: "
> +               "%d (%s).\n", vas, vaErrorStr(vas));
> +        err = AVERROR(EIO);
> +        goto fail;
> +    }
> +
> +    vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context,
> +                         VAProcPipelineParameterBufferType,
> +                         sizeof(*params), 1, params, &params_id);
> +    if (vas != VA_STATUS_SUCCESS) {
> +        av_log(avctx, AV_LOG_ERROR, "Failed to create parameter buffer: "
> +               "%d (%s).\n", vas, vaErrorStr(vas));
> +        err = AVERROR(EIO);
> +        goto fail_after_begin;
> +    }
> +    av_log(avctx, AV_LOG_DEBUG, "Pipeline parameter buffer is %#x.\n",
> +           params_id);
> +
> +
> +    vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context,
> +                         VAProcPipelineParameterBufferType,
> +                         sizeof(*subpic_params), 1, subpic_params, &subpic_params_id);
> +    if (vas != VA_STATUS_SUCCESS) {
> +        av_log(avctx, AV_LOG_ERROR, "Failed to create parameter buffer: "
> +               "%d (%s).\n", vas, vaErrorStr(vas));
> +        err = AVERROR(EIO);
> +        goto fail_after_begin;
> +    }
> +    av_log(avctx, AV_LOG_DEBUG, "Pipeline subpic parameter buffer is %#x.\n",
> +           subpic_params_id);
> +
> +    vas = vaRenderPicture(ctx->hwctx->display, ctx->va_context,
> +                          &params_id, 1);
> +    if (vas != VA_STATUS_SUCCESS) {
> +        av_log(avctx, AV_LOG_ERROR, "Failed to render parameter buffer: "
> +               "%d (%s).\n", vas, vaErrorStr(vas));
> +        err = AVERROR(EIO);
> +        goto fail_after_begin;
> +    }
> +
> +    vas = vaRenderPicture(ctx->hwctx->display, ctx->va_context,
> +                          &subpic_params_id, 1);
> +    if (vas != VA_STATUS_SUCCESS) {
> +        av_log(avctx, AV_LOG_ERROR, "Failed to render subpic parameter buffer: "
> +               "%d (%s).\n", vas, vaErrorStr(vas));
> +        err = AVERROR(EIO);
> +        goto fail_after_begin;
> +    }
> +
> +    vas = vaEndPicture(ctx->hwctx->display, ctx->va_context);
> +    if (vas != VA_STATUS_SUCCESS) {
> +        av_log(avctx, AV_LOG_ERROR, "Failed to start picture processing: "
> +               "%d (%s).\n", vas, vaErrorStr(vas));
> +        err = AVERROR(EIO);
> +        goto fail_after_render;
> +    }
> +
> +    if (CONFIG_VAAPI_1 || ctx->hwctx->driver_quirks &
> +        AV_VAAPI_DRIVER_QUIRK_RENDER_PARAM_BUFFERS) {
> +        vas = vaDestroyBuffer(ctx->hwctx->display, params_id);
> +        if (vas != VA_STATUS_SUCCESS) {
> +            av_log(avctx, AV_LOG_ERROR, "Failed to free parameter buffer: "
> +                   "%d (%s).\n", vas, vaErrorStr(vas));
> +            // And ignore.
> +        }
> +    }
> +
> +    return 0;
> +
> +    // We want to make sure that if vaBeginPicture has been called, we also
> +    // call vaRenderPicture and vaEndPicture.  These calls may well fail or
> +    // do something else nasty, but once we're in this failure case there
> +    // isn't much else we can do.
> +fail_after_begin:
> +    vaRenderPicture(ctx->hwctx->display, ctx->va_context, &params_id, 1);
> +fail_after_render:
> +    vaEndPicture(ctx->hwctx->display, ctx->va_context);
> +fail:
> +    return err;
> +}
> +
> +static int overlay_vaapi_blend(FFFrameSync *fs)
> +{
> +    AVFilterContext    *avctx = fs->parent;
> +    AVFilterLink     *outlink = avctx->outputs[0];
> +    OverlayVAAPIContext *ctx  = avctx->priv;
> +    VAAPIVPPContext *vpp_ctx  = avctx->priv;
> +    AVFrame *input_main, *input_overlay;
> +    AVFrame *output;
> +    int err;
> +    VASurfaceID main_surface, overlay_surface, output_surface;
> +    VARectangle overlay_region, output_region;
> +    VAProcPipelineParameterBuffer params, subpic_params;
> +
> +#if VA_CHECK_VERSION(2, 1, 0)
> +    VABlendState blend_state; // Blend State
> +#endif
> +
> +    err = ff_framesync_get_frame(fs, 0, &input_main, 0);
> +    if (err < 0)
> +        return err;
> +    err = ff_framesync_get_frame(fs, 1, &input_overlay, 0);
> +    if (err < 0)
> +        return err;
> +
> +    av_log(avctx, AV_LOG_DEBUG, "Filter main: %s, %ux%u (%"PRId64").\n",
> +           av_get_pix_fmt_name(input_main->format),
> +           input_main->width, input_main->height, input_main->pts);
> +
> +    av_log(avctx, AV_LOG_DEBUG, "Filter overlay: %s, %ux%u (%"PRId64").\n",
> +           av_get_pix_fmt_name(input_overlay->format),
> +           input_overlay->width, input_overlay->height, input_overlay->pts);
> +
> +    if (vpp_ctx->va_context == VA_INVALID_ID)
> +        return AVERROR(EINVAL);
> +
> +    main_surface = (VASurfaceID)(uintptr_t)input_main->data[3];
> +    av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp main.\n",
> +           main_surface);
> +
> +    overlay_surface = (VASurfaceID)(uintptr_t)input_overlay->data[3];
> +    av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp overlay.\n",
> +           overlay_surface);
> +
> +    output = ff_get_video_buffer(outlink, outlink->w, outlink->h);
> +    if (!output) {
> +        err = AVERROR(ENOMEM);
> +        goto fail;
> +    }
> +
> +    output_surface = (VASurfaceID)(uintptr_t)output->data[3];
> +    av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp output.\n",
> +           output_surface);
> +
> +    memset(&params, 0, sizeof(params));
> +    memset(&subpic_params, 0, sizeof(subpic_params));
> +
> +    overlay_region = (VARectangle) {
> +        .x      = ctx->overlay_x,
> +        .y      = ctx->overlay_y,
> +        .width  = input_overlay->width,
> +        .height = input_overlay->height,
> +    };
> +
> +    output_region = (VARectangle) {
> +        .x      = 0,
> +        .y      = 0,
> +        .width  = output->width,
> +        .height = output->height,
> +    };
> +
> +    if (ctx->overlay_w && ctx->overlay_h) {
> +        overlay_region.width = ctx->overlay_w;
> +        overlay_region.height = ctx->overlay_h;
> +    }
> +
> +    if (overlay_region.width > input_main->width ||
> +        overlay_region.height > input_main->height) {
> +        av_log(ctx, AV_LOG_ERROR,
> +               "Overlay width > main width or overlay height > main height is not supported.\n");
> +        err =  AVERROR(EINVAL);
> +        goto fail;
> +    }
> +
> +    params.filters     = &vpp_ctx->filter_buffers[0];
> +    params.num_filters = vpp_ctx->nb_filter_buffers;
> +
> +    err = ff_vaapi_vpp_init_params(avctx, &params,
> +                                   input_main, output);
> +    if (err < 0)
> +        goto fail;
> +
> +    params.output_region = &output_region;
> +    params.output_background_color = VAAPI_VPP_BACKGROUND_BLACK;
> +    params.output_color_standard = params.surface_color_standard;
> +
> +    memcpy(&subpic_params, &params, sizeof(subpic_params));
> +
> +#if VA_CHECK_VERSION(2, 1, 0)
> +    blend_state.flag = VA_BLEND_GLOBAL_ALPHA;
> +    blend_state.global_alpha = ctx->overlay_alpha;
> +    subpic_params.blend_state = &blend_state;
> +#endif
> +
> +    subpic_params.surface = overlay_surface;
> +    subpic_params.output_region = &overlay_region;
> +
> +    err = overlay_vaapi_render(avctx, &params, &subpic_params, output_surface);
> +    if (err < 0)
> +        goto fail;
> +
> +    err = av_frame_copy_props(output, input_main);
> +    if (err < 0)
> +        goto fail;
> +
> +    av_log(avctx, AV_LOG_DEBUG, "Filter output: %s, %ux%u (%"PRId64").\n",
> +           av_get_pix_fmt_name(output->format),
> +           output->width, output->height, output->pts);
> +
> +    return ff_filter_frame(outlink, output);
> +
> +fail:
> +    av_frame_free(&output);
> +    return err;
> +}
> +
> +static int overlay_vaapi_init_framesync(AVFilterContext *avctx)
> +{
> +    OverlayVAAPIContext *ctx = avctx->priv;
> +    int ret, i;
> +
> +    ctx->fs.on_event = overlay_vaapi_blend;
> +    ctx->fs.opaque   = ctx;
> +    ret = ff_framesync_init(&ctx->fs, avctx, avctx->nb_inputs);
> +    if (ret < 0)
> +        return ret;
> +
> +    for (i = 0; i < avctx->nb_inputs; i++) {
> +        FFFrameSyncIn *in = &ctx->fs.in[i];
> +        in->before    = EXT_STOP;
> +        in->after     = EXT_INFINITY;
> +        in->sync      = i ? 1 : 2;
> +        in->time_base = avctx->inputs[i]->time_base;
> +    }
> +
> +    return ff_framesync_configure(&ctx->fs);
> +}
> +
> +static int overlay_vaapi_config_output(AVFilterLink *outlink)
> +{
> +    AVFilterContext  *avctx  = outlink->src;
> +    OverlayVAAPIContext *ctx = avctx->priv;
> +    VAAPIVPPContext *vpp_ctx = avctx->priv;
> +    AVFilterLink        *in0 = avctx->inputs[0];
> +    AVFilterLink        *in1 = avctx->inputs[1];
> +    int err;
> +
> +    av_log(ctx, AV_LOG_DEBUG, "Output is of %s.\n", av_get_pix_fmt_name(outlink->format));
> +    if ((in0->format == AV_PIX_FMT_VAAPI && in1->format != AV_PIX_FMT_VAAPI) ||
> +        (in0->format != AV_PIX_FMT_VAAPI && in1->format == AV_PIX_FMT_VAAPI)) {
> +        av_log(ctx, AV_LOG_ERROR, "Mixing hardware and software pixel formats is not supported.\n");
> +        return AVERROR(EINVAL);
> +    }
> +
> +    err = overlay_vaapi_init_framesync(avctx);
> +    if (err < 0)
> +        return err;
> +
> +    vpp_ctx->output_width  = avctx->inputs[0]->w;
> +    vpp_ctx->output_height = avctx->inputs[0]->h;
> +
> +    err = ff_vaapi_vpp_config_output(outlink);
> +    if (err < 0)
> +        return err;
> +
> +    err = ff_framesync_init_dualinput(&ctx->fs, avctx);
> +    if (err < 0)
> +        return err;
> +
> +    return ff_framesync_configure(&ctx->fs);
> +}
> +
> +static av_cold int overlay_vaapi_init(AVFilterContext *avctx)
> +{
> +    VAAPIVPPContext *vpp_ctx = avctx->priv;
> +
> +    ff_vaapi_vpp_ctx_init(avctx);
> +    vpp_ctx->output_format = AV_PIX_FMT_NONE;
> +
> +    return 0;
> +}
> +
> +static int overlay_vaapi_activate(AVFilterContext *avctx)
> +{
> +    OverlayVAAPIContext *ctx = avctx->priv;
> +
> +    return ff_framesync_activate(&ctx->fs);
> +}
> +
> +static av_cold void overlay_vaapi_uninit(AVFilterContext *avctx)
> +{
> +    OverlayVAAPIContext *ctx = avctx->priv;
> +
> +    ff_framesync_uninit(&ctx->fs);
> +}
> +
> +#define OFFSET(x) offsetof(OverlayVAAPIContext, x)
> +#define FLAGS (AV_OPT_FLAG_FILTERING_PARAM | AV_OPT_FLAG_VIDEO_PARAM)
> +static const AVOption overlay_vaapi_options[] = {
> +    { "x", "Overlay x position",
> +      OFFSET(overlay_x), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = FLAGS },
> +    { "y", "Overlay y position",
> +      OFFSET(overlay_y), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = FLAGS },
> +    { "w", "Overlay width",
> +      OFFSET(overlay_w), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = FLAGS },
> +    { "h", "Overlay hight",
> +      OFFSET(overlay_h), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = FLAGS },
> +    { "alpha", "Overlay global alpha",
> +      OFFSET(overlay_alpha), AV_OPT_TYPE_FLOAT, { .dbl = 0.0}, 0.0, 1.0, .flags = FLAGS},
> +    { NULL },
> +};
> +
> +AVFILTER_DEFINE_CLASS(overlay_vaapi);
> +
> +static const AVFilterPad overlay_vaapi_inputs[] = {
> +    {
> +        .name             = "main",
> +        .type             = AVMEDIA_TYPE_VIDEO,
> +        .get_video_buffer = ff_default_get_video_buffer,
> +        .config_props     = &ff_vaapi_vpp_config_input,
> +        .needs_fifo       = 1,
> +    },
> +    {
> +        .name             = "overlay",
> +        .type             = AVMEDIA_TYPE_VIDEO,
> +        .get_video_buffer = ff_default_get_video_buffer,
> +        .needs_fifo       = 1,
> +    },
> +    { NULL }
> +};
> +
> +static const AVFilterPad overlay_vaapi_outputs[] = {
> +    {
> +        .name          = "default",
> +        .type          = AVMEDIA_TYPE_VIDEO,
> +        .config_props  = &overlay_vaapi_config_output,
> +    },
> +    { NULL }
> +};
> +
> +AVFilter ff_vf_overlay_vaapi = {
> +    .name            = "overlay_vaapi",
> +    .description     = NULL_IF_CONFIG_SMALL("Overlay one video on top of another"),
> +    .priv_size       = sizeof(OverlayVAAPIContext),
> +    .priv_class      = &overlay_vaapi_class,
> +    .init            = &overlay_vaapi_init,
> +    .uninit          = &overlay_vaapi_uninit,
> +    .query_formats   = &overlay_vaapi_query_formats,
> +    .activate        = &overlay_vaapi_activate,
> +    .inputs          = overlay_vaapi_inputs,
> +    .outputs         = overlay_vaapi_outputs,
> +    .flags_internal  = FF_FILTER_FLAG_HWFRAME_AWARE,
> +};
Zhong Li Aug. 28, 2019, 8:37 a.m. UTC | #2
> From: ffmpeg-devel [mailto:ffmpeg-devel-bounces@ffmpeg.org] On Behalf

> Of Zachary Zhou

> Sent: Thursday, July 11, 2019 5:30 PM

> To: ffmpeg-devel@ffmpeg.org

> Cc: Zhou, Zachary <zachary.zhou@intel.com>

> Subject: [FFmpeg-devel] [PATCH v4] avfilter/vaapi: add overlay_vaapi filter

> 

> ---

>  configure                      |   1 +

>  libavfilter/Makefile           |   1 +

>  libavfilter/allfilters.c       |   1 +

>  libavfilter/vf_overlay_vaapi.c | 424


Would be good to update the filter doc when you add a new one.

> +++++++++++++++++++++++++++++++++

>  4 files changed, 427 insertions(+)

>  create mode 100644 libavfilter/vf_overlay_vaapi.c

> 

> diff --git a/configure b/configure

> index 32fc26356c..f469e6a3b1 100755

> --- a/configure

> +++ b/configure

> @@ -3478,6 +3478,7 @@ openclsrc_filter_deps="opencl"

>  overlay_opencl_filter_deps="opencl"

>  overlay_qsv_filter_deps="libmfx"

>  overlay_qsv_filter_select="qsvvpp"

> +overlay_vaapi_filter_deps="vaapi"

>  owdenoise_filter_deps="gpl"

>  pan_filter_deps="swresample"

>  perspective_filter_deps="gpl"

> diff --git a/libavfilter/Makefile b/libavfilter/Makefile index

> 07ea8d7edc..ccaad0d6a4 100644

> --- a/libavfilter/Makefile

> +++ b/libavfilter/Makefile

> @@ -311,6 +311,7 @@ OBJS-$(CONFIG_OVERLAY_FILTER)

> += vf_overlay.o framesync.o

>  OBJS-$(CONFIG_OVERLAY_OPENCL_FILTER)         +=

> vf_overlay_opencl.o opencl.o \

> 

> opencl/overlay.o framesync.o

>  OBJS-$(CONFIG_OVERLAY_QSV_FILTER)            += vf_overlay_qsv.o

> framesync.o

> +OBJS-$(CONFIG_OVERLAY_VAAPI_FILTER)          +=

> vf_overlay_vaapi.o framesync.o vaapi_vpp.o

>  OBJS-$(CONFIG_OWDENOISE_FILTER)              += vf_owdenoise.o

>  OBJS-$(CONFIG_PAD_FILTER)                    += vf_pad.o

>  OBJS-$(CONFIG_PALETTEGEN_FILTER)             += vf_palettegen.o

> diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c index

> 9c846b1ddd..27ee1df78b 100644

> --- a/libavfilter/allfilters.c

> +++ b/libavfilter/allfilters.c

> @@ -295,6 +295,7 @@ extern AVFilter ff_vf_oscilloscope;  extern AVFilter

> ff_vf_overlay;  extern AVFilter ff_vf_overlay_opencl;  extern AVFilter

> ff_vf_overlay_qsv;

> +extern AVFilter ff_vf_overlay_vaapi;

>  extern AVFilter ff_vf_owdenoise;

>  extern AVFilter ff_vf_pad;

>  extern AVFilter ff_vf_palettegen;

> diff --git a/libavfilter/vf_overlay_vaapi.c b/libavfilter/vf_overlay_vaapi.c

> new file mode 100644 index 0000000000..9fffa0fcb9

> --- /dev/null

> +++ b/libavfilter/vf_overlay_vaapi.c

> @@ -0,0 +1,424 @@

> +/*

> + * This file is part of FFmpeg.

> + *

> + * FFmpeg is free software; you can redistribute it and/or

> + * modify it under the terms of the GNU Lesser General Public

> + * License as published by the Free Software Foundation; either

> + * version 2.1 of the License, or (at your option) any later version.

> + *

> + * FFmpeg is distributed in the hope that it will be useful,

> + * but WITHOUT ANY WARRANTY; without even the implied warranty of

> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the

> GNU

> + * Lesser General Public License for more details.

> + *

> + * You should have received a copy of the GNU Lesser General Public

> + * License along with FFmpeg; if not, write to the Free Software

> + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA

> +02110-1301 USA  */ #include <string.h>

> +

> +#include "libavutil/avassert.h"

> +#include "libavutil/mem.h"

> +#include "libavutil/opt.h"

> +#include "libavutil/pixdesc.h"

> +

> +#include "avfilter.h"

> +#include "framesync.h"

> +#include "formats.h"

> +#include "internal.h"

> +#include "vaapi_vpp.h"

> +

> +typedef struct OverlayVAAPIContext {

> +    VAAPIVPPContext  vpp_ctx; // must be the first field

> +    FFFrameSync      fs;

> +    int              overlay_x;

> +    int              overlay_y;

> +    int              overlay_w;

> +    int              overlay_h;

> +    float            overlay_alpha;


Personally I would like to remove the redundant "overlay_" prefix since they are belong to Overlay context.

> +} OverlayVAAPIContext;

> +

> +static int overlay_vaapi_query_formats(AVFilterContext *ctx) {

> +    int i;

> +    int ret;

> +

> +    static const enum AVPixelFormat main_in_fmts[] = {

> +        AV_PIX_FMT_VAAPI,

> +        AV_PIX_FMT_NONE

> +    };

> +    static const enum AVPixelFormat out_pix_fmts[] = {

> +        AV_PIX_FMT_VAAPI,

> +        AV_PIX_FMT_NONE

> +    };


What are the software pixel formats can be supported for input and output? 

> +    for (i = 0; i < ctx->nb_inputs; i++) {


What is the maximum input can be support? 
If only two, probably using ctx->inputs[MAIN]/ ctx->inputs[OVERLAY] is clearer. 

> +        ret = ff_formats_ref(ff_make_format_list(main_in_fmts),

> &ctx->inputs[i]->out_formats);

> +        if (ret < 0)

> +            return ret;

> +    }

>

> +    ret = ff_formats_ref(ff_make_format_list(out_pix_fmts),

> &ctx->outputs[0]->in_formats);

> +    if (ret < 0)

> +        return ret;

> +

> +    return 0;

> +}

> +

> +static int overlay_vaapi_render(AVFilterContext *avctx,

> +                                VAProcPipelineParameterBuffer

> *params,

> +                                VAProcPipelineParameterBuffer

> *subpic_params,

> +                                VASurfaceID output_surface) {


Is it a good idea to extend ff_vaapi_vpp_render_picture() to support multiple parameter buffer inputs instead of copying lots of code? 

> +    VABufferID params_id;

> +    VABufferID subpic_params_id;

> +    VAStatus vas;

> +    int err = 0;

> +    VAAPIVPPContext *ctx   = avctx->priv;

> +

> +    vas = vaBeginPicture(ctx->hwctx->display,

> +                         ctx->va_context, output_surface);

> +    if (vas != VA_STATUS_SUCCESS) {

> +        av_log(avctx, AV_LOG_ERROR, "Failed to attach new picture: "

> +               "%d (%s).\n", vas, vaErrorStr(vas));

> +        err = AVERROR(EIO);

> +        goto fail;

> +    }

> +

> +    vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context,

> +                         VAProcPipelineParameterBufferType,

> +                         sizeof(*params), 1, params, &params_id);

> +    if (vas != VA_STATUS_SUCCESS) {

> +        av_log(avctx, AV_LOG_ERROR, "Failed to create parameter

> buffer: "

> +               "%d (%s).\n", vas, vaErrorStr(vas));

> +        err = AVERROR(EIO);

> +        goto fail_after_begin;

> +    }

> +    av_log(avctx, AV_LOG_DEBUG, "Pipeline parameter buffer is %#x.\n",

> +           params_id);

> +

> +

> +    vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context,

> +                         VAProcPipelineParameterBufferType,

> +                         sizeof(*subpic_params), 1, subpic_params,

> &subpic_params_id);

> +    if (vas != VA_STATUS_SUCCESS) {

> +        av_log(avctx, AV_LOG_ERROR, "Failed to create parameter

> buffer: "

> +               "%d (%s).\n", vas, vaErrorStr(vas));

> +        err = AVERROR(EIO);

> +        goto fail_after_begin;

> +    }

> +    av_log(avctx, AV_LOG_DEBUG, "Pipeline subpic parameter buffer

> is %#x.\n",

> +           subpic_params_id);

> +

> +    vas = vaRenderPicture(ctx->hwctx->display, ctx->va_context,

> +                          &params_id, 1);

> +    if (vas != VA_STATUS_SUCCESS) {

> +        av_log(avctx, AV_LOG_ERROR, "Failed to render parameter

> buffer: "

> +               "%d (%s).\n", vas, vaErrorStr(vas));

> +        err = AVERROR(EIO);

> +        goto fail_after_begin;

> +    }

> +

> +    vas = vaRenderPicture(ctx->hwctx->display, ctx->va_context,

> +                          &subpic_params_id, 1);

> +    if (vas != VA_STATUS_SUCCESS) {

> +        av_log(avctx, AV_LOG_ERROR, "Failed to render subpic

> parameter buffer: "

> +               "%d (%s).\n", vas, vaErrorStr(vas));

> +        err = AVERROR(EIO);

> +        goto fail_after_begin;

> +    }

> +

> +    vas = vaEndPicture(ctx->hwctx->display, ctx->va_context);

> +    if (vas != VA_STATUS_SUCCESS) {

> +        av_log(avctx, AV_LOG_ERROR, "Failed to start picture processing:

> "

> +               "%d (%s).\n", vas, vaErrorStr(vas));

> +        err = AVERROR(EIO);

> +        goto fail_after_render;

> +    }

> +

> +    if (CONFIG_VAAPI_1 || ctx->hwctx->driver_quirks &

> +        AV_VAAPI_DRIVER_QUIRK_RENDER_PARAM_BUFFERS) {

> +        vas = vaDestroyBuffer(ctx->hwctx->display, params_id);

> +        if (vas != VA_STATUS_SUCCESS) {

> +            av_log(avctx, AV_LOG_ERROR, "Failed to free parameter

> buffer: "

> +                   "%d (%s).\n", vas, vaErrorStr(vas));

> +            // And ignore.

> +        }

> +    }

> +

> +    return 0;

> +

> +    // We want to make sure that if vaBeginPicture has been called, we

> also

> +    // call vaRenderPicture and vaEndPicture.  These calls may well fail or

> +    // do something else nasty, but once we're in this failure case there

> +    // isn't much else we can do.

> +fail_after_begin:

> +    vaRenderPicture(ctx->hwctx->display, ctx->va_context, &params_id,

> +1);

> +fail_after_render:

> +    vaEndPicture(ctx->hwctx->display, ctx->va_context);

> +fail:

> +    return err;

> +}

> +

> +static int overlay_vaapi_blend(FFFrameSync *fs) {

> +    AVFilterContext    *avctx = fs->parent;

> +    AVFilterLink     *outlink = avctx->outputs[0];

> +    OverlayVAAPIContext *ctx  = avctx->priv;

> +    VAAPIVPPContext *vpp_ctx  = avctx->priv;

> +    AVFrame *input_main, *input_overlay;

> +    AVFrame *output;

> +    int err;

> +    VASurfaceID main_surface, overlay_surface, output_surface;

> +    VARectangle overlay_region, output_region;

> +    VAProcPipelineParameterBuffer params, subpic_params;

> +

> +#if VA_CHECK_VERSION(2, 1, 0)

> +    VABlendState blend_state; // Blend State #endif

> +

> +    err = ff_framesync_get_frame(fs, 0, &input_main, 0);

> +    if (err < 0)

> +        return err;

> +    err = ff_framesync_get_frame(fs, 1, &input_overlay, 0);

> +    if (err < 0)

> +        return err;

> +

> +    av_log(avctx, AV_LOG_DEBUG, "Filter main: %s, %ux%u

> (%"PRId64").\n",

> +           av_get_pix_fmt_name(input_main->format),

> +           input_main->width, input_main->height, input_main->pts);

> +

> +    av_log(avctx, AV_LOG_DEBUG, "Filter overlay: %s, %ux%u

> (%"PRId64").\n",

> +           av_get_pix_fmt_name(input_overlay->format),

> +           input_overlay->width, input_overlay->height,

> + input_overlay->pts);

> +

> +    if (vpp_ctx->va_context == VA_INVALID_ID)

> +        return AVERROR(EINVAL);

> +

> +    main_surface = (VASurfaceID)(uintptr_t)input_main->data[3];

> +    av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp

> main.\n",

> +           main_surface);

> +

> +    overlay_surface = (VASurfaceID)(uintptr_t)input_overlay->data[3];

> +    av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp

> overlay.\n",

> +           overlay_surface);

> +

> +    output = ff_get_video_buffer(outlink, outlink->w, outlink->h);

> +    if (!output) {

> +        err = AVERROR(ENOMEM);

> +        goto fail;

> +    }

> +

> +    output_surface = (VASurfaceID)(uintptr_t)output->data[3];

> +    av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp

> output.\n",

> +           output_surface);

> +

> +    memset(&params, 0, sizeof(params));

> +    memset(&subpic_params, 0, sizeof(subpic_params));

> +

> +    overlay_region = (VARectangle) {

> +        .x      = ctx->overlay_x,

> +        .y      = ctx->overlay_y,

> +        .width  = input_overlay->width,

> +        .height = input_overlay->height,

> +    };

> +

> +    output_region = (VARectangle) {

> +        .x      = 0,

> +        .y      = 0,

> +        .width  = output->width,

> +        .height = output->height,

> +    };

> +

> +    if (ctx->overlay_w && ctx->overlay_h) {

> +        overlay_region.width = ctx->overlay_w;

> +        overlay_region.height = ctx->overlay_h;

> +    }

> +

> +    if (overlay_region.width > input_main->width ||

> +        overlay_region.height > input_main->height) {

> +        av_log(ctx, AV_LOG_ERROR,

> +               "Overlay width > main width or overlay height > main

> height is not supported.\n");

> +        err =  AVERROR(EINVAL);

> +        goto fail;

> +    }

> +

> +    params.filters     = &vpp_ctx->filter_buffers[0];

> +    params.num_filters = vpp_ctx->nb_filter_buffers;

> +

> +    err = ff_vaapi_vpp_init_params(avctx, &params,

> +                                   input_main, output);

> +    if (err < 0)

> +        goto fail;

> +

> +    params.output_region = &output_region;

> +    params.output_background_color =

> VAAPI_VPP_BACKGROUND_BLACK;

> +    params.output_color_standard = params.surface_color_standard;

> +

> +    memcpy(&subpic_params, &params, sizeof(subpic_params));

> +

> +#if VA_CHECK_VERSION(2, 1, 0)

> +    blend_state.flag = VA_BLEND_GLOBAL_ALPHA;

> +    blend_state.global_alpha = ctx->overlay_alpha;

> +    subpic_params.blend_state = &blend_state; #endif

> +

> +    subpic_params.surface = overlay_surface;

> +    subpic_params.output_region = &overlay_region;

> +

> +    err = overlay_vaapi_render(avctx, &params, &subpic_params,

> output_surface);

> +    if (err < 0)

> +        goto fail;

> +

> +    err = av_frame_copy_props(output, input_main);

> +    if (err < 0)

> +        goto fail;

> +

> +    av_log(avctx, AV_LOG_DEBUG, "Filter output: %s, %ux%u

> (%"PRId64").\n",

> +           av_get_pix_fmt_name(output->format),

> +           output->width, output->height, output->pts);

> +

> +    return ff_filter_frame(outlink, output);

> +

> +fail:

> +    av_frame_free(&output);

> +    return err;

> +}

> +

> +static int overlay_vaapi_init_framesync(AVFilterContext *avctx) {

> +    OverlayVAAPIContext *ctx = avctx->priv;

> +    int ret, i;

> +

> +    ctx->fs.on_event = overlay_vaapi_blend;

> +    ctx->fs.opaque   = ctx;

> +    ret = ff_framesync_init(&ctx->fs, avctx, avctx->nb_inputs);

> +    if (ret < 0)

> +        return ret;

> +

> +    for (i = 0; i < avctx->nb_inputs; i++) {

> +        FFFrameSyncIn *in = &ctx->fs.in[i];

> +        in->before    = EXT_STOP;

> +        in->after     = EXT_INFINITY;

> +        in->sync      = i ? 1 : 2;

> +        in->time_base = avctx->inputs[i]->time_base;

> +    }

> +

> +    return ff_framesync_configure(&ctx->fs); }

> +

> +static int overlay_vaapi_config_output(AVFilterLink *outlink) {

> +    AVFilterContext  *avctx  = outlink->src;

> +    OverlayVAAPIContext *ctx = avctx->priv;

> +    VAAPIVPPContext *vpp_ctx = avctx->priv;

> +    AVFilterLink        *in0 = avctx->inputs[0];

> +    AVFilterLink        *in1 = avctx->inputs[1];

> +    int err;


Some vaapi drivers may not support overlay, please query the driver capability at run-time.

> +    av_log(ctx, AV_LOG_DEBUG, "Output is of %s.\n",

> av_get_pix_fmt_name(outlink->format));

> +    if ((in0->format == AV_PIX_FMT_VAAPI && in1->format !=

> AV_PIX_FMT_VAAPI) ||

> +        (in0->format != AV_PIX_FMT_VAAPI && in1->format ==

> AV_PIX_FMT_VAAPI)) {

> +        av_log(ctx, AV_LOG_ERROR, "Mixing hardware and software

> pixel formats is not supported.\n");

> +        return AVERROR(EINVAL);

> +    }

> +

> +    err = overlay_vaapi_init_framesync(avctx);

> +    if (err < 0)

> +        return err;

> +

> +    vpp_ctx->output_width  = avctx->inputs[0]->w;

> +    vpp_ctx->output_height = avctx->inputs[0]->h;

> +

> +    err = ff_vaapi_vpp_config_output(outlink);

> +    if (err < 0)

> +        return err;

> +

> +    err = ff_framesync_init_dualinput(&ctx->fs, avctx);

> +    if (err < 0)

> +        return err;

> +

> +    return ff_framesync_configure(&ctx->fs); }

> +

> +static av_cold int overlay_vaapi_init(AVFilterContext *avctx) {

> +    VAAPIVPPContext *vpp_ctx = avctx->priv;

> +

> +    ff_vaapi_vpp_ctx_init(avctx);

> +    vpp_ctx->output_format = AV_PIX_FMT_NONE;

> +

> +    return 0;

> +}

> +

> +static int overlay_vaapi_activate(AVFilterContext *avctx) {

> +    OverlayVAAPIContext *ctx = avctx->priv;

> +

> +    return ff_framesync_activate(&ctx->fs); }

> +

> +static av_cold void overlay_vaapi_uninit(AVFilterContext *avctx) {

> +    OverlayVAAPIContext *ctx = avctx->priv;

> +

> +    ff_framesync_uninit(&ctx->fs);

> +}

> +

> +#define OFFSET(x) offsetof(OverlayVAAPIContext, x) #define FLAGS

> +(AV_OPT_FLAG_FILTERING_PARAM | AV_OPT_FLAG_VIDEO_PARAM) static

> const

> +AVOption overlay_vaapi_options[] = {

> +    { "x", "Overlay x position",

> +      OFFSET(overlay_x), AV_OPT_TYPE_INT, { .i64 = 0 }, 0,

> INT_MAX, .flags = FLAGS },

> +    { "y", "Overlay y position",

> +      OFFSET(overlay_y), AV_OPT_TYPE_INT, { .i64 = 0 }, 0,

> INT_MAX, .flags = FLAGS },

> +    { "w", "Overlay width",

> +      OFFSET(overlay_w), AV_OPT_TYPE_INT, { .i64 = 0 }, 0,

> INT_MAX, .flags = FLAGS },

> +    { "h", "Overlay hight",


height, not hight

> +      OFFSET(overlay_h), AV_OPT_TYPE_INT, { .i64 = 0 }, 0,

> INT_MAX, .flags = FLAGS },

> +    { "alpha", "Overlay global alpha",

> +      OFFSET(overlay_alpha), AV_OPT_TYPE_FLOAT, { .dbl = 0.0}, 0.0,

> 1.0, .flags = FLAGS},

> +    { NULL },

> +};

> +

> +AVFILTER_DEFINE_CLASS(overlay_vaapi);

> +

> +static const AVFilterPad overlay_vaapi_inputs[] = {

> +    {

> +        .name             = "main",

> +        .type             = AVMEDIA_TYPE_VIDEO,

> +        .get_video_buffer = ff_default_get_video_buffer,

> +        .config_props     = &ff_vaapi_vpp_config_input,

> +        .needs_fifo       = 1,

> +    },

> +    {

> +        .name             = "overlay",

> +        .type             = AVMEDIA_TYPE_VIDEO,

> +        .get_video_buffer = ff_default_get_video_buffer,

> +        .needs_fifo       = 1,

> +    },

> +    { NULL }

> +};

> +

> +static const AVFilterPad overlay_vaapi_outputs[] = {

> +    {

> +        .name          = "default",

> +        .type          = AVMEDIA_TYPE_VIDEO,

> +        .config_props  = &overlay_vaapi_config_output,

> +    },

> +    { NULL }

> +};

> +

> +AVFilter ff_vf_overlay_vaapi = {

> +    .name            = "overlay_vaapi",

> +    .description     = NULL_IF_CONFIG_SMALL("Overlay one video on

> top of another"),

> +    .priv_size       = sizeof(OverlayVAAPIContext),

> +    .priv_class      = &overlay_vaapi_class,

> +    .init            = &overlay_vaapi_init,

> +    .uninit          = &overlay_vaapi_uninit,

> +    .query_formats   = &overlay_vaapi_query_formats,

> +    .activate        = &overlay_vaapi_activate,

> +    .inputs          = overlay_vaapi_inputs,

> +    .outputs         = overlay_vaapi_outputs,

> +    .flags_internal  = FF_FILTER_FLAG_HWFRAME_AWARE, };

> --

> 2.17.1
Zachary Zhou Aug. 29, 2019, 5:29 a.m. UTC | #3
> -----Original Message-----

> From: Li, Zhong

> Sent: Wednesday, August 28, 2019 4:37 PM

> To: FFmpeg development discussions and patches <ffmpeg-devel@ffmpeg.org>

> Cc: Zhou, Zachary <zachary.zhou@intel.com>

> Subject: RE: [FFmpeg-devel] [PATCH v4] avfilter/vaapi: add overlay_vaapi filter

> 

> > From: ffmpeg-devel [mailto:ffmpeg-devel-bounces@ffmpeg.org] On Behalf

> > Of Zachary Zhou

> > Sent: Thursday, July 11, 2019 5:30 PM

> > To: ffmpeg-devel@ffmpeg.org

> > Cc: Zhou, Zachary <zachary.zhou@intel.com>

> > Subject: [FFmpeg-devel] [PATCH v4] avfilter/vaapi: add overlay_vaapi

> > filter

> >

> > ---

> >  configure                      |   1 +

> >  libavfilter/Makefile           |   1 +

> >  libavfilter/allfilters.c       |   1 +

> >  libavfilter/vf_overlay_vaapi.c | 424

> 

> Would be good to update the filter doc when you add a new one.


will update the filter doc in the next version.

> 

> > +++++++++++++++++++++++++++++++++

> >  4 files changed, 427 insertions(+)

> >  create mode 100644 libavfilter/vf_overlay_vaapi.c

> >

> > diff --git a/configure b/configure

> > index 32fc26356c..f469e6a3b1 100755

> > --- a/configure

> > +++ b/configure

> > @@ -3478,6 +3478,7 @@ openclsrc_filter_deps="opencl"

> >  overlay_opencl_filter_deps="opencl"

> >  overlay_qsv_filter_deps="libmfx"

> >  overlay_qsv_filter_select="qsvvpp"

> > +overlay_vaapi_filter_deps="vaapi"

> >  owdenoise_filter_deps="gpl"

> >  pan_filter_deps="swresample"

> >  perspective_filter_deps="gpl"

> > diff --git a/libavfilter/Makefile b/libavfilter/Makefile index

> > 07ea8d7edc..ccaad0d6a4 100644

> > --- a/libavfilter/Makefile

> > +++ b/libavfilter/Makefile

> > @@ -311,6 +311,7 @@ OBJS-$(CONFIG_OVERLAY_FILTER)

> > += vf_overlay.o framesync.o

> >  OBJS-$(CONFIG_OVERLAY_OPENCL_FILTER)         +=

> > vf_overlay_opencl.o opencl.o \

> >

> > opencl/overlay.o framesync.o

> >  OBJS-$(CONFIG_OVERLAY_QSV_FILTER)            += vf_overlay_qsv.o

> > framesync.o

> > +OBJS-$(CONFIG_OVERLAY_VAAPI_FILTER)          +=

> > vf_overlay_vaapi.o framesync.o vaapi_vpp.o

> >  OBJS-$(CONFIG_OWDENOISE_FILTER)              += vf_owdenoise.o

> >  OBJS-$(CONFIG_PAD_FILTER)                    += vf_pad.o

> >  OBJS-$(CONFIG_PALETTEGEN_FILTER)             += vf_palettegen.o

> > diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c index

> > 9c846b1ddd..27ee1df78b 100644

> > --- a/libavfilter/allfilters.c

> > +++ b/libavfilter/allfilters.c

> > @@ -295,6 +295,7 @@ extern AVFilter ff_vf_oscilloscope;  extern

> > AVFilter ff_vf_overlay;  extern AVFilter ff_vf_overlay_opencl;  extern

> > AVFilter ff_vf_overlay_qsv;

> > +extern AVFilter ff_vf_overlay_vaapi;

> >  extern AVFilter ff_vf_owdenoise;

> >  extern AVFilter ff_vf_pad;

> >  extern AVFilter ff_vf_palettegen;

> > diff --git a/libavfilter/vf_overlay_vaapi.c

> > b/libavfilter/vf_overlay_vaapi.c new file mode 100644 index

> > 0000000000..9fffa0fcb9

> > --- /dev/null

> > +++ b/libavfilter/vf_overlay_vaapi.c

> > @@ -0,0 +1,424 @@

> > +/*

> > + * This file is part of FFmpeg.

> > + *

> > + * FFmpeg is free software; you can redistribute it and/or

> > + * modify it under the terms of the GNU Lesser General Public

> > + * License as published by the Free Software Foundation; either

> > + * version 2.1 of the License, or (at your option) any later version.

> > + *

> > + * FFmpeg is distributed in the hope that it will be useful,

> > + * but WITHOUT ANY WARRANTY; without even the implied warranty of

> > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the

> > GNU

> > + * Lesser General Public License for more details.

> > + *

> > + * You should have received a copy of the GNU Lesser General Public

> > + * License along with FFmpeg; if not, write to the Free Software

> > + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA

> > +02110-1301 USA  */ #include <string.h>

> > +

> > +#include "libavutil/avassert.h"

> > +#include "libavutil/mem.h"

> > +#include "libavutil/opt.h"

> > +#include "libavutil/pixdesc.h"

> > +

> > +#include "avfilter.h"

> > +#include "framesync.h"

> > +#include "formats.h"

> > +#include "internal.h"

> > +#include "vaapi_vpp.h"

> > +

> > +typedef struct OverlayVAAPIContext {

> > +    VAAPIVPPContext  vpp_ctx; // must be the first field

> > +    FFFrameSync      fs;

> > +    int              overlay_x;

> > +    int              overlay_y;

> > +    int              overlay_w;

> > +    int              overlay_h;

> > +    float            overlay_alpha;

> 

> Personally I would like to remove the redundant "overlay_" prefix since they

> are belong to Overlay context.


will do the change.

> 

> > +} OverlayVAAPIContext;

> > +

> > +static int overlay_vaapi_query_formats(AVFilterContext *ctx) {

> > +    int i;

> > +    int ret;

> > +

> > +    static const enum AVPixelFormat main_in_fmts[] = {

> > +        AV_PIX_FMT_VAAPI,

> > +        AV_PIX_FMT_NONE

> > +    };

> > +    static const enum AVPixelFormat out_pix_fmts[] = {

> > +        AV_PIX_FMT_VAAPI,

> > +        AV_PIX_FMT_NONE

> > +    };

> 

> What are the software pixel formats can be supported for input and output?

Because this filter only support HW acceleration, so remove software pixel formats here.
This is based on MarkT previous comment.

> 

> > +    for (i = 0; i < ctx->nb_inputs; i++) {

> 

> What is the maximum input can be support?

> If only two, probably using ctx->inputs[MAIN]/ ctx->inputs[OVERLAY] is clearer.


will do the changes

> 

> > +        ret = ff_formats_ref(ff_make_format_list(main_in_fmts),

> > &ctx->inputs[i]->out_formats);

> > +        if (ret < 0)

> > +            return ret;

> > +    }

> >

> > +    ret = ff_formats_ref(ff_make_format_list(out_pix_fmts),

> > &ctx->outputs[0]->in_formats);

> > +    if (ret < 0)

> > +        return ret;

> > +

> > +    return 0;

> > +}

> > +

> > +static int overlay_vaapi_render(AVFilterContext *avctx,

> > +                                VAProcPipelineParameterBuffer

> > *params,

> > +                                VAProcPipelineParameterBuffer

> > *subpic_params,

> > +                                VASurfaceID output_surface) {

> 

> Is it a good idea to extend ff_vaapi_vpp_render_picture() to support multiple

> parameter buffer inputs instead of copying lots of code?

Actually this function will call two times vaRenderPicture, and ff_vaapi_vpp_render_picture is used by many vaapi vaapi filters, maybe it is not good to change global function ff_vaapi_vpp_render_picture.
please let me know if you still want to change ff_vaapi_vpp_render_picture.

> 

> > +    VABufferID params_id;

> > +    VABufferID subpic_params_id;

> > +    VAStatus vas;

> > +    int err = 0;

> > +    VAAPIVPPContext *ctx   = avctx->priv;

> > +

> > +    vas = vaBeginPicture(ctx->hwctx->display,

> > +                         ctx->va_context, output_surface);

> > +    if (vas != VA_STATUS_SUCCESS) {

> > +        av_log(avctx, AV_LOG_ERROR, "Failed to attach new picture: "

> > +               "%d (%s).\n", vas, vaErrorStr(vas));

> > +        err = AVERROR(EIO);

> > +        goto fail;

> > +    }

> > +

> > +    vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context,

> > +                         VAProcPipelineParameterBufferType,

> > +                         sizeof(*params), 1, params, &params_id);

> > +    if (vas != VA_STATUS_SUCCESS) {

> > +        av_log(avctx, AV_LOG_ERROR, "Failed to create parameter

> > buffer: "

> > +               "%d (%s).\n", vas, vaErrorStr(vas));

> > +        err = AVERROR(EIO);

> > +        goto fail_after_begin;

> > +    }

> > +    av_log(avctx, AV_LOG_DEBUG, "Pipeline parameter buffer is %#x.\n",

> > +           params_id);

> > +

> > +

> > +    vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context,

> > +                         VAProcPipelineParameterBufferType,

> > +                         sizeof(*subpic_params), 1, subpic_params,

> > &subpic_params_id);

> > +    if (vas != VA_STATUS_SUCCESS) {

> > +        av_log(avctx, AV_LOG_ERROR, "Failed to create parameter

> > buffer: "

> > +               "%d (%s).\n", vas, vaErrorStr(vas));

> > +        err = AVERROR(EIO);

> > +        goto fail_after_begin;

> > +    }

> > +    av_log(avctx, AV_LOG_DEBUG, "Pipeline subpic parameter buffer

> > is %#x.\n",

> > +           subpic_params_id);

> > +

> > +    vas = vaRenderPicture(ctx->hwctx->display, ctx->va_context,

> > +                          &params_id, 1);

> > +    if (vas != VA_STATUS_SUCCESS) {

> > +        av_log(avctx, AV_LOG_ERROR, "Failed to render parameter

> > buffer: "

> > +               "%d (%s).\n", vas, vaErrorStr(vas));

> > +        err = AVERROR(EIO);

> > +        goto fail_after_begin;

> > +    }

> > +

> > +    vas = vaRenderPicture(ctx->hwctx->display, ctx->va_context,

> > +                          &subpic_params_id, 1);

> > +    if (vas != VA_STATUS_SUCCESS) {

> > +        av_log(avctx, AV_LOG_ERROR, "Failed to render subpic

> > parameter buffer: "

> > +               "%d (%s).\n", vas, vaErrorStr(vas));

> > +        err = AVERROR(EIO);

> > +        goto fail_after_begin;

> > +    }

> > +

> > +    vas = vaEndPicture(ctx->hwctx->display, ctx->va_context);

> > +    if (vas != VA_STATUS_SUCCESS) {

> > +        av_log(avctx, AV_LOG_ERROR, "Failed to start picture processing:

> > "

> > +               "%d (%s).\n", vas, vaErrorStr(vas));

> > +        err = AVERROR(EIO);

> > +        goto fail_after_render;

> > +    }

> > +

> > +    if (CONFIG_VAAPI_1 || ctx->hwctx->driver_quirks &

> > +        AV_VAAPI_DRIVER_QUIRK_RENDER_PARAM_BUFFERS) {

> > +        vas = vaDestroyBuffer(ctx->hwctx->display, params_id);

> > +        if (vas != VA_STATUS_SUCCESS) {

> > +            av_log(avctx, AV_LOG_ERROR, "Failed to free parameter

> > buffer: "

> > +                   "%d (%s).\n", vas, vaErrorStr(vas));

> > +            // And ignore.

> > +        }

> > +    }

> > +

> > +    return 0;

> > +

> > +    // We want to make sure that if vaBeginPicture has been called,

> > + we

> > also

> > +    // call vaRenderPicture and vaEndPicture.  These calls may well fail or

> > +    // do something else nasty, but once we're in this failure case there

> > +    // isn't much else we can do.

> > +fail_after_begin:

> > +    vaRenderPicture(ctx->hwctx->display, ctx->va_context, &params_id,

> > +1);

> > +fail_after_render:

> > +    vaEndPicture(ctx->hwctx->display, ctx->va_context);

> > +fail:

> > +    return err;

> > +}

> > +

> > +static int overlay_vaapi_blend(FFFrameSync *fs) {

> > +    AVFilterContext    *avctx = fs->parent;

> > +    AVFilterLink     *outlink = avctx->outputs[0];

> > +    OverlayVAAPIContext *ctx  = avctx->priv;

> > +    VAAPIVPPContext *vpp_ctx  = avctx->priv;

> > +    AVFrame *input_main, *input_overlay;

> > +    AVFrame *output;

> > +    int err;

> > +    VASurfaceID main_surface, overlay_surface, output_surface;

> > +    VARectangle overlay_region, output_region;

> > +    VAProcPipelineParameterBuffer params, subpic_params;

> > +

> > +#if VA_CHECK_VERSION(2, 1, 0)

> > +    VABlendState blend_state; // Blend State #endif

> > +

> > +    err = ff_framesync_get_frame(fs, 0, &input_main, 0);

> > +    if (err < 0)

> > +        return err;

> > +    err = ff_framesync_get_frame(fs, 1, &input_overlay, 0);

> > +    if (err < 0)

> > +        return err;

> > +

> > +    av_log(avctx, AV_LOG_DEBUG, "Filter main: %s, %ux%u

> > (%"PRId64").\n",

> > +           av_get_pix_fmt_name(input_main->format),

> > +           input_main->width, input_main->height, input_main->pts);

> > +

> > +    av_log(avctx, AV_LOG_DEBUG, "Filter overlay: %s, %ux%u

> > (%"PRId64").\n",

> > +           av_get_pix_fmt_name(input_overlay->format),

> > +           input_overlay->width, input_overlay->height,

> > + input_overlay->pts);

> > +

> > +    if (vpp_ctx->va_context == VA_INVALID_ID)

> > +        return AVERROR(EINVAL);

> > +

> > +    main_surface = (VASurfaceID)(uintptr_t)input_main->data[3];

> > +    av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp

> > main.\n",

> > +           main_surface);

> > +

> > +    overlay_surface = (VASurfaceID)(uintptr_t)input_overlay->data[3];

> > +    av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp

> > overlay.\n",

> > +           overlay_surface);

> > +

> > +    output = ff_get_video_buffer(outlink, outlink->w, outlink->h);

> > +    if (!output) {

> > +        err = AVERROR(ENOMEM);

> > +        goto fail;

> > +    }

> > +

> > +    output_surface = (VASurfaceID)(uintptr_t)output->data[3];

> > +    av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp

> > output.\n",

> > +           output_surface);

> > +

> > +    memset(&params, 0, sizeof(params));

> > +    memset(&subpic_params, 0, sizeof(subpic_params));

> > +

> > +    overlay_region = (VARectangle) {

> > +        .x      = ctx->overlay_x,

> > +        .y      = ctx->overlay_y,

> > +        .width  = input_overlay->width,

> > +        .height = input_overlay->height,

> > +    };

> > +

> > +    output_region = (VARectangle) {

> > +        .x      = 0,

> > +        .y      = 0,

> > +        .width  = output->width,

> > +        .height = output->height,

> > +    };

> > +

> > +    if (ctx->overlay_w && ctx->overlay_h) {

> > +        overlay_region.width = ctx->overlay_w;

> > +        overlay_region.height = ctx->overlay_h;

> > +    }

> > +

> > +    if (overlay_region.width > input_main->width ||

> > +        overlay_region.height > input_main->height) {

> > +        av_log(ctx, AV_LOG_ERROR,

> > +               "Overlay width > main width or overlay height > main

> > height is not supported.\n");

> > +        err =  AVERROR(EINVAL);

> > +        goto fail;

> > +    }

> > +

> > +    params.filters     = &vpp_ctx->filter_buffers[0];

> > +    params.num_filters = vpp_ctx->nb_filter_buffers;

> > +

> > +    err = ff_vaapi_vpp_init_params(avctx, &params,

> > +                                   input_main, output);

> > +    if (err < 0)

> > +        goto fail;

> > +

> > +    params.output_region = &output_region;

> > +    params.output_background_color =

> > VAAPI_VPP_BACKGROUND_BLACK;

> > +    params.output_color_standard = params.surface_color_standard;

> > +

> > +    memcpy(&subpic_params, &params, sizeof(subpic_params));

> > +

> > +#if VA_CHECK_VERSION(2, 1, 0)

> > +    blend_state.flag = VA_BLEND_GLOBAL_ALPHA;

> > +    blend_state.global_alpha = ctx->overlay_alpha;

> > +    subpic_params.blend_state = &blend_state; #endif

> > +

> > +    subpic_params.surface = overlay_surface;

> > +    subpic_params.output_region = &overlay_region;

> > +

> > +    err = overlay_vaapi_render(avctx, &params, &subpic_params,

> > output_surface);

> > +    if (err < 0)

> > +        goto fail;

> > +

> > +    err = av_frame_copy_props(output, input_main);

> > +    if (err < 0)

> > +        goto fail;

> > +

> > +    av_log(avctx, AV_LOG_DEBUG, "Filter output: %s, %ux%u

> > (%"PRId64").\n",

> > +           av_get_pix_fmt_name(output->format),

> > +           output->width, output->height, output->pts);

> > +

> > +    return ff_filter_frame(outlink, output);

> > +

> > +fail:

> > +    av_frame_free(&output);

> > +    return err;

> > +}

> > +

> > +static int overlay_vaapi_init_framesync(AVFilterContext *avctx) {

> > +    OverlayVAAPIContext *ctx = avctx->priv;

> > +    int ret, i;

> > +

> > +    ctx->fs.on_event = overlay_vaapi_blend;

> > +    ctx->fs.opaque   = ctx;

> > +    ret = ff_framesync_init(&ctx->fs, avctx, avctx->nb_inputs);

> > +    if (ret < 0)

> > +        return ret;

> > +

> > +    for (i = 0; i < avctx->nb_inputs; i++) {

> > +        FFFrameSyncIn *in = &ctx->fs.in[i];

> > +        in->before    = EXT_STOP;

> > +        in->after     = EXT_INFINITY;

> > +        in->sync      = i ? 1 : 2;

> > +        in->time_base = avctx->inputs[i]->time_base;

> > +    }

> > +

> > +    return ff_framesync_configure(&ctx->fs); }

> > +

> > +static int overlay_vaapi_config_output(AVFilterLink *outlink) {

> > +    AVFilterContext  *avctx  = outlink->src;

> > +    OverlayVAAPIContext *ctx = avctx->priv;

> > +    VAAPIVPPContext *vpp_ctx = avctx->priv;

> > +    AVFilterLink        *in0 = avctx->inputs[0];

> > +    AVFilterLink        *in1 = avctx->inputs[1];

> > +    int err;

> 

> Some vaapi drivers may not support overlay, please query the driver

> capability at run-time.


I totally agree this, will add this check.

> 

> > +    av_log(ctx, AV_LOG_DEBUG, "Output is of %s.\n",

> > av_get_pix_fmt_name(outlink->format));

> > +    if ((in0->format == AV_PIX_FMT_VAAPI && in1->format !=

> > AV_PIX_FMT_VAAPI) ||

> > +        (in0->format != AV_PIX_FMT_VAAPI && in1->format ==

> > AV_PIX_FMT_VAAPI)) {

> > +        av_log(ctx, AV_LOG_ERROR, "Mixing hardware and software

> > pixel formats is not supported.\n");

> > +        return AVERROR(EINVAL);

> > +    }

> > +

> > +    err = overlay_vaapi_init_framesync(avctx);

> > +    if (err < 0)

> > +        return err;

> > +

> > +    vpp_ctx->output_width  = avctx->inputs[0]->w;

> > +    vpp_ctx->output_height = avctx->inputs[0]->h;

> > +

> > +    err = ff_vaapi_vpp_config_output(outlink);

> > +    if (err < 0)

> > +        return err;

> > +

> > +    err = ff_framesync_init_dualinput(&ctx->fs, avctx);

> > +    if (err < 0)

> > +        return err;

> > +

> > +    return ff_framesync_configure(&ctx->fs); }

> > +

> > +static av_cold int overlay_vaapi_init(AVFilterContext *avctx) {

> > +    VAAPIVPPContext *vpp_ctx = avctx->priv;

> > +

> > +    ff_vaapi_vpp_ctx_init(avctx);

> > +    vpp_ctx->output_format = AV_PIX_FMT_NONE;

> > +

> > +    return 0;

> > +}

> > +

> > +static int overlay_vaapi_activate(AVFilterContext *avctx) {

> > +    OverlayVAAPIContext *ctx = avctx->priv;

> > +

> > +    return ff_framesync_activate(&ctx->fs); }

> > +

> > +static av_cold void overlay_vaapi_uninit(AVFilterContext *avctx) {

> > +    OverlayVAAPIContext *ctx = avctx->priv;

> > +

> > +    ff_framesync_uninit(&ctx->fs);

> > +}

> > +

> > +#define OFFSET(x) offsetof(OverlayVAAPIContext, x) #define FLAGS

> > +(AV_OPT_FLAG_FILTERING_PARAM | AV_OPT_FLAG_VIDEO_PARAM) static

> > const

> > +AVOption overlay_vaapi_options[] = {

> > +    { "x", "Overlay x position",

> > +      OFFSET(overlay_x), AV_OPT_TYPE_INT, { .i64 = 0 }, 0,

> > INT_MAX, .flags = FLAGS },

> > +    { "y", "Overlay y position",

> > +      OFFSET(overlay_y), AV_OPT_TYPE_INT, { .i64 = 0 }, 0,

> > INT_MAX, .flags = FLAGS },

> > +    { "w", "Overlay width",

> > +      OFFSET(overlay_w), AV_OPT_TYPE_INT, { .i64 = 0 }, 0,

> > INT_MAX, .flags = FLAGS },

> > +    { "h", "Overlay hight",

> 

> height, not hight

> 

> > +      OFFSET(overlay_h), AV_OPT_TYPE_INT, { .i64 = 0 }, 0,

> > INT_MAX, .flags = FLAGS },

> > +    { "alpha", "Overlay global alpha",

> > +      OFFSET(overlay_alpha), AV_OPT_TYPE_FLOAT, { .dbl = 0.0}, 0.0,

> > 1.0, .flags = FLAGS},

> > +    { NULL },

> > +};

> > +

> > +AVFILTER_DEFINE_CLASS(overlay_vaapi);

> > +

> > +static const AVFilterPad overlay_vaapi_inputs[] = {

> > +    {

> > +        .name             = "main",

> > +        .type             = AVMEDIA_TYPE_VIDEO,

> > +        .get_video_buffer = ff_default_get_video_buffer,

> > +        .config_props     = &ff_vaapi_vpp_config_input,

> > +        .needs_fifo       = 1,

> > +    },

> > +    {

> > +        .name             = "overlay",

> > +        .type             = AVMEDIA_TYPE_VIDEO,

> > +        .get_video_buffer = ff_default_get_video_buffer,

> > +        .needs_fifo       = 1,

> > +    },

> > +    { NULL }

> > +};

> > +

> > +static const AVFilterPad overlay_vaapi_outputs[] = {

> > +    {

> > +        .name          = "default",

> > +        .type          = AVMEDIA_TYPE_VIDEO,

> > +        .config_props  = &overlay_vaapi_config_output,

> > +    },

> > +    { NULL }

> > +};

> > +

> > +AVFilter ff_vf_overlay_vaapi = {

> > +    .name            = "overlay_vaapi",

> > +    .description     = NULL_IF_CONFIG_SMALL("Overlay one video on

> > top of another"),

> > +    .priv_size       = sizeof(OverlayVAAPIContext),

> > +    .priv_class      = &overlay_vaapi_class,

> > +    .init            = &overlay_vaapi_init,

> > +    .uninit          = &overlay_vaapi_uninit,

> > +    .query_formats   = &overlay_vaapi_query_formats,

> > +    .activate        = &overlay_vaapi_activate,

> > +    .inputs          = overlay_vaapi_inputs,

> > +    .outputs         = overlay_vaapi_outputs,

> > +    .flags_internal  = FF_FILTER_FLAG_HWFRAME_AWARE, };

> > --


Thank you Zhong for the review.

Thanks
Zachary

> > 2.17.1
diff mbox

Patch

diff --git a/configure b/configure
index 32fc26356c..f469e6a3b1 100755
--- a/configure
+++ b/configure
@@ -3478,6 +3478,7 @@  openclsrc_filter_deps="opencl"
 overlay_opencl_filter_deps="opencl"
 overlay_qsv_filter_deps="libmfx"
 overlay_qsv_filter_select="qsvvpp"
+overlay_vaapi_filter_deps="vaapi"
 owdenoise_filter_deps="gpl"
 pan_filter_deps="swresample"
 perspective_filter_deps="gpl"
diff --git a/libavfilter/Makefile b/libavfilter/Makefile
index 07ea8d7edc..ccaad0d6a4 100644
--- a/libavfilter/Makefile
+++ b/libavfilter/Makefile
@@ -311,6 +311,7 @@  OBJS-$(CONFIG_OVERLAY_FILTER)                += vf_overlay.o framesync.o
 OBJS-$(CONFIG_OVERLAY_OPENCL_FILTER)         += vf_overlay_opencl.o opencl.o \
                                                 opencl/overlay.o framesync.o
 OBJS-$(CONFIG_OVERLAY_QSV_FILTER)            += vf_overlay_qsv.o framesync.o
+OBJS-$(CONFIG_OVERLAY_VAAPI_FILTER)          += vf_overlay_vaapi.o framesync.o vaapi_vpp.o
 OBJS-$(CONFIG_OWDENOISE_FILTER)              += vf_owdenoise.o
 OBJS-$(CONFIG_PAD_FILTER)                    += vf_pad.o
 OBJS-$(CONFIG_PALETTEGEN_FILTER)             += vf_palettegen.o
diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c
index 9c846b1ddd..27ee1df78b 100644
--- a/libavfilter/allfilters.c
+++ b/libavfilter/allfilters.c
@@ -295,6 +295,7 @@  extern AVFilter ff_vf_oscilloscope;
 extern AVFilter ff_vf_overlay;
 extern AVFilter ff_vf_overlay_opencl;
 extern AVFilter ff_vf_overlay_qsv;
+extern AVFilter ff_vf_overlay_vaapi;
 extern AVFilter ff_vf_owdenoise;
 extern AVFilter ff_vf_pad;
 extern AVFilter ff_vf_palettegen;
diff --git a/libavfilter/vf_overlay_vaapi.c b/libavfilter/vf_overlay_vaapi.c
new file mode 100644
index 0000000000..9fffa0fcb9
--- /dev/null
+++ b/libavfilter/vf_overlay_vaapi.c
@@ -0,0 +1,424 @@ 
+/*
+ * This file is part of FFmpeg.
+ *
+ * FFmpeg is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Lesser General Public
+ * License as published by the Free Software Foundation; either
+ * version 2.1 of the License, or (at your option) any later version.
+ *
+ * FFmpeg is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+ * Lesser General Public License for more details.
+ *
+ * You should have received a copy of the GNU Lesser General Public
+ * License along with FFmpeg; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
+ */
+#include <string.h>
+
+#include "libavutil/avassert.h"
+#include "libavutil/mem.h"
+#include "libavutil/opt.h"
+#include "libavutil/pixdesc.h"
+
+#include "avfilter.h"
+#include "framesync.h"
+#include "formats.h"
+#include "internal.h"
+#include "vaapi_vpp.h"
+
+typedef struct OverlayVAAPIContext {
+    VAAPIVPPContext  vpp_ctx; // must be the first field
+    FFFrameSync      fs;
+    int              overlay_x;
+    int              overlay_y;
+    int              overlay_w;
+    int              overlay_h;
+    float            overlay_alpha;
+} OverlayVAAPIContext;
+
+static int overlay_vaapi_query_formats(AVFilterContext *ctx)
+{
+    int i;
+    int ret;
+
+    static const enum AVPixelFormat main_in_fmts[] = {
+        AV_PIX_FMT_VAAPI,
+        AV_PIX_FMT_NONE
+    };
+    static const enum AVPixelFormat out_pix_fmts[] = {
+        AV_PIX_FMT_VAAPI,
+        AV_PIX_FMT_NONE
+    };
+
+    for (i = 0; i < ctx->nb_inputs; i++) {
+        ret = ff_formats_ref(ff_make_format_list(main_in_fmts), &ctx->inputs[i]->out_formats);
+        if (ret < 0)
+            return ret;
+    }
+
+    ret = ff_formats_ref(ff_make_format_list(out_pix_fmts), &ctx->outputs[0]->in_formats);
+    if (ret < 0)
+        return ret;
+
+    return 0;
+}
+
+static int overlay_vaapi_render(AVFilterContext *avctx,
+                                VAProcPipelineParameterBuffer *params,
+                                VAProcPipelineParameterBuffer *subpic_params,
+                                VASurfaceID output_surface)
+{
+    VABufferID params_id;
+    VABufferID subpic_params_id;
+    VAStatus vas;
+    int err = 0;
+    VAAPIVPPContext *ctx   = avctx->priv;
+
+    vas = vaBeginPicture(ctx->hwctx->display,
+                         ctx->va_context, output_surface);
+    if (vas != VA_STATUS_SUCCESS) {
+        av_log(avctx, AV_LOG_ERROR, "Failed to attach new picture: "
+               "%d (%s).\n", vas, vaErrorStr(vas));
+        err = AVERROR(EIO);
+        goto fail;
+    }
+
+    vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context,
+                         VAProcPipelineParameterBufferType,
+                         sizeof(*params), 1, params, &params_id);
+    if (vas != VA_STATUS_SUCCESS) {
+        av_log(avctx, AV_LOG_ERROR, "Failed to create parameter buffer: "
+               "%d (%s).\n", vas, vaErrorStr(vas));
+        err = AVERROR(EIO);
+        goto fail_after_begin;
+    }
+    av_log(avctx, AV_LOG_DEBUG, "Pipeline parameter buffer is %#x.\n",
+           params_id);
+
+
+    vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context,
+                         VAProcPipelineParameterBufferType,
+                         sizeof(*subpic_params), 1, subpic_params, &subpic_params_id);
+    if (vas != VA_STATUS_SUCCESS) {
+        av_log(avctx, AV_LOG_ERROR, "Failed to create parameter buffer: "
+               "%d (%s).\n", vas, vaErrorStr(vas));
+        err = AVERROR(EIO);
+        goto fail_after_begin;
+    }
+    av_log(avctx, AV_LOG_DEBUG, "Pipeline subpic parameter buffer is %#x.\n",
+           subpic_params_id);
+
+    vas = vaRenderPicture(ctx->hwctx->display, ctx->va_context,
+                          &params_id, 1);
+    if (vas != VA_STATUS_SUCCESS) {
+        av_log(avctx, AV_LOG_ERROR, "Failed to render parameter buffer: "
+               "%d (%s).\n", vas, vaErrorStr(vas));
+        err = AVERROR(EIO);
+        goto fail_after_begin;
+    }
+
+    vas = vaRenderPicture(ctx->hwctx->display, ctx->va_context,
+                          &subpic_params_id, 1);
+    if (vas != VA_STATUS_SUCCESS) {
+        av_log(avctx, AV_LOG_ERROR, "Failed to render subpic parameter buffer: "
+               "%d (%s).\n", vas, vaErrorStr(vas));
+        err = AVERROR(EIO);
+        goto fail_after_begin;
+    }
+
+    vas = vaEndPicture(ctx->hwctx->display, ctx->va_context);
+    if (vas != VA_STATUS_SUCCESS) {
+        av_log(avctx, AV_LOG_ERROR, "Failed to start picture processing: "
+               "%d (%s).\n", vas, vaErrorStr(vas));
+        err = AVERROR(EIO);
+        goto fail_after_render;
+    }
+
+    if (CONFIG_VAAPI_1 || ctx->hwctx->driver_quirks &
+        AV_VAAPI_DRIVER_QUIRK_RENDER_PARAM_BUFFERS) {
+        vas = vaDestroyBuffer(ctx->hwctx->display, params_id);
+        if (vas != VA_STATUS_SUCCESS) {
+            av_log(avctx, AV_LOG_ERROR, "Failed to free parameter buffer: "
+                   "%d (%s).\n", vas, vaErrorStr(vas));
+            // And ignore.
+        }
+    }
+
+    return 0;
+
+    // We want to make sure that if vaBeginPicture has been called, we also
+    // call vaRenderPicture and vaEndPicture.  These calls may well fail or
+    // do something else nasty, but once we're in this failure case there
+    // isn't much else we can do.
+fail_after_begin:
+    vaRenderPicture(ctx->hwctx->display, ctx->va_context, &params_id, 1);
+fail_after_render:
+    vaEndPicture(ctx->hwctx->display, ctx->va_context);
+fail:
+    return err;
+}
+
+static int overlay_vaapi_blend(FFFrameSync *fs)
+{
+    AVFilterContext    *avctx = fs->parent;
+    AVFilterLink     *outlink = avctx->outputs[0];
+    OverlayVAAPIContext *ctx  = avctx->priv;
+    VAAPIVPPContext *vpp_ctx  = avctx->priv;
+    AVFrame *input_main, *input_overlay;
+    AVFrame *output;
+    int err;
+    VASurfaceID main_surface, overlay_surface, output_surface;
+    VARectangle overlay_region, output_region;
+    VAProcPipelineParameterBuffer params, subpic_params;
+
+#if VA_CHECK_VERSION(2, 1, 0)
+    VABlendState blend_state; // Blend State
+#endif
+
+    err = ff_framesync_get_frame(fs, 0, &input_main, 0);
+    if (err < 0)
+        return err;
+    err = ff_framesync_get_frame(fs, 1, &input_overlay, 0);
+    if (err < 0)
+        return err;
+
+    av_log(avctx, AV_LOG_DEBUG, "Filter main: %s, %ux%u (%"PRId64").\n",
+           av_get_pix_fmt_name(input_main->format),
+           input_main->width, input_main->height, input_main->pts);
+
+    av_log(avctx, AV_LOG_DEBUG, "Filter overlay: %s, %ux%u (%"PRId64").\n",
+           av_get_pix_fmt_name(input_overlay->format),
+           input_overlay->width, input_overlay->height, input_overlay->pts);
+
+    if (vpp_ctx->va_context == VA_INVALID_ID)
+        return AVERROR(EINVAL);
+
+    main_surface = (VASurfaceID)(uintptr_t)input_main->data[3];
+    av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp main.\n",
+           main_surface);
+
+    overlay_surface = (VASurfaceID)(uintptr_t)input_overlay->data[3];
+    av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp overlay.\n",
+           overlay_surface);
+
+    output = ff_get_video_buffer(outlink, outlink->w, outlink->h);
+    if (!output) {
+        err = AVERROR(ENOMEM);
+        goto fail;
+    }
+
+    output_surface = (VASurfaceID)(uintptr_t)output->data[3];
+    av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp output.\n",
+           output_surface);
+
+    memset(&params, 0, sizeof(params));
+    memset(&subpic_params, 0, sizeof(subpic_params));
+
+    overlay_region = (VARectangle) {
+        .x      = ctx->overlay_x,
+        .y      = ctx->overlay_y,
+        .width  = input_overlay->width,
+        .height = input_overlay->height,
+    };
+
+    output_region = (VARectangle) {
+        .x      = 0,
+        .y      = 0,
+        .width  = output->width,
+        .height = output->height,
+    };
+
+    if (ctx->overlay_w && ctx->overlay_h) {
+        overlay_region.width = ctx->overlay_w;
+        overlay_region.height = ctx->overlay_h;
+    }
+
+    if (overlay_region.width > input_main->width ||
+        overlay_region.height > input_main->height) {
+        av_log(ctx, AV_LOG_ERROR,
+               "Overlay width > main width or overlay height > main height is not supported.\n");
+        err =  AVERROR(EINVAL);
+        goto fail;
+    }
+
+    params.filters     = &vpp_ctx->filter_buffers[0];
+    params.num_filters = vpp_ctx->nb_filter_buffers;
+
+    err = ff_vaapi_vpp_init_params(avctx, &params,
+                                   input_main, output);
+    if (err < 0)
+        goto fail;
+
+    params.output_region = &output_region;
+    params.output_background_color = VAAPI_VPP_BACKGROUND_BLACK;
+    params.output_color_standard = params.surface_color_standard;
+
+    memcpy(&subpic_params, &params, sizeof(subpic_params));
+
+#if VA_CHECK_VERSION(2, 1, 0)
+    blend_state.flag = VA_BLEND_GLOBAL_ALPHA;
+    blend_state.global_alpha = ctx->overlay_alpha;
+    subpic_params.blend_state = &blend_state;
+#endif
+
+    subpic_params.surface = overlay_surface;
+    subpic_params.output_region = &overlay_region;
+
+    err = overlay_vaapi_render(avctx, &params, &subpic_params, output_surface);
+    if (err < 0)
+        goto fail;
+
+    err = av_frame_copy_props(output, input_main);
+    if (err < 0)
+        goto fail;
+
+    av_log(avctx, AV_LOG_DEBUG, "Filter output: %s, %ux%u (%"PRId64").\n",
+           av_get_pix_fmt_name(output->format),
+           output->width, output->height, output->pts);
+
+    return ff_filter_frame(outlink, output);
+
+fail:
+    av_frame_free(&output);
+    return err;
+}
+
+static int overlay_vaapi_init_framesync(AVFilterContext *avctx)
+{
+    OverlayVAAPIContext *ctx = avctx->priv;
+    int ret, i;
+
+    ctx->fs.on_event = overlay_vaapi_blend;
+    ctx->fs.opaque   = ctx;
+    ret = ff_framesync_init(&ctx->fs, avctx, avctx->nb_inputs);
+    if (ret < 0)
+        return ret;
+
+    for (i = 0; i < avctx->nb_inputs; i++) {
+        FFFrameSyncIn *in = &ctx->fs.in[i];
+        in->before    = EXT_STOP;
+        in->after     = EXT_INFINITY;
+        in->sync      = i ? 1 : 2;
+        in->time_base = avctx->inputs[i]->time_base;
+    }
+
+    return ff_framesync_configure(&ctx->fs);
+}
+
+static int overlay_vaapi_config_output(AVFilterLink *outlink)
+{
+    AVFilterContext  *avctx  = outlink->src;
+    OverlayVAAPIContext *ctx = avctx->priv;
+    VAAPIVPPContext *vpp_ctx = avctx->priv;
+    AVFilterLink        *in0 = avctx->inputs[0];
+    AVFilterLink        *in1 = avctx->inputs[1];
+    int err;
+
+    av_log(ctx, AV_LOG_DEBUG, "Output is of %s.\n", av_get_pix_fmt_name(outlink->format));
+    if ((in0->format == AV_PIX_FMT_VAAPI && in1->format != AV_PIX_FMT_VAAPI) ||
+        (in0->format != AV_PIX_FMT_VAAPI && in1->format == AV_PIX_FMT_VAAPI)) {
+        av_log(ctx, AV_LOG_ERROR, "Mixing hardware and software pixel formats is not supported.\n");
+        return AVERROR(EINVAL);
+    }
+
+    err = overlay_vaapi_init_framesync(avctx);
+    if (err < 0)
+        return err;
+
+    vpp_ctx->output_width  = avctx->inputs[0]->w;
+    vpp_ctx->output_height = avctx->inputs[0]->h;
+
+    err = ff_vaapi_vpp_config_output(outlink);
+    if (err < 0)
+        return err;
+
+    err = ff_framesync_init_dualinput(&ctx->fs, avctx);
+    if (err < 0)
+        return err;
+
+    return ff_framesync_configure(&ctx->fs);
+}
+
+static av_cold int overlay_vaapi_init(AVFilterContext *avctx)
+{
+    VAAPIVPPContext *vpp_ctx = avctx->priv;
+
+    ff_vaapi_vpp_ctx_init(avctx);
+    vpp_ctx->output_format = AV_PIX_FMT_NONE;
+
+    return 0;
+}
+
+static int overlay_vaapi_activate(AVFilterContext *avctx)
+{
+    OverlayVAAPIContext *ctx = avctx->priv;
+
+    return ff_framesync_activate(&ctx->fs);
+}
+
+static av_cold void overlay_vaapi_uninit(AVFilterContext *avctx)
+{
+    OverlayVAAPIContext *ctx = avctx->priv;
+
+    ff_framesync_uninit(&ctx->fs);
+}
+
+#define OFFSET(x) offsetof(OverlayVAAPIContext, x)
+#define FLAGS (AV_OPT_FLAG_FILTERING_PARAM | AV_OPT_FLAG_VIDEO_PARAM)
+static const AVOption overlay_vaapi_options[] = {
+    { "x", "Overlay x position",
+      OFFSET(overlay_x), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = FLAGS },
+    { "y", "Overlay y position",
+      OFFSET(overlay_y), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = FLAGS },
+    { "w", "Overlay width",
+      OFFSET(overlay_w), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = FLAGS },
+    { "h", "Overlay hight",
+      OFFSET(overlay_h), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = FLAGS },
+    { "alpha", "Overlay global alpha",
+      OFFSET(overlay_alpha), AV_OPT_TYPE_FLOAT, { .dbl = 0.0}, 0.0, 1.0, .flags = FLAGS},
+    { NULL },
+};
+
+AVFILTER_DEFINE_CLASS(overlay_vaapi);
+
+static const AVFilterPad overlay_vaapi_inputs[] = {
+    {
+        .name             = "main",
+        .type             = AVMEDIA_TYPE_VIDEO,
+        .get_video_buffer = ff_default_get_video_buffer,
+        .config_props     = &ff_vaapi_vpp_config_input,
+        .needs_fifo       = 1,
+    },
+    {
+        .name             = "overlay",
+        .type             = AVMEDIA_TYPE_VIDEO,
+        .get_video_buffer = ff_default_get_video_buffer,
+        .needs_fifo       = 1,
+    },
+    { NULL }
+};
+
+static const AVFilterPad overlay_vaapi_outputs[] = {
+    {
+        .name          = "default",
+        .type          = AVMEDIA_TYPE_VIDEO,
+        .config_props  = &overlay_vaapi_config_output,
+    },
+    { NULL }
+};
+
+AVFilter ff_vf_overlay_vaapi = {
+    .name            = "overlay_vaapi",
+    .description     = NULL_IF_CONFIG_SMALL("Overlay one video on top of another"),
+    .priv_size       = sizeof(OverlayVAAPIContext),
+    .priv_class      = &overlay_vaapi_class,
+    .init            = &overlay_vaapi_init,
+    .uninit          = &overlay_vaapi_uninit,
+    .query_formats   = &overlay_vaapi_query_formats,
+    .activate        = &overlay_vaapi_activate,
+    .inputs          = overlay_vaapi_inputs,
+    .outputs         = overlay_vaapi_outputs,
+    .flags_internal  = FF_FILTER_FLAG_HWFRAME_AWARE,
+};