Message ID | 20190620074228.25159-1-zachary.zhou@intel.com |
---|---|
State | Superseded |
Headers | show |
On 20/06/2019 08:42, Zachary Zhou wrote: > --- > configure | 1 + > libavfilter/Makefile | 1 + > libavfilter/allfilters.c | 1 + > libavfilter/vaapi_vpp.c | 95 +++++++++ > libavfilter/vaapi_vpp.h | 5 + > libavfilter/vf_overlay_vaapi.c | 365 +++++++++++++++++++++++++++++++++ > 6 files changed, 468 insertions(+) > create mode 100644 libavfilter/vf_overlay_vaapi.c > > diff --git a/configure b/configure > index 32fc26356c..f469e6a3b1 100755 > --- a/configure > +++ b/configure > @@ -3478,6 +3478,7 @@ openclsrc_filter_deps="opencl" > overlay_opencl_filter_deps="opencl" > overlay_qsv_filter_deps="libmfx" > overlay_qsv_filter_select="qsvvpp" > +overlay_vaapi_filter_deps="vaapi" > owdenoise_filter_deps="gpl" > pan_filter_deps="swresample" > perspective_filter_deps="gpl" > diff --git a/libavfilter/Makefile b/libavfilter/Makefile > index 07ea8d7edc..5cbf1a7e41 100644 > --- a/libavfilter/Makefile > +++ b/libavfilter/Makefile > @@ -311,6 +311,7 @@ OBJS-$(CONFIG_OVERLAY_FILTER) += vf_overlay.o framesync.o > OBJS-$(CONFIG_OVERLAY_OPENCL_FILTER) += vf_overlay_opencl.o opencl.o \ > opencl/overlay.o framesync.o > OBJS-$(CONFIG_OVERLAY_QSV_FILTER) += vf_overlay_qsv.o framesync.o > +OBJS-$(CONFIG_OVERLAY_VAAPI_FILTER) += vf_overlay_vaapi.o framesync.o and vaapi_vpp.o? (Try building the filter on its own without anything else enabled.) > OBJS-$(CONFIG_OWDENOISE_FILTER) += vf_owdenoise.o > OBJS-$(CONFIG_PAD_FILTER) += vf_pad.o > OBJS-$(CONFIG_PALETTEGEN_FILTER) += vf_palettegen.o > diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c > index 9c846b1ddd..27ee1df78b 100644 > --- a/libavfilter/allfilters.c > +++ b/libavfilter/allfilters.c > @@ -295,6 +295,7 @@ extern AVFilter ff_vf_oscilloscope; > extern AVFilter ff_vf_overlay; > extern AVFilter ff_vf_overlay_opencl; > extern AVFilter ff_vf_overlay_qsv; > +extern AVFilter ff_vf_overlay_vaapi; > extern AVFilter ff_vf_owdenoise; > extern AVFilter ff_vf_pad; > extern AVFilter ff_vf_palettegen; > diff --git a/libavfilter/vaapi_vpp.c b/libavfilter/vaapi_vpp.c > index b5b245c8af..a8caa5b532 100644 > --- a/libavfilter/vaapi_vpp.c > +++ b/libavfilter/vaapi_vpp.c > @@ -663,6 +663,101 @@ fail: > return err; > } > > +int ff_vaapi_vpp_render_overlay(AVFilterContext *avctx, > + VAProcPipelineParameterBuffer *params, > + VAProcPipelineParameterBuffer *subpic_params, > + VASurfaceID output_surface) > +{ > + VABufferID params_id; > + VABufferID subpic_params_id; > + VAStatus vas; > + int err = 0; > + VAAPIVPPContext *ctx = avctx->priv; > + > + vas = vaBeginPicture(ctx->hwctx->display, > + ctx->va_context, output_surface); > + if (vas != VA_STATUS_SUCCESS) { > + av_log(avctx, AV_LOG_ERROR, "Failed to attach new picture: " > + "%d (%s).\n", vas, vaErrorStr(vas)); > + err = AVERROR(EIO); > + goto fail; > + } > + > + vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context, > + VAProcPipelineParameterBufferType, > + sizeof(*params), 1, params, ¶ms_id); > + if (vas != VA_STATUS_SUCCESS) { > + av_log(avctx, AV_LOG_ERROR, "Failed to create parameter buffer: " > + "%d (%s).\n", vas, vaErrorStr(vas)); > + err = AVERROR(EIO); > + goto fail_after_begin; > + } > + av_log(avctx, AV_LOG_DEBUG, "Pipeline parameter buffer is %#x.\n", > + params_id); > + > + > + vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context, > + VAProcPipelineParameterBufferType, > + sizeof(*subpic_params), 1, subpic_params, &subpic_params_id); > + if (vas != VA_STATUS_SUCCESS) { > + av_log(avctx, AV_LOG_ERROR, "Failed to create parameter buffer: " > + "%d (%s).\n", vas, vaErrorStr(vas)); > + err = AVERROR(EIO); > + goto fail_after_begin; > + } > + av_log(avctx, AV_LOG_DEBUG, "Pipeline subpic parameter buffer is %#x.\n", > + subpic_params_id); > + > + vas = vaRenderPicture(ctx->hwctx->display, ctx->va_context, > + ¶ms_id, 1); > + if (vas != VA_STATUS_SUCCESS) { > + av_log(avctx, AV_LOG_ERROR, "Failed to render parameter buffer: " > + "%d (%s).\n", vas, vaErrorStr(vas)); > + err = AVERROR(EIO); > + goto fail_after_begin; > + } > + > + vas = vaRenderPicture(ctx->hwctx->display, ctx->va_context, > + &subpic_params_id, 1); > + if (vas != VA_STATUS_SUCCESS) { > + av_log(avctx, AV_LOG_ERROR, "Failed to render subpic parameter buffer: " > + "%d (%s).\n", vas, vaErrorStr(vas)); > + err = AVERROR(EIO); > + goto fail_after_begin; > + } > + > + vas = vaEndPicture(ctx->hwctx->display, ctx->va_context); > + if (vas != VA_STATUS_SUCCESS) { > + av_log(avctx, AV_LOG_ERROR, "Failed to start picture processing: " > + "%d (%s).\n", vas, vaErrorStr(vas)); > + err = AVERROR(EIO); > + goto fail_after_render; > + } > + > + if (CONFIG_VAAPI_1 || ctx->hwctx->driver_quirks & > + AV_VAAPI_DRIVER_QUIRK_RENDER_PARAM_BUFFERS) { > + vas = vaDestroyBuffer(ctx->hwctx->display, params_id); > + if (vas != VA_STATUS_SUCCESS) { > + av_log(avctx, AV_LOG_ERROR, "Failed to free parameter buffer: " > + "%d (%s).\n", vas, vaErrorStr(vas)); > + // And ignore. > + } > + } > + > + return 0; > + > + // We want to make sure that if vaBeginPicture has been called, we also > + // call vaRenderPicture and vaEndPicture. These calls may well fail or > + // do something else nasty, but once we're in this failure case there > + // isn't much else we can do. > +fail_after_begin: > + vaRenderPicture(ctx->hwctx->display, ctx->va_context, ¶ms_id, 1); > +fail_after_render: > + vaEndPicture(ctx->hwctx->display, ctx->va_context); > +fail: > + return err; > +} > + > void ff_vaapi_vpp_ctx_init(AVFilterContext *avctx) > { > int i; > diff --git a/libavfilter/vaapi_vpp.h b/libavfilter/vaapi_vpp.h > index c3da91717c..8d53acdec9 100644 > --- a/libavfilter/vaapi_vpp.h > +++ b/libavfilter/vaapi_vpp.h > @@ -83,4 +83,9 @@ int ff_vaapi_vpp_render_picture(AVFilterContext *avctx, > VAProcPipelineParameterBuffer *params, > AVFrame *output_frame); > > +int ff_vaapi_vpp_render_overlay(AVFilterContext *avctx, > + VAProcPipelineParameterBuffer *primary_params, > + VAProcPipelineParameterBuffer *subpic_params, > + VASurfaceID output_surface); In the absence of a second user I don't think it's worth putting this in the generic file. > + > #endif /* AVFILTER_VAAPI_VPP_H */ > diff --git a/libavfilter/vf_overlay_vaapi.c b/libavfilter/vf_overlay_vaapi.c > new file mode 100644 > index 0000000000..bcea76a806 > --- /dev/null > +++ b/libavfilter/vf_overlay_vaapi.c > @@ -0,0 +1,365 @@ > +/* > + * This file is part of FFmpeg. > + * > + * FFmpeg is free software; you can redistribute it and/or > + * modify it under the terms of the GNU Lesser General Public > + * License as published by the Free Software Foundation; either > + * version 2.1 of the License, or (at your option) any later version. > + * > + * FFmpeg is distributed in the hope that it will be useful, > + * but WITHOUT ANY WARRANTY; without even the implied warranty of > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU > + * Lesser General Public License for more details. > + * > + * You should have received a copy of the GNU Lesser General Public > + * License along with FFmpeg; if not, write to the Free Software > + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA > + */ > +#include <string.h> > + > +#include "libavutil/avassert.h" > +#include "libavutil/mem.h" > +#include "libavutil/opt.h" > +#include "libavutil/pixdesc.h" > + > +#include "avfilter.h" > +#include "framesync.h" > +#include "formats.h" > +#include "internal.h" > +#include "vaapi_vpp.h" > + > +#if VA_CHECK_VERSION(2, 1, 0) > +// Blend State > +static VABlendState blend_state = { /** \brief Video blending flags. */ Static data in libraries must be const. > + .flags = VA_BLEND_PREMULTIPLIED_ALPHA, //VA_BLEND_GLOBAL_ALPHA, > + .global_alpha = 1, > + .min_luma = 0, > + .max_luma = 1 These are all floats, use 0.0f and 1.0f to avoid confusion. > +}; > +#endif > + > +typedef struct OverlayVAAPIContext { > + VAAPIVPPContext vpp_ctx; // must be the first field > + > + FFFrameSync fs; > + > + int overlay_x; > + int overlay_y; > + int overlay_w; > + int overlay_h; > + float overlay_alpha; > +} OverlayVAAPIContext; > + > +static int overlay_vaapi_query_formats(AVFilterContext *ctx) > +{ > + int i; > + int ret; > + > + static const enum AVPixelFormat main_in_fmts[] = { > + AV_PIX_FMT_NV12, > + AV_PIX_FMT_RGB32, > + AV_PIX_FMT_VAAPI, > + AV_PIX_FMT_NONE > + }; > + static const enum AVPixelFormat out_pix_fmts[] = { > + AV_PIX_FMT_NV12, > + AV_PIX_FMT_RGB32, > + AV_PIX_FMT_VAAPI, > + AV_PIX_FMT_NONE > + }; I don't see how this can handle the software pixfmts? > + > + for (i = 0; i < ctx->nb_inputs; i++) { > + ret = ff_formats_ref(ff_make_format_list(main_in_fmts), &ctx->inputs[i]->out_formats); > + if (ret < 0) > + return ret; > + } > + > + ret = ff_formats_ref(ff_make_format_list(out_pix_fmts), &ctx->outputs[0]->in_formats); > + if (ret < 0) > + return ret; > + > + return 0; > +} > + > +static int overlay_vaapi_blend(FFFrameSync *fs) > +{ > + AVFilterContext *avctx = fs->parent; > + AVFilterLink *outlink = avctx->outputs[0]; > + OverlayVAAPIContext *ctx = avctx->priv; > + VAAPIVPPContext *vpp_ctx = avctx->priv; > + AVFrame *input_main, *input_overlay; > + AVFrame *output; > + int err; > + VASurfaceID main_surface, overlay_surface, output_surface; > + VARectangle main_region, overlay_region, output_region; > + VAProcPipelineParameterBuffer params, subpic_params; > + > + err = ff_framesync_get_frame(fs, 0, &input_main, 0); > + if (err < 0) > + return err; > + err = ff_framesync_get_frame(fs, 1, &input_overlay, 0); > + if (err < 0) > + return err; > + > + av_log(avctx, AV_LOG_DEBUG, "Filter main: %s, %ux%u (%"PRId64").\n", > + av_get_pix_fmt_name(input_main->format), > + input_main->width, input_main->height, input_main->pts); > + > + av_log(avctx, AV_LOG_DEBUG, "Filter overlay: %s, %ux%u (%"PRId64").\n", > + av_get_pix_fmt_name(input_overlay->format), > + input_overlay->width, input_overlay->height, input_overlay->pts); > + > + if (vpp_ctx->va_context == VA_INVALID_ID) > + return AVERROR(EINVAL); > + > + main_surface = (VASurfaceID)(uintptr_t)input_main->data[3]; > + av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp main.\n", > + main_surface); > + > + overlay_surface = (VASurfaceID)(uintptr_t)input_overlay->data[3]; > + av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp overlay.\n", > + overlay_surface); > + > + output = ff_get_video_buffer(outlink, outlink->w, outlink->h); > + if (!output) { > + err = AVERROR(ENOMEM); > + goto fail; > + } > + > + output_surface = (VASurfaceID)(uintptr_t)output->data[3]; > + av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp output.\n", > + output_surface); > + > + memset(¶ms, 0, sizeof(params)); > + memset(&subpic_params, 0, sizeof(subpic_params)); > + > + main_region = (VARectangle) { > + .x = 0, > + .y = 0, > + .width = input_main->width, > + .height = input_main->height, > + }; > + > + overlay_region = (VARectangle) { > + .x = ctx->overlay_x, > + .y = ctx->overlay_y, > + .width = input_overlay->width, > + .height = input_overlay->height, > + }; > + > + output_region = (VARectangle) { > + .x = 0, > + .y = 0, > + .width = output->width, > + .height = output->height, > + }; > + > + if (!ctx->overlay_w && !ctx->overlay_h) { > + overlay_region.width = ctx->overlay_w; > + overlay_region.height = ctx->overlay_h; > + } > + > + if (overlay_region.width > main_region.width || > + overlay_region.height > main_region.height) { > + av_log(ctx, AV_LOG_ERROR, > + "Overlay width > main width or overlay height > main height is not supported.\n"); > + err = AVERROR(EINVAL); > + goto fail; > + } What about overlay_region.x + overlay_region.with > main_region.width - does that work? > + > + params.filters = &vpp_ctx->filter_buffers[0]; > + params.num_filters = vpp_ctx->nb_filter_buffers; > + > + params.surface = main_surface; > + params.surface_region = &main_region; > + > + switch (input_main->colorspace) { > + case AVCOL_SPC_BT470BG: > + params.surface_color_standard = VAProcColorStandardBT601; > + break; > + case AVCOL_SPC_BT709: > + params.surface_color_standard = VAProcColorStandardBT709; > + break; > + default: > + params.surface_color_standard = VAProcColorStandardNone; > + break; > + } We have better support for colours - see ff_vaapi_vpp_init_params(). > + > + params.output_region = &output_region; > + params.output_background_color = VAAPI_VPP_BACKGROUND_BLACK; > + params.output_color_standard = params.surface_color_standard; > + > + memcpy(&subpic_params, ¶ms, sizeof(subpic_params)); > + subpic_params.pipeline_flags |= VA_PROC_PIPELINE_FAST; > + subpic_params.filter_flags |= VA_FILTER_SCALING_FAST; Why are these flags set? > + > +#if VA_CHECK_VERSION(2, 1, 0) > + blend_state.global_alpha = ctx->overlay_alpha; > + subpic_params.blend_state = &blend_state; > +#endif > + > + subpic_params.surface = overlay_surface; > + subpic_params.output_region = &overlay_region; > + > + err = ff_vaapi_vpp_render_overlay(avctx, ¶ms, &subpic_params, output_surface); > + if (err < 0) > + goto fail; > + > + err = av_frame_copy_props(output, input_main); > + if (err < 0) > + goto fail; > + > + av_log(avctx, AV_LOG_DEBUG, "Filter output: %s, %ux%u (%"PRId64").\n", > + av_get_pix_fmt_name(output->format), > + output->width, output->height, output->pts); > + > + return ff_filter_frame(outlink, output); > + > +fail: > + av_frame_free(&output); > + return err; > +} > + > +static int overlay_vaapi_init_framesync(AVFilterContext *avctx) > +{ > + OverlayVAAPIContext *ctx = avctx->priv; > + int ret, i; > + > + ctx->fs.on_event = overlay_vaapi_blend; > + ctx->fs.opaque = ctx; > + ret = ff_framesync_init(&ctx->fs, avctx, avctx->nb_inputs); > + if (ret < 0) > + return ret; > + > + for (i = 0; i < avctx->nb_inputs; i++) { > + FFFrameSyncIn *in = &ctx->fs.in[i]; > + in->before = EXT_STOP; > + in->after = EXT_INFINITY; > + in->sync = i ? 1 : 2; > + in->time_base = avctx->inputs[i]->time_base; > + } > + > + return ff_framesync_configure(&ctx->fs); > +} > + > +static int overlay_vaapi_config_output(AVFilterLink *outlink) > +{ > + AVFilterContext *avctx = outlink->src; > + OverlayVAAPIContext *ctx = avctx->priv; > + VAAPIVPPContext *vpp_ctx = avctx->priv; > + AVFilterLink *in0 = avctx->inputs[0]; > + AVFilterLink *in1 = avctx->inputs[1]; > + int err; > + > + av_log(ctx, AV_LOG_DEBUG, "Output is of %s.\n", av_get_pix_fmt_name(outlink->format)); > + if ((in0->format == AV_PIX_FMT_VAAPI && in1->format != AV_PIX_FMT_VAAPI) || > + (in0->format != AV_PIX_FMT_VAAPI && in1->format == AV_PIX_FMT_VAAPI)) { > + av_log(ctx, AV_LOG_ERROR, "Mixing hardware and software pixel formats is not supported.\n"); Software pixel formats don't appear to be supported at all, though? > + return AVERROR(EINVAL); > + } > + > + err = overlay_vaapi_init_framesync(avctx); > + if (err < 0) > + return err; > + > + vpp_ctx->output_width = avctx->inputs[0]->w; > + vpp_ctx->output_height = avctx->inputs[0]->h; > + > + err = ff_vaapi_vpp_config_output(outlink); > + if (err < 0) > + return err; Is it ok to copy the format of the output frames from the last-initialised input, as this does? What happens if the inputs aren't in the same format? > + > + err = ff_framesync_init_dualinput(&ctx->fs, avctx); > + if (err < 0) > + return err; > + > + return ff_framesync_configure(&ctx->fs); > +} > + > +static av_cold int overlay_vaapi_init(AVFilterContext *avctx) > +{ > + VAAPIVPPContext *vpp_ctx = avctx->priv; > + > + ff_vaapi_vpp_ctx_init(avctx); > + vpp_ctx->output_format = AV_PIX_FMT_NONE; > + > + return 0; > +} > + > +static int overlay_vaapi_activate(AVFilterContext *avctx) > +{ > + OverlayVAAPIContext *ctx = avctx->priv; > + > + return ff_framesync_activate(&ctx->fs); > +} > + > +static av_cold void overlay_vaapi_uninit(AVFilterContext *avctx) > +{ > + OverlayVAAPIContext *ctx = avctx->priv; > + > + ff_framesync_uninit(&ctx->fs); > +} > + > +static AVFrame *get_video_buffer(AVFilterLink *inlink, int w, int h) > +{ > + return ff_default_get_video_buffer(inlink, w, h); > +} I don't think this function is needed - the ff_default_get_video_buffer() is already the default if it isn't set. > + > +#define OFFSET(x) offsetof(OverlayVAAPIContext, x) > +#define FLAGS (AV_OPT_FLAG_FILTERING_PARAM | AV_OPT_FLAG_VIDEO_PARAM) > +static const AVOption overlay_vaapi_options[] = { > + { "x", "Overlay x position", > + OFFSET(overlay_x), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = FLAGS }, > + { "y", "Overlay y position", > + OFFSET(overlay_y), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = FLAGS }, > + { "w", "Overlay width", > + OFFSET(overlay_w), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = FLAGS }, > + { "h", "Overlay hight", > + OFFSET(overlay_h), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = FLAGS }, > + { "alpha", "Overlay global alpha", > + OFFSET(overlay_alpha), AV_OPT_TYPE_FLOAT, { .dbl = 0.0}, 0.0, 1.0, .flags = FLAGS}, > + { NULL }, > +}; > + > +AVFILTER_DEFINE_CLASS(overlay_vaapi); > + > +static const AVFilterPad overlay_vaapi_inputs[] = { > + { > + .name = "main", > + .type = AVMEDIA_TYPE_VIDEO, > + .get_video_buffer = get_video_buffer, > + .config_props = &ff_vaapi_vpp_config_input, > + .needs_fifo = 1, > + }, > + { > + .name = "overlay", > + .type = AVMEDIA_TYPE_VIDEO, > + .get_video_buffer = get_video_buffer, > + .config_props = &ff_vaapi_vpp_config_input, > + .needs_fifo = 1, > + }, Calling ff_vaapi_vpp_config_input() twice doesn't look like something you wanted to do. In particular, it leaks a reference to an inlink hw_frames_ctx. > + { NULL } > +}; > + > +static const AVFilterPad overlay_vaapi_outputs[] = { > + { > + .name = "default", > + .type = AVMEDIA_TYPE_VIDEO, > + .config_props = &overlay_vaapi_config_output, > + }, > + { NULL } > +}; > + > +AVFilter ff_vf_overlay_vaapi = { > + .name = "overlay_vaapi", > + .description = NULL_IF_CONFIG_SMALL("Overlay one video on top of another"), > + .priv_size = sizeof(OverlayVAAPIContext), > + .priv_class = &overlay_vaapi_class, > + .init = &overlay_vaapi_init, > + .uninit = &overlay_vaapi_uninit, > + .query_formats = &overlay_vaapi_query_formats, > + .activate = &overlay_vaapi_activate, > + .inputs = overlay_vaapi_inputs, > + .outputs = overlay_vaapi_outputs, > + .flags_internal = FF_FILTER_FLAG_HWFRAME_AWARE, > +}; > - Mark
> -----Original Message----- > From: ffmpeg-devel [mailto:ffmpeg-devel-bounces@ffmpeg.org] On Behalf Of > Mark Thompson > Sent: Sunday, July 7, 2019 8:56 PM > To: ffmpeg-devel@ffmpeg.org > Subject: Re: [FFmpeg-devel] [PATCH v3] avfilter/vaapi: add overlay_vaapi filter > > On 20/06/2019 08:42, Zachary Zhou wrote: > > --- > > configure | 1 + > > libavfilter/Makefile | 1 + > > libavfilter/allfilters.c | 1 + > > libavfilter/vaapi_vpp.c | 95 +++++++++ > > libavfilter/vaapi_vpp.h | 5 + > > libavfilter/vf_overlay_vaapi.c | 365 > > +++++++++++++++++++++++++++++++++ > > 6 files changed, 468 insertions(+) > > create mode 100644 libavfilter/vf_overlay_vaapi.c > > > > diff --git a/configure b/configure > > index 32fc26356c..f469e6a3b1 100755 > > --- a/configure > > +++ b/configure > > @@ -3478,6 +3478,7 @@ openclsrc_filter_deps="opencl" > > overlay_opencl_filter_deps="opencl" > > overlay_qsv_filter_deps="libmfx" > > overlay_qsv_filter_select="qsvvpp" > > +overlay_vaapi_filter_deps="vaapi" > > owdenoise_filter_deps="gpl" > > pan_filter_deps="swresample" > > perspective_filter_deps="gpl" > > diff --git a/libavfilter/Makefile b/libavfilter/Makefile index > > 07ea8d7edc..5cbf1a7e41 100644 > > --- a/libavfilter/Makefile > > +++ b/libavfilter/Makefile > > @@ -311,6 +311,7 @@ OBJS-$(CONFIG_OVERLAY_FILTER) += > vf_overlay.o framesync.o > > OBJS-$(CONFIG_OVERLAY_OPENCL_FILTER) += vf_overlay_opencl.o > opencl.o \ > > opencl/overlay.o framesync.o > > OBJS-$(CONFIG_OVERLAY_QSV_FILTER) += vf_overlay_qsv.o > framesync.o > > +OBJS-$(CONFIG_OVERLAY_VAAPI_FILTER) += vf_overlay_vaapi.o > framesync.o > > and vaapi_vpp.o? will add vaapi_vpp.o to the Makefile > > (Try building the filter on its own without anything else enabled.) > > > OBJS-$(CONFIG_OWDENOISE_FILTER) += vf_owdenoise.o > > OBJS-$(CONFIG_PAD_FILTER) += vf_pad.o > > OBJS-$(CONFIG_PALETTEGEN_FILTER) += vf_palettegen.o > > diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c index > > 9c846b1ddd..27ee1df78b 100644 > > --- a/libavfilter/allfilters.c > > +++ b/libavfilter/allfilters.c > > @@ -295,6 +295,7 @@ extern AVFilter ff_vf_oscilloscope; extern > > AVFilter ff_vf_overlay; extern AVFilter ff_vf_overlay_opencl; extern > > AVFilter ff_vf_overlay_qsv; > > +extern AVFilter ff_vf_overlay_vaapi; > > extern AVFilter ff_vf_owdenoise; > > extern AVFilter ff_vf_pad; > > extern AVFilter ff_vf_palettegen; > > diff --git a/libavfilter/vaapi_vpp.c b/libavfilter/vaapi_vpp.c index > > b5b245c8af..a8caa5b532 100644 > > --- a/libavfilter/vaapi_vpp.c > > +++ b/libavfilter/vaapi_vpp.c > > @@ -663,6 +663,101 @@ fail: > > return err; > > } > > > > +int ff_vaapi_vpp_render_overlay(AVFilterContext *avctx, > > + VAProcPipelineParameterBuffer *params, > > + VAProcPipelineParameterBuffer *subpic_params, > > + VASurfaceID output_surface) { > > + VABufferID params_id; > > + VABufferID subpic_params_id; > > + VAStatus vas; > > + int err = 0; > > + VAAPIVPPContext *ctx = avctx->priv; > > + > > + vas = vaBeginPicture(ctx->hwctx->display, > > + ctx->va_context, output_surface); > > + if (vas != VA_STATUS_SUCCESS) { > > + av_log(avctx, AV_LOG_ERROR, "Failed to attach new picture: " > > + "%d (%s).\n", vas, vaErrorStr(vas)); > > + err = AVERROR(EIO); > > + goto fail; > > + } > > + > > + vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context, > > + VAProcPipelineParameterBufferType, > > + sizeof(*params), 1, params, ¶ms_id); > > + if (vas != VA_STATUS_SUCCESS) { > > + av_log(avctx, AV_LOG_ERROR, "Failed to create parameter buffer: " > > + "%d (%s).\n", vas, vaErrorStr(vas)); > > + err = AVERROR(EIO); > > + goto fail_after_begin; > > + } > > + av_log(avctx, AV_LOG_DEBUG, "Pipeline parameter buffer is %#x.\n", > > + params_id); > > + > > + > > + vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context, > > + VAProcPipelineParameterBufferType, > > + sizeof(*subpic_params), 1, subpic_params, > &subpic_params_id); > > + if (vas != VA_STATUS_SUCCESS) { > > + av_log(avctx, AV_LOG_ERROR, "Failed to create parameter buffer: " > > + "%d (%s).\n", vas, vaErrorStr(vas)); > > + err = AVERROR(EIO); > > + goto fail_after_begin; > > + } > > + av_log(avctx, AV_LOG_DEBUG, "Pipeline subpic parameter buffer > is %#x.\n", > > + subpic_params_id); > > + > > + vas = vaRenderPicture(ctx->hwctx->display, ctx->va_context, > > + ¶ms_id, 1); > > + if (vas != VA_STATUS_SUCCESS) { > > + av_log(avctx, AV_LOG_ERROR, "Failed to render parameter buffer: " > > + "%d (%s).\n", vas, vaErrorStr(vas)); > > + err = AVERROR(EIO); > > + goto fail_after_begin; > > + } > > + > > + vas = vaRenderPicture(ctx->hwctx->display, ctx->va_context, > > + &subpic_params_id, 1); > > + if (vas != VA_STATUS_SUCCESS) { > > + av_log(avctx, AV_LOG_ERROR, "Failed to render subpic parameter > buffer: " > > + "%d (%s).\n", vas, vaErrorStr(vas)); > > + err = AVERROR(EIO); > > + goto fail_after_begin; > > + } > > + > > + vas = vaEndPicture(ctx->hwctx->display, ctx->va_context); > > + if (vas != VA_STATUS_SUCCESS) { > > + av_log(avctx, AV_LOG_ERROR, "Failed to start picture processing: " > > + "%d (%s).\n", vas, vaErrorStr(vas)); > > + err = AVERROR(EIO); > > + goto fail_after_render; > > + } > > + > > + if (CONFIG_VAAPI_1 || ctx->hwctx->driver_quirks & > > + AV_VAAPI_DRIVER_QUIRK_RENDER_PARAM_BUFFERS) { > > + vas = vaDestroyBuffer(ctx->hwctx->display, params_id); > > + if (vas != VA_STATUS_SUCCESS) { > > + av_log(avctx, AV_LOG_ERROR, "Failed to free parameter buffer: " > > + "%d (%s).\n", vas, vaErrorStr(vas)); > > + // And ignore. > > + } > > + } > > + > > + return 0; > > + > > + // We want to make sure that if vaBeginPicture has been called, we also > > + // call vaRenderPicture and vaEndPicture. These calls may well fail or > > + // do something else nasty, but once we're in this failure case there > > + // isn't much else we can do. > > +fail_after_begin: > > + vaRenderPicture(ctx->hwctx->display, ctx->va_context, ¶ms_id, > > +1); > > +fail_after_render: > > + vaEndPicture(ctx->hwctx->display, ctx->va_context); > > +fail: > > + return err; > > +} > > + > > void ff_vaapi_vpp_ctx_init(AVFilterContext *avctx) { > > int i; > > diff --git a/libavfilter/vaapi_vpp.h b/libavfilter/vaapi_vpp.h index > > c3da91717c..8d53acdec9 100644 > > --- a/libavfilter/vaapi_vpp.h > > +++ b/libavfilter/vaapi_vpp.h > > @@ -83,4 +83,9 @@ int ff_vaapi_vpp_render_picture(AVFilterContext > *avctx, > > VAProcPipelineParameterBuffer *params, > > AVFrame *output_frame); > > > > +int ff_vaapi_vpp_render_overlay(AVFilterContext *avctx, > > + VAProcPipelineParameterBuffer *primary_params, > > + VAProcPipelineParameterBuffer *subpic_params, > > + VASurfaceID output_surface); > > In the absence of a second user I don't think it's worth putting this in the > generic file. I am considering the do the change, planning to add it to vf_overlay_vaapi.c > > > + > > #endif /* AVFILTER_VAAPI_VPP_H */ > > diff --git a/libavfilter/vf_overlay_vaapi.c > > b/libavfilter/vf_overlay_vaapi.c new file mode 100644 index > > 0000000000..bcea76a806 > > --- /dev/null > > +++ b/libavfilter/vf_overlay_vaapi.c > > @@ -0,0 +1,365 @@ > > +/* > > + * This file is part of FFmpeg. > > + * > > + * FFmpeg is free software; you can redistribute it and/or > > + * modify it under the terms of the GNU Lesser General Public > > + * License as published by the Free Software Foundation; either > > + * version 2.1 of the License, or (at your option) any later version. > > + * > > + * FFmpeg is distributed in the hope that it will be useful, > > + * but WITHOUT ANY WARRANTY; without even the implied warranty of > > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the > GNU > > + * Lesser General Public License for more details. > > + * > > + * You should have received a copy of the GNU Lesser General Public > > + * License along with FFmpeg; if not, write to the Free Software > > + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA > > +02110-1301 USA */ #include <string.h> > > + > > +#include "libavutil/avassert.h" > > +#include "libavutil/mem.h" > > +#include "libavutil/opt.h" > > +#include "libavutil/pixdesc.h" > > + > > +#include "avfilter.h" > > +#include "framesync.h" > > +#include "formats.h" > > +#include "internal.h" > > +#include "vaapi_vpp.h" > > + > > +#if VA_CHECK_VERSION(2, 1, 0) > > +// Blend State > > +static VABlendState blend_state = { /** \brief Video blending flags. > > +*/ > > Static data in libraries must be const. will change it by add const. > > > + .flags = VA_BLEND_PREMULTIPLIED_ALPHA, > //VA_BLEND_GLOBAL_ALPHA, > > + .global_alpha = 1, > > + .min_luma = 0, > > + .max_luma = 1 > > These are all floats, use 0.0f and 1.0f to avoid confusion. will do the change. > > > +}; > > +#endif > > + > > +typedef struct OverlayVAAPIContext { > > + VAAPIVPPContext vpp_ctx; // must be the first field > > + > > + FFFrameSync fs; > > + > > + int overlay_x; > > + int overlay_y; > > + int overlay_w; > > + int overlay_h; > > + float overlay_alpha; > > +} OverlayVAAPIContext; > > + > > +static int overlay_vaapi_query_formats(AVFilterContext *ctx) { > > + int i; > > + int ret; > > + > > + static const enum AVPixelFormat main_in_fmts[] = { > > + AV_PIX_FMT_NV12, > > + AV_PIX_FMT_RGB32, > > + AV_PIX_FMT_VAAPI, > > + AV_PIX_FMT_NONE > > + }; > > + static const enum AVPixelFormat out_pix_fmts[] = { > > + AV_PIX_FMT_NV12, > > + AV_PIX_FMT_RGB32, > > + AV_PIX_FMT_VAAPI, > > + AV_PIX_FMT_NONE > > + }; > > I don't see how this can handle the software pixfmts? will remove software pixfmts. > > > + > > + for (i = 0; i < ctx->nb_inputs; i++) { > > + ret = ff_formats_ref(ff_make_format_list(main_in_fmts), &ctx- > >inputs[i]->out_formats); > > + if (ret < 0) > > + return ret; > > + } > > + > > + ret = ff_formats_ref(ff_make_format_list(out_pix_fmts), &ctx- > >outputs[0]->in_formats); > > + if (ret < 0) > > + return ret; > > + > > + return 0; > > +} > > + > > +static int overlay_vaapi_blend(FFFrameSync *fs) { > > + AVFilterContext *avctx = fs->parent; > > + AVFilterLink *outlink = avctx->outputs[0]; > > + OverlayVAAPIContext *ctx = avctx->priv; > > + VAAPIVPPContext *vpp_ctx = avctx->priv; > > + AVFrame *input_main, *input_overlay; > > + AVFrame *output; > > + int err; > > + VASurfaceID main_surface, overlay_surface, output_surface; > > + VARectangle main_region, overlay_region, output_region; > > + VAProcPipelineParameterBuffer params, subpic_params; > > + > > + err = ff_framesync_get_frame(fs, 0, &input_main, 0); > > + if (err < 0) > > + return err; > > + err = ff_framesync_get_frame(fs, 1, &input_overlay, 0); > > + if (err < 0) > > + return err; > > + > > + av_log(avctx, AV_LOG_DEBUG, "Filter main: %s, %ux%u (%"PRId64").\n", > > + av_get_pix_fmt_name(input_main->format), > > + input_main->width, input_main->height, input_main->pts); > > + > > + av_log(avctx, AV_LOG_DEBUG, "Filter overlay: %s, %ux%u > (%"PRId64").\n", > > + av_get_pix_fmt_name(input_overlay->format), > > + input_overlay->width, input_overlay->height, > > + input_overlay->pts); > > + > > + if (vpp_ctx->va_context == VA_INVALID_ID) > > + return AVERROR(EINVAL); > > + > > + main_surface = (VASurfaceID)(uintptr_t)input_main->data[3]; > > + av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp > main.\n", > > + main_surface); > > + > > + overlay_surface = (VASurfaceID)(uintptr_t)input_overlay->data[3]; > > + av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp > overlay.\n", > > + overlay_surface); > > + > > + output = ff_get_video_buffer(outlink, outlink->w, outlink->h); > > + if (!output) { > > + err = AVERROR(ENOMEM); > > + goto fail; > > + } > > + > > + output_surface = (VASurfaceID)(uintptr_t)output->data[3]; > > + av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp > output.\n", > > + output_surface); > > + > > + memset(¶ms, 0, sizeof(params)); > > + memset(&subpic_params, 0, sizeof(subpic_params)); > > + > > + main_region = (VARectangle) { > > + .x = 0, > > + .y = 0, > > + .width = input_main->width, > > + .height = input_main->height, > > + }; > > + > > + overlay_region = (VARectangle) { > > + .x = ctx->overlay_x, > > + .y = ctx->overlay_y, > > + .width = input_overlay->width, > > + .height = input_overlay->height, > > + }; > > + > > + output_region = (VARectangle) { > > + .x = 0, > > + .y = 0, > > + .width = output->width, > > + .height = output->height, > > + }; > > + > > + if (!ctx->overlay_w && !ctx->overlay_h) { > > + overlay_region.width = ctx->overlay_w; > > + overlay_region.height = ctx->overlay_h; > > + } > > + > > + if (overlay_region.width > main_region.width || > > + overlay_region.height > main_region.height) { > > + av_log(ctx, AV_LOG_ERROR, > > + "Overlay width > main width or overlay height > main height is not > supported.\n"); > > + err = AVERROR(EINVAL); > > + goto fail; > > + } > > What about overlay_region.x + overlay_region.with > main_region.width - > does that work? it should do the crop and output the main_region.width, will test it. > > > + > > + params.filters = &vpp_ctx->filter_buffers[0]; > > + params.num_filters = vpp_ctx->nb_filter_buffers; > > + > > + params.surface = main_surface; > > + params.surface_region = &main_region; > > + > > + switch (input_main->colorspace) { > > + case AVCOL_SPC_BT470BG: > > + params.surface_color_standard = VAProcColorStandardBT601; > > + break; > > + case AVCOL_SPC_BT709: > > + params.surface_color_standard = VAProcColorStandardBT709; > > + break; > > + default: > > + params.surface_color_standard = VAProcColorStandardNone; > > + break; > > + } > > We have better support for colours - see ff_vaapi_vpp_init_params(). will do the change > > > + > > + params.output_region = &output_region; > > + params.output_background_color = VAAPI_VPP_BACKGROUND_BLACK; > > + params.output_color_standard = params.surface_color_standard; > > + > > + memcpy(&subpic_params, ¶ms, sizeof(subpic_params)); > > + subpic_params.pipeline_flags |= VA_PROC_PIPELINE_FAST; > > + subpic_params.filter_flags |= VA_FILTER_SCALING_FAST; > > Why are these flags set? As I tested, this flags can work before, will try to test if it okay to remove them. > > > + > > +#if VA_CHECK_VERSION(2, 1, 0) > > + blend_state.global_alpha = ctx->overlay_alpha; > > + subpic_params.blend_state = &blend_state; #endif > > + > > + subpic_params.surface = overlay_surface; > > + subpic_params.output_region = &overlay_region; > > + > > + err = ff_vaapi_vpp_render_overlay(avctx, ¶ms, &subpic_params, > output_surface); > > + if (err < 0) > > + goto fail; > > + > > + err = av_frame_copy_props(output, input_main); > > + if (err < 0) > > + goto fail; > > + > > + av_log(avctx, AV_LOG_DEBUG, "Filter output: %s, %ux%u > (%"PRId64").\n", > > + av_get_pix_fmt_name(output->format), > > + output->width, output->height, output->pts); > > + > > + return ff_filter_frame(outlink, output); > > + > > +fail: > > + av_frame_free(&output); > > + return err; > > +} > > + > > +static int overlay_vaapi_init_framesync(AVFilterContext *avctx) { > > + OverlayVAAPIContext *ctx = avctx->priv; > > + int ret, i; > > + > > + ctx->fs.on_event = overlay_vaapi_blend; > > + ctx->fs.opaque = ctx; > > + ret = ff_framesync_init(&ctx->fs, avctx, avctx->nb_inputs); > > + if (ret < 0) > > + return ret; > > + > > + for (i = 0; i < avctx->nb_inputs; i++) { > > + FFFrameSyncIn *in = &ctx->fs.in[i]; > > + in->before = EXT_STOP; > > + in->after = EXT_INFINITY; > > + in->sync = i ? 1 : 2; > > + in->time_base = avctx->inputs[i]->time_base; > > + } > > + > > + return ff_framesync_configure(&ctx->fs); } > > + > > +static int overlay_vaapi_config_output(AVFilterLink *outlink) { > > + AVFilterContext *avctx = outlink->src; > > + OverlayVAAPIContext *ctx = avctx->priv; > > + VAAPIVPPContext *vpp_ctx = avctx->priv; > > + AVFilterLink *in0 = avctx->inputs[0]; > > + AVFilterLink *in1 = avctx->inputs[1]; > > + int err; > > + > > + av_log(ctx, AV_LOG_DEBUG, "Output is of %s.\n", > av_get_pix_fmt_name(outlink->format)); > > + if ((in0->format == AV_PIX_FMT_VAAPI && in1->format != > AV_PIX_FMT_VAAPI) || > > + (in0->format != AV_PIX_FMT_VAAPI && in1->format == > AV_PIX_FMT_VAAPI)) { > > + av_log(ctx, AV_LOG_ERROR, "Mixing hardware and software pixel > > + formats is not supported.\n"); > > Software pixel formats don't appear to be supported at all, though? Got it. > > > + return AVERROR(EINVAL); > > + } > > + > > + err = overlay_vaapi_init_framesync(avctx); > > + if (err < 0) > > + return err; > > + > > + vpp_ctx->output_width = avctx->inputs[0]->w; > > + vpp_ctx->output_height = avctx->inputs[0]->h; > > + > > + err = ff_vaapi_vpp_config_output(outlink); > > + if (err < 0) > > + return err; > > Is it ok to copy the format of the output frames from the last-initialised input, > as this does? What happens if the inputs aren't in the same format? As my test, the filter doesn't support the different format input. > > > + > > + err = ff_framesync_init_dualinput(&ctx->fs, avctx); > > + if (err < 0) > > + return err; > > + > > + return ff_framesync_configure(&ctx->fs); } > > + > > +static av_cold int overlay_vaapi_init(AVFilterContext *avctx) { > > + VAAPIVPPContext *vpp_ctx = avctx->priv; > > + > > + ff_vaapi_vpp_ctx_init(avctx); > > + vpp_ctx->output_format = AV_PIX_FMT_NONE; > > + > > + return 0; > > +} > > + > > +static int overlay_vaapi_activate(AVFilterContext *avctx) { > > + OverlayVAAPIContext *ctx = avctx->priv; > > + > > + return ff_framesync_activate(&ctx->fs); } > > + > > +static av_cold void overlay_vaapi_uninit(AVFilterContext *avctx) { > > + OverlayVAAPIContext *ctx = avctx->priv; > > + > > + ff_framesync_uninit(&ctx->fs); > > +} > > + > > +static AVFrame *get_video_buffer(AVFilterLink *inlink, int w, int h) > > +{ > > + return ff_default_get_video_buffer(inlink, w, h); } > > I don't think this function is needed - the ff_default_get_video_buffer() is > already the default if it isn't set. will do the change. > > > + > > +#define OFFSET(x) offsetof(OverlayVAAPIContext, x) #define FLAGS > > +(AV_OPT_FLAG_FILTERING_PARAM | AV_OPT_FLAG_VIDEO_PARAM) static > const > > +AVOption overlay_vaapi_options[] = { > > + { "x", "Overlay x position", > > + OFFSET(overlay_x), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = > FLAGS }, > > + { "y", "Overlay y position", > > + OFFSET(overlay_y), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = > FLAGS }, > > + { "w", "Overlay width", > > + OFFSET(overlay_w), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = > FLAGS }, > > + { "h", "Overlay hight", > > + OFFSET(overlay_h), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = > FLAGS }, > > + { "alpha", "Overlay global alpha", > > + OFFSET(overlay_alpha), AV_OPT_TYPE_FLOAT, { .dbl = 0.0}, 0.0, > 1.0, .flags = FLAGS}, > > + { NULL }, > > +}; > > + > > +AVFILTER_DEFINE_CLASS(overlay_vaapi); > > + > > +static const AVFilterPad overlay_vaapi_inputs[] = { > > + { > > + .name = "main", > > + .type = AVMEDIA_TYPE_VIDEO, > > + .get_video_buffer = get_video_buffer, > > + .config_props = &ff_vaapi_vpp_config_input, > > + .needs_fifo = 1, > > + }, > > + { > > + .name = "overlay", > > + .type = AVMEDIA_TYPE_VIDEO, > > + .get_video_buffer = get_video_buffer, > > + .config_props = &ff_vaapi_vpp_config_input, > > + .needs_fifo = 1, > > + }, > > Calling ff_vaapi_vpp_config_input() twice doesn't look like something you > wanted to do. In particular, it leaks a reference to an inlink hw_frames_ctx. I will try to remove one of ff_vaapi_vpp_config_input(). > > > + { NULL } > > +}; > > + > > +static const AVFilterPad overlay_vaapi_outputs[] = { > > + { > > + .name = "default", > > + .type = AVMEDIA_TYPE_VIDEO, > > + .config_props = &overlay_vaapi_config_output, > > + }, > > + { NULL } > > +}; > > + > > +AVFilter ff_vf_overlay_vaapi = { > > + .name = "overlay_vaapi", > > + .description = NULL_IF_CONFIG_SMALL("Overlay one video on top of > another"), > > + .priv_size = sizeof(OverlayVAAPIContext), > > + .priv_class = &overlay_vaapi_class, > > + .init = &overlay_vaapi_init, > > + .uninit = &overlay_vaapi_uninit, > > + .query_formats = &overlay_vaapi_query_formats, > > + .activate = &overlay_vaapi_activate, > > + .inputs = overlay_vaapi_inputs, > > + .outputs = overlay_vaapi_outputs, > > + .flags_internal = FF_FILTER_FLAG_HWFRAME_AWARE, }; > > > > - Mark Thank you Mark for the review, will send the next version soon. -Zach > _______________________________________________ > ffmpeg-devel mailing list > ffmpeg-devel@ffmpeg.org > https://ffmpeg.org/mailman/listinfo/ffmpeg-devel > > To unsubscribe, visit link above, or email ffmpeg-devel-request@ffmpeg.org > with subject "unsubscribe".
diff --git a/configure b/configure index 32fc26356c..f469e6a3b1 100755 --- a/configure +++ b/configure @@ -3478,6 +3478,7 @@ openclsrc_filter_deps="opencl" overlay_opencl_filter_deps="opencl" overlay_qsv_filter_deps="libmfx" overlay_qsv_filter_select="qsvvpp" +overlay_vaapi_filter_deps="vaapi" owdenoise_filter_deps="gpl" pan_filter_deps="swresample" perspective_filter_deps="gpl" diff --git a/libavfilter/Makefile b/libavfilter/Makefile index 07ea8d7edc..5cbf1a7e41 100644 --- a/libavfilter/Makefile +++ b/libavfilter/Makefile @@ -311,6 +311,7 @@ OBJS-$(CONFIG_OVERLAY_FILTER) += vf_overlay.o framesync.o OBJS-$(CONFIG_OVERLAY_OPENCL_FILTER) += vf_overlay_opencl.o opencl.o \ opencl/overlay.o framesync.o OBJS-$(CONFIG_OVERLAY_QSV_FILTER) += vf_overlay_qsv.o framesync.o +OBJS-$(CONFIG_OVERLAY_VAAPI_FILTER) += vf_overlay_vaapi.o framesync.o OBJS-$(CONFIG_OWDENOISE_FILTER) += vf_owdenoise.o OBJS-$(CONFIG_PAD_FILTER) += vf_pad.o OBJS-$(CONFIG_PALETTEGEN_FILTER) += vf_palettegen.o diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c index 9c846b1ddd..27ee1df78b 100644 --- a/libavfilter/allfilters.c +++ b/libavfilter/allfilters.c @@ -295,6 +295,7 @@ extern AVFilter ff_vf_oscilloscope; extern AVFilter ff_vf_overlay; extern AVFilter ff_vf_overlay_opencl; extern AVFilter ff_vf_overlay_qsv; +extern AVFilter ff_vf_overlay_vaapi; extern AVFilter ff_vf_owdenoise; extern AVFilter ff_vf_pad; extern AVFilter ff_vf_palettegen; diff --git a/libavfilter/vaapi_vpp.c b/libavfilter/vaapi_vpp.c index b5b245c8af..a8caa5b532 100644 --- a/libavfilter/vaapi_vpp.c +++ b/libavfilter/vaapi_vpp.c @@ -663,6 +663,101 @@ fail: return err; } +int ff_vaapi_vpp_render_overlay(AVFilterContext *avctx, + VAProcPipelineParameterBuffer *params, + VAProcPipelineParameterBuffer *subpic_params, + VASurfaceID output_surface) +{ + VABufferID params_id; + VABufferID subpic_params_id; + VAStatus vas; + int err = 0; + VAAPIVPPContext *ctx = avctx->priv; + + vas = vaBeginPicture(ctx->hwctx->display, + ctx->va_context, output_surface); + if (vas != VA_STATUS_SUCCESS) { + av_log(avctx, AV_LOG_ERROR, "Failed to attach new picture: " + "%d (%s).\n", vas, vaErrorStr(vas)); + err = AVERROR(EIO); + goto fail; + } + + vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context, + VAProcPipelineParameterBufferType, + sizeof(*params), 1, params, ¶ms_id); + if (vas != VA_STATUS_SUCCESS) { + av_log(avctx, AV_LOG_ERROR, "Failed to create parameter buffer: " + "%d (%s).\n", vas, vaErrorStr(vas)); + err = AVERROR(EIO); + goto fail_after_begin; + } + av_log(avctx, AV_LOG_DEBUG, "Pipeline parameter buffer is %#x.\n", + params_id); + + + vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context, + VAProcPipelineParameterBufferType, + sizeof(*subpic_params), 1, subpic_params, &subpic_params_id); + if (vas != VA_STATUS_SUCCESS) { + av_log(avctx, AV_LOG_ERROR, "Failed to create parameter buffer: " + "%d (%s).\n", vas, vaErrorStr(vas)); + err = AVERROR(EIO); + goto fail_after_begin; + } + av_log(avctx, AV_LOG_DEBUG, "Pipeline subpic parameter buffer is %#x.\n", + subpic_params_id); + + vas = vaRenderPicture(ctx->hwctx->display, ctx->va_context, + ¶ms_id, 1); + if (vas != VA_STATUS_SUCCESS) { + av_log(avctx, AV_LOG_ERROR, "Failed to render parameter buffer: " + "%d (%s).\n", vas, vaErrorStr(vas)); + err = AVERROR(EIO); + goto fail_after_begin; + } + + vas = vaRenderPicture(ctx->hwctx->display, ctx->va_context, + &subpic_params_id, 1); + if (vas != VA_STATUS_SUCCESS) { + av_log(avctx, AV_LOG_ERROR, "Failed to render subpic parameter buffer: " + "%d (%s).\n", vas, vaErrorStr(vas)); + err = AVERROR(EIO); + goto fail_after_begin; + } + + vas = vaEndPicture(ctx->hwctx->display, ctx->va_context); + if (vas != VA_STATUS_SUCCESS) { + av_log(avctx, AV_LOG_ERROR, "Failed to start picture processing: " + "%d (%s).\n", vas, vaErrorStr(vas)); + err = AVERROR(EIO); + goto fail_after_render; + } + + if (CONFIG_VAAPI_1 || ctx->hwctx->driver_quirks & + AV_VAAPI_DRIVER_QUIRK_RENDER_PARAM_BUFFERS) { + vas = vaDestroyBuffer(ctx->hwctx->display, params_id); + if (vas != VA_STATUS_SUCCESS) { + av_log(avctx, AV_LOG_ERROR, "Failed to free parameter buffer: " + "%d (%s).\n", vas, vaErrorStr(vas)); + // And ignore. + } + } + + return 0; + + // We want to make sure that if vaBeginPicture has been called, we also + // call vaRenderPicture and vaEndPicture. These calls may well fail or + // do something else nasty, but once we're in this failure case there + // isn't much else we can do. +fail_after_begin: + vaRenderPicture(ctx->hwctx->display, ctx->va_context, ¶ms_id, 1); +fail_after_render: + vaEndPicture(ctx->hwctx->display, ctx->va_context); +fail: + return err; +} + void ff_vaapi_vpp_ctx_init(AVFilterContext *avctx) { int i; diff --git a/libavfilter/vaapi_vpp.h b/libavfilter/vaapi_vpp.h index c3da91717c..8d53acdec9 100644 --- a/libavfilter/vaapi_vpp.h +++ b/libavfilter/vaapi_vpp.h @@ -83,4 +83,9 @@ int ff_vaapi_vpp_render_picture(AVFilterContext *avctx, VAProcPipelineParameterBuffer *params, AVFrame *output_frame); +int ff_vaapi_vpp_render_overlay(AVFilterContext *avctx, + VAProcPipelineParameterBuffer *primary_params, + VAProcPipelineParameterBuffer *subpic_params, + VASurfaceID output_surface); + #endif /* AVFILTER_VAAPI_VPP_H */ diff --git a/libavfilter/vf_overlay_vaapi.c b/libavfilter/vf_overlay_vaapi.c new file mode 100644 index 0000000000..bcea76a806 --- /dev/null +++ b/libavfilter/vf_overlay_vaapi.c @@ -0,0 +1,365 @@ +/* + * This file is part of FFmpeg. + * + * FFmpeg is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * FFmpeg is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with FFmpeg; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + */ +#include <string.h> + +#include "libavutil/avassert.h" +#include "libavutil/mem.h" +#include "libavutil/opt.h" +#include "libavutil/pixdesc.h" + +#include "avfilter.h" +#include "framesync.h" +#include "formats.h" +#include "internal.h" +#include "vaapi_vpp.h" + +#if VA_CHECK_VERSION(2, 1, 0) +// Blend State +static VABlendState blend_state = { /** \brief Video blending flags. */ + .flags = VA_BLEND_PREMULTIPLIED_ALPHA, //VA_BLEND_GLOBAL_ALPHA, + .global_alpha = 1, + .min_luma = 0, + .max_luma = 1 +}; +#endif + +typedef struct OverlayVAAPIContext { + VAAPIVPPContext vpp_ctx; // must be the first field + + FFFrameSync fs; + + int overlay_x; + int overlay_y; + int overlay_w; + int overlay_h; + float overlay_alpha; +} OverlayVAAPIContext; + +static int overlay_vaapi_query_formats(AVFilterContext *ctx) +{ + int i; + int ret; + + static const enum AVPixelFormat main_in_fmts[] = { + AV_PIX_FMT_NV12, + AV_PIX_FMT_RGB32, + AV_PIX_FMT_VAAPI, + AV_PIX_FMT_NONE + }; + static const enum AVPixelFormat out_pix_fmts[] = { + AV_PIX_FMT_NV12, + AV_PIX_FMT_RGB32, + AV_PIX_FMT_VAAPI, + AV_PIX_FMT_NONE + }; + + for (i = 0; i < ctx->nb_inputs; i++) { + ret = ff_formats_ref(ff_make_format_list(main_in_fmts), &ctx->inputs[i]->out_formats); + if (ret < 0) + return ret; + } + + ret = ff_formats_ref(ff_make_format_list(out_pix_fmts), &ctx->outputs[0]->in_formats); + if (ret < 0) + return ret; + + return 0; +} + +static int overlay_vaapi_blend(FFFrameSync *fs) +{ + AVFilterContext *avctx = fs->parent; + AVFilterLink *outlink = avctx->outputs[0]; + OverlayVAAPIContext *ctx = avctx->priv; + VAAPIVPPContext *vpp_ctx = avctx->priv; + AVFrame *input_main, *input_overlay; + AVFrame *output; + int err; + VASurfaceID main_surface, overlay_surface, output_surface; + VARectangle main_region, overlay_region, output_region; + VAProcPipelineParameterBuffer params, subpic_params; + + err = ff_framesync_get_frame(fs, 0, &input_main, 0); + if (err < 0) + return err; + err = ff_framesync_get_frame(fs, 1, &input_overlay, 0); + if (err < 0) + return err; + + av_log(avctx, AV_LOG_DEBUG, "Filter main: %s, %ux%u (%"PRId64").\n", + av_get_pix_fmt_name(input_main->format), + input_main->width, input_main->height, input_main->pts); + + av_log(avctx, AV_LOG_DEBUG, "Filter overlay: %s, %ux%u (%"PRId64").\n", + av_get_pix_fmt_name(input_overlay->format), + input_overlay->width, input_overlay->height, input_overlay->pts); + + if (vpp_ctx->va_context == VA_INVALID_ID) + return AVERROR(EINVAL); + + main_surface = (VASurfaceID)(uintptr_t)input_main->data[3]; + av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp main.\n", + main_surface); + + overlay_surface = (VASurfaceID)(uintptr_t)input_overlay->data[3]; + av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp overlay.\n", + overlay_surface); + + output = ff_get_video_buffer(outlink, outlink->w, outlink->h); + if (!output) { + err = AVERROR(ENOMEM); + goto fail; + } + + output_surface = (VASurfaceID)(uintptr_t)output->data[3]; + av_log(avctx, AV_LOG_DEBUG, "Using surface %#x for overlay vpp output.\n", + output_surface); + + memset(¶ms, 0, sizeof(params)); + memset(&subpic_params, 0, sizeof(subpic_params)); + + main_region = (VARectangle) { + .x = 0, + .y = 0, + .width = input_main->width, + .height = input_main->height, + }; + + overlay_region = (VARectangle) { + .x = ctx->overlay_x, + .y = ctx->overlay_y, + .width = input_overlay->width, + .height = input_overlay->height, + }; + + output_region = (VARectangle) { + .x = 0, + .y = 0, + .width = output->width, + .height = output->height, + }; + + if (!ctx->overlay_w && !ctx->overlay_h) { + overlay_region.width = ctx->overlay_w; + overlay_region.height = ctx->overlay_h; + } + + if (overlay_region.width > main_region.width || + overlay_region.height > main_region.height) { + av_log(ctx, AV_LOG_ERROR, + "Overlay width > main width or overlay height > main height is not supported.\n"); + err = AVERROR(EINVAL); + goto fail; + } + + params.filters = &vpp_ctx->filter_buffers[0]; + params.num_filters = vpp_ctx->nb_filter_buffers; + + params.surface = main_surface; + params.surface_region = &main_region; + + switch (input_main->colorspace) { + case AVCOL_SPC_BT470BG: + params.surface_color_standard = VAProcColorStandardBT601; + break; + case AVCOL_SPC_BT709: + params.surface_color_standard = VAProcColorStandardBT709; + break; + default: + params.surface_color_standard = VAProcColorStandardNone; + break; + } + + params.output_region = &output_region; + params.output_background_color = VAAPI_VPP_BACKGROUND_BLACK; + params.output_color_standard = params.surface_color_standard; + + memcpy(&subpic_params, ¶ms, sizeof(subpic_params)); + subpic_params.pipeline_flags |= VA_PROC_PIPELINE_FAST; + subpic_params.filter_flags |= VA_FILTER_SCALING_FAST; + +#if VA_CHECK_VERSION(2, 1, 0) + blend_state.global_alpha = ctx->overlay_alpha; + subpic_params.blend_state = &blend_state; +#endif + + subpic_params.surface = overlay_surface; + subpic_params.output_region = &overlay_region; + + err = ff_vaapi_vpp_render_overlay(avctx, ¶ms, &subpic_params, output_surface); + if (err < 0) + goto fail; + + err = av_frame_copy_props(output, input_main); + if (err < 0) + goto fail; + + av_log(avctx, AV_LOG_DEBUG, "Filter output: %s, %ux%u (%"PRId64").\n", + av_get_pix_fmt_name(output->format), + output->width, output->height, output->pts); + + return ff_filter_frame(outlink, output); + +fail: + av_frame_free(&output); + return err; +} + +static int overlay_vaapi_init_framesync(AVFilterContext *avctx) +{ + OverlayVAAPIContext *ctx = avctx->priv; + int ret, i; + + ctx->fs.on_event = overlay_vaapi_blend; + ctx->fs.opaque = ctx; + ret = ff_framesync_init(&ctx->fs, avctx, avctx->nb_inputs); + if (ret < 0) + return ret; + + for (i = 0; i < avctx->nb_inputs; i++) { + FFFrameSyncIn *in = &ctx->fs.in[i]; + in->before = EXT_STOP; + in->after = EXT_INFINITY; + in->sync = i ? 1 : 2; + in->time_base = avctx->inputs[i]->time_base; + } + + return ff_framesync_configure(&ctx->fs); +} + +static int overlay_vaapi_config_output(AVFilterLink *outlink) +{ + AVFilterContext *avctx = outlink->src; + OverlayVAAPIContext *ctx = avctx->priv; + VAAPIVPPContext *vpp_ctx = avctx->priv; + AVFilterLink *in0 = avctx->inputs[0]; + AVFilterLink *in1 = avctx->inputs[1]; + int err; + + av_log(ctx, AV_LOG_DEBUG, "Output is of %s.\n", av_get_pix_fmt_name(outlink->format)); + if ((in0->format == AV_PIX_FMT_VAAPI && in1->format != AV_PIX_FMT_VAAPI) || + (in0->format != AV_PIX_FMT_VAAPI && in1->format == AV_PIX_FMT_VAAPI)) { + av_log(ctx, AV_LOG_ERROR, "Mixing hardware and software pixel formats is not supported.\n"); + return AVERROR(EINVAL); + } + + err = overlay_vaapi_init_framesync(avctx); + if (err < 0) + return err; + + vpp_ctx->output_width = avctx->inputs[0]->w; + vpp_ctx->output_height = avctx->inputs[0]->h; + + err = ff_vaapi_vpp_config_output(outlink); + if (err < 0) + return err; + + err = ff_framesync_init_dualinput(&ctx->fs, avctx); + if (err < 0) + return err; + + return ff_framesync_configure(&ctx->fs); +} + +static av_cold int overlay_vaapi_init(AVFilterContext *avctx) +{ + VAAPIVPPContext *vpp_ctx = avctx->priv; + + ff_vaapi_vpp_ctx_init(avctx); + vpp_ctx->output_format = AV_PIX_FMT_NONE; + + return 0; +} + +static int overlay_vaapi_activate(AVFilterContext *avctx) +{ + OverlayVAAPIContext *ctx = avctx->priv; + + return ff_framesync_activate(&ctx->fs); +} + +static av_cold void overlay_vaapi_uninit(AVFilterContext *avctx) +{ + OverlayVAAPIContext *ctx = avctx->priv; + + ff_framesync_uninit(&ctx->fs); +} + +static AVFrame *get_video_buffer(AVFilterLink *inlink, int w, int h) +{ + return ff_default_get_video_buffer(inlink, w, h); +} + +#define OFFSET(x) offsetof(OverlayVAAPIContext, x) +#define FLAGS (AV_OPT_FLAG_FILTERING_PARAM | AV_OPT_FLAG_VIDEO_PARAM) +static const AVOption overlay_vaapi_options[] = { + { "x", "Overlay x position", + OFFSET(overlay_x), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = FLAGS }, + { "y", "Overlay y position", + OFFSET(overlay_y), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = FLAGS }, + { "w", "Overlay width", + OFFSET(overlay_w), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = FLAGS }, + { "h", "Overlay hight", + OFFSET(overlay_h), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = FLAGS }, + { "alpha", "Overlay global alpha", + OFFSET(overlay_alpha), AV_OPT_TYPE_FLOAT, { .dbl = 0.0}, 0.0, 1.0, .flags = FLAGS}, + { NULL }, +}; + +AVFILTER_DEFINE_CLASS(overlay_vaapi); + +static const AVFilterPad overlay_vaapi_inputs[] = { + { + .name = "main", + .type = AVMEDIA_TYPE_VIDEO, + .get_video_buffer = get_video_buffer, + .config_props = &ff_vaapi_vpp_config_input, + .needs_fifo = 1, + }, + { + .name = "overlay", + .type = AVMEDIA_TYPE_VIDEO, + .get_video_buffer = get_video_buffer, + .config_props = &ff_vaapi_vpp_config_input, + .needs_fifo = 1, + }, + { NULL } +}; + +static const AVFilterPad overlay_vaapi_outputs[] = { + { + .name = "default", + .type = AVMEDIA_TYPE_VIDEO, + .config_props = &overlay_vaapi_config_output, + }, + { NULL } +}; + +AVFilter ff_vf_overlay_vaapi = { + .name = "overlay_vaapi", + .description = NULL_IF_CONFIG_SMALL("Overlay one video on top of another"), + .priv_size = sizeof(OverlayVAAPIContext), + .priv_class = &overlay_vaapi_class, + .init = &overlay_vaapi_init, + .uninit = &overlay_vaapi_uninit, + .query_formats = &overlay_vaapi_query_formats, + .activate = &overlay_vaapi_activate, + .inputs = overlay_vaapi_inputs, + .outputs = overlay_vaapi_outputs, + .flags_internal = FF_FILTER_FLAG_HWFRAME_AWARE, +};