From patchwork Tue Feb 27 16:35:03 2024 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: gnattu X-Patchwork-Id: 46577 Delivered-To: ffmpegpatchwork2@gmail.com Received: by 2002:a05:6a20:c51b:b0:19e:cdac:8cce with SMTP id gm27csp2110468pzb; Tue, 27 Feb 2024 08:35:57 -0800 (PST) X-Forwarded-Encrypted: i=2; AJvYcCXg63L0qHKfv4tjzc72AFnzRH0okGp1kxb2sL9jTs1UbPjOZ7yQdQFtKYptSscnODon7GPKMI/Iit4vXMJoolkiWkx9WI+FJi/kwA== X-Google-Smtp-Source: AGHT+IFh2bNdIShG4y5hTJFrvs882crhkARscP2eBal37RTaabvdpSTNAbpCRUAJINoMfaC/B4RM X-Received: by 2002:a05:6402:b14:b0:565:2b6:6a7 with SMTP id bm20-20020a0564020b1400b0056502b606a7mr7643504edb.34.1709051756923; Tue, 27 Feb 2024 08:35:56 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1709051756; cv=none; d=google.com; s=arc-20160816; b=MIkEfDJqUhae7tok/IB3hDteM89pOaq/FeDzV/vRfRUcqL9Meql5C4r4kvb0nLmXEk SxaLCIzrT6WmvOLY5vxAALguR+1qWSPfW4R7OZ9ZEdmlUtEQkUDv0dbHxEvNSu58tAI+ FgzmlucXEXfAliipN//Y5jb3QSquhXAiMM3PURYojS8aU0a87zSWSof/qEutZ3OLvD+d kXHJSRJqItUDvJKDPeCtbeENGscGwI3rk6PKEuivmwX7KGl8qL0Uli75kOs03vug3Mmk H0HKdu7C8bZ0+yfCkDcSjCqUGX6H1XCwmYky5o+rezN0cDu/3QBhg8v1TjOPHYcapN5e cB5w== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=sender:errors-to:content-transfer-encoding:cc:reply-to:from :list-subscribe:list-help:list-post:list-archive:list-unsubscribe :list-id:precedence:subject:mime-version:message-id:date:to :delivered-to; bh=Iw1WaGxL15fYjgyXJqzstaFe4Hc5onDT+NopC/8DB0U=; fh=xM9Q7RyOgaJGxYG6eQTK8seY7Q6k7Q0+biksTdW2Uc8=; b=PGbLw7etKX9IW6M+mHjeOWxmDo83eK+EwGMjHZHDxzpXdcYWsNRdDCHVUZm+m6MyrB 4ixYydU0PMw1HKYUlqqqlKOfVN0Medb/lTCV2ElA37Dqu66Zx4SkUKho4WEm3tSS62a8 PsrsqtjZgtDBvIQs713pPp6f5m1BhW/+o/vJM9zK7OIuEQD8QJPvj6KWx0Qu3VbWpqKh B+ES88FrKKoctsETLmjeqPk7AyqXzD1iPpVNM4dmRPyKkaFE84k79zTEnb7XKO8FEzcL DJxaKeekSfbZlSYZKVbvWiJl+Fn/5enD4u1Fxn1mq6ZlYerpfHpjXihbBcMEek949w/g pt6g==; dara=google.com ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of ffmpeg-devel-bounces@ffmpeg.org designates 79.124.17.100 as permitted sender) smtp.mailfrom=ffmpeg-devel-bounces@ffmpeg.org Return-Path: Received: from ffbox0-bg.mplayerhq.hu (ffbox0-bg.ffmpeg.org. [79.124.17.100]) by mx.google.com with ESMTP id s19-20020a056402521300b005664e589501si259127edd.231.2024.02.27.08.35.56; Tue, 27 Feb 2024 08:35:56 -0800 (PST) Received-SPF: pass (google.com: domain of ffmpeg-devel-bounces@ffmpeg.org designates 79.124.17.100 as permitted sender) client-ip=79.124.17.100; Authentication-Results: mx.google.com; spf=pass (google.com: domain of ffmpeg-devel-bounces@ffmpeg.org designates 79.124.17.100 as permitted sender) smtp.mailfrom=ffmpeg-devel-bounces@ffmpeg.org Received: from [127.0.1.1] (localhost [127.0.0.1]) by ffbox0-bg.mplayerhq.hu (Postfix) with ESMTP id 3223668CC1D; Tue, 27 Feb 2024 18:35:52 +0200 (EET) X-Original-To: ffmpeg-devel@ffmpeg.org Delivered-To: ffmpeg-devel@ffmpeg.org Received: from mr85p00im-ztdg06021701.me.com (mr85p00im-ztdg06021701.me.com [17.58.23.196]) by ffbox0-bg.mplayerhq.hu (Postfix) with ESMTPS id 9C8E768C8FB for ; Tue, 27 Feb 2024 18:35:45 +0200 (EET) Received: from Yakumo-Yukari.lan (mr38p00im-dlb-asmtp-mailmevip.me.com [17.57.152.18]) by mr85p00im-ztdg06021701.me.com (Postfix) with ESMTPSA id 714D326334A0; Tue, 27 Feb 2024 16:35:40 +0000 (UTC) To: ffmpeg-devel@ffmpeg.org Date: Wed, 28 Feb 2024 00:35:03 +0800 Message-Id: <20240227163502.28268-1-gnattuoc@me.com> X-Mailer: git-send-email 2.39.3 (Apple Git-145) MIME-Version: 1.0 X-Proofpoint-GUID: N0aMSUa6CLk3k5Wa3WFtw58c0_TeeIpu X-Proofpoint-ORIG-GUID: N0aMSUa6CLk3k5Wa3WFtw58c0_TeeIpu X-Proofpoint-Virus-Version: vendor=baseguard engine=ICAP:2.0.272,Aquarius:18.0.1011,Hydra:6.0.619,FMLib:17.11.176.26 definitions=2024-02-27_03,2024-02-27_01,2023-05-22_02 X-Proofpoint-Spam-Details: rule=notspam policy=default score=0 mlxlogscore=999 phishscore=0 suspectscore=0 adultscore=0 clxscore=1015 malwarescore=0 spamscore=0 mlxscore=0 bulkscore=0 classifier=spam adjust=0 reason=mlx scancount=1 engine=8.19.0-2308100000 definitions=main-2402270128 Subject: [FFmpeg-devel] [PATCH v4] avfilter: add vf_overlay_videotoolbox X-BeenThere: ffmpeg-devel@ffmpeg.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: FFmpeg development discussions and patches List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-Patchwork-Original-From: gnattu via ffmpeg-devel From: gnattu Reply-To: FFmpeg development discussions and patches Cc: gnattu , zhilizhao@tencent.com Errors-To: ffmpeg-devel-bounces@ffmpeg.org Sender: "ffmpeg-devel" X-TUID: 6YHUhIPHpUMJ Overlay filter for VideoToolbox hwframes. Unlike most hardware overlay filters, this filter does not require the two inputs to have the same pixel format; instead, it will perform format conversion automatically with hardware accelerated methods. Signed-off-by: Gnattu OC --- Changes from v3: - Fixes an issue that 8bit depth BGRA overlay frames are not correctly converted to 16bit - Added a constraint to input pixel formats as VideoToolbox cannot convert all of its hardwareframes Changelog | 1 + configure | 1 + doc/filters.texi | 52 ++ libavfilter/Makefile | 3 + libavfilter/allfilters.c | 1 + libavfilter/metal/utils.m | 7 +- .../metal/vf_overlay_videotoolbox.metal | 58 ++ libavfilter/vf_overlay_videotoolbox.m | 609 ++++++++++++++++++ 8 files changed, 730 insertions(+), 2 deletions(-) create mode 100644 libavfilter/metal/vf_overlay_videotoolbox.metal create mode 100644 libavfilter/vf_overlay_videotoolbox.m diff --git a/Changelog b/Changelog index 610ee61dd6..3ecfdab81b 100644 --- a/Changelog +++ b/Changelog @@ -27,6 +27,7 @@ version : - a C11-compliant compiler is now required; note that this requirement will be bumped to C17 in the near future, so consider updating your build environment if it lacks C17 support +- VideoToolbox overlay filter version 6.1: - libaribcaption decoder diff --git a/configure b/configure index 23066efa32..a7c349d126 100755 --- a/configure +++ b/configure @@ -3807,6 +3807,7 @@ overlay_qsv_filter_deps="libmfx" overlay_qsv_filter_select="qsvvpp" overlay_vaapi_filter_deps="vaapi VAProcPipelineCaps_blend_flags" overlay_vulkan_filter_deps="vulkan spirv_compiler" +overlay_videotoolbox_filter_deps="metal corevideo coreimage videotoolbox" owdenoise_filter_deps="gpl" pad_opencl_filter_deps="opencl" pan_filter_deps="swresample" diff --git a/doc/filters.texi b/doc/filters.texi index e0436a5755..bfb77562cb 100644 --- a/doc/filters.texi +++ b/doc/filters.texi @@ -19033,6 +19033,58 @@ See @ref{framesync}. This filter also supports the @ref{framesync} options. +@section overlay_videotoolbox + +Overlay one video on top of another. + +This is the VideoToolbox variant of the @ref{overlay} filter. +It takes two inputs and has one output. The first input is the "main" video on which the second input is overlaid. +It only accepts VideoToolbox frames. The underlying input pixel formats do not have to match. +Different input pixel formats and color spaces will be automatically converted using hardware accelerated methods. +The final output will have the same pixel format and color space as the "main" input. + +The filter accepts the following options: + +@table @option + +@item x +Set the x coordinate of the overlaid video on the main video. +Default value is @code{0}. + +@item y +Set the y coordinate of the overlaid video on the main video. +Default value is @code{0}. + +@item eof_action +See @ref{framesync}. + +@item shortest +See @ref{framesync}. + +@item repeatlast +See @ref{framesync}. + +@end table + +@subsection Examples + +@itemize +@item +Overlay an image LOGO at the top-left corner of the INPUT video. +The INPUT video is in nv12 format and the LOGO image is in rgba format. +@example +-hwaccel videotoolbox -i INPUT -i LOGO -codec:v:0 h264_videotoolbox -filter_complex "[0:v]format=nv12,hwupload[a], [1:v]format=rgba,hwupload[b], [a][b]overlay_videotoolbox" OUTPUT +@end example +@item +Overlay an SDR video OVERLAY at the top-left corner of the HDR video MAIN. +The INPUT video is in p010 format and the LOGO image is in nv12 format. +The OUTPUT video will also be an HDR video with OVERLAY mapped to HDR. +@example +-hwaccel videotoolbox -i MAIN -i OVERLAY -codec:v:0 hevc_videotoolbox -tag:v hvc1 -filter_complex "[0:v]format=p010,hwupload[a], [1:v]format=nv12,hwupload[b], [a][b]overlay_videotoolbox" OUTPUT +@end example + +@end itemize + @section owdenoise Apply Overcomplete Wavelet denoiser. diff --git a/libavfilter/Makefile b/libavfilter/Makefile index f6c1d641d6..ea1389ab57 100644 --- a/libavfilter/Makefile +++ b/libavfilter/Makefile @@ -413,6 +413,9 @@ OBJS-$(CONFIG_OVERLAY_OPENCL_FILTER) += vf_overlay_opencl.o opencl.o \ OBJS-$(CONFIG_OVERLAY_QSV_FILTER) += vf_overlay_qsv.o framesync.o OBJS-$(CONFIG_OVERLAY_VAAPI_FILTER) += vf_overlay_vaapi.o framesync.o vaapi_vpp.o OBJS-$(CONFIG_OVERLAY_VULKAN_FILTER) += vf_overlay_vulkan.o vulkan.o vulkan_filter.o +OBJS-$(CONFIG_OVERLAY_VIDEOTOOLBOX_FILTER) += vf_overlay_videotoolbox.o framesync.o \ + metal/vf_overlay_videotoolbox.metallib.o \ + metal/utils.o OBJS-$(CONFIG_OWDENOISE_FILTER) += vf_owdenoise.o OBJS-$(CONFIG_PAD_FILTER) += vf_pad.o OBJS-$(CONFIG_PAD_OPENCL_FILTER) += vf_pad_opencl.o opencl.o opencl/pad.o diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c index 149bf50997..ec9d975ecb 100644 --- a/libavfilter/allfilters.c +++ b/libavfilter/allfilters.c @@ -389,6 +389,7 @@ extern const AVFilter ff_vf_overlay_qsv; extern const AVFilter ff_vf_overlay_vaapi; extern const AVFilter ff_vf_overlay_vulkan; extern const AVFilter ff_vf_overlay_cuda; +extern const AVFilter ff_vf_overlay_videotoolbox; extern const AVFilter ff_vf_owdenoise; extern const AVFilter ff_vf_pad; extern const AVFilter ff_vf_pad_opencl; diff --git a/libavfilter/metal/utils.m b/libavfilter/metal/utils.m index f365d3ceea..db5c5f6f10 100644 --- a/libavfilter/metal/utils.m +++ b/libavfilter/metal/utils.m @@ -55,6 +55,9 @@ CVMetalTextureRef ff_metal_texture_from_pixbuf(void *ctx, { CVMetalTextureRef tex = NULL; CVReturn ret; + bool is_planer = CVPixelBufferIsPlanar(pixbuf); + size_t width = is_planer ? CVPixelBufferGetWidthOfPlane(pixbuf, plane) : CVPixelBufferGetWidth(pixbuf); + size_t height = is_planer ? CVPixelBufferGetHeightOfPlane(pixbuf, plane) : CVPixelBufferGetHeight(pixbuf); ret = CVMetalTextureCacheCreateTextureFromImage( NULL, @@ -62,8 +65,8 @@ CVMetalTextureRef ff_metal_texture_from_pixbuf(void *ctx, pixbuf, NULL, format, - CVPixelBufferGetWidthOfPlane(pixbuf, plane), - CVPixelBufferGetHeightOfPlane(pixbuf, plane), + width, + height, plane, &tex ); diff --git a/libavfilter/metal/vf_overlay_videotoolbox.metal b/libavfilter/metal/vf_overlay_videotoolbox.metal new file mode 100644 index 0000000000..936e57e03e --- /dev/null +++ b/libavfilter/metal/vf_overlay_videotoolbox.metal @@ -0,0 +1,58 @@ +/* + * Copyright (C) 2024 Gnattu OC + * + * This file is part of FFmpeg. + * + * FFmpeg is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * FFmpeg is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with FFmpeg; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + */ + +#include +#include +#include + +using namespace metal; + +struct mtlBlendParams { + uint x_position; + uint y_position; +}; + +/* + * Blend shader for premultiplied alpha textures + */ +kernel void blend_shader( + texture2d source [[ texture(0) ]], + texture2d mask [[ texture(1) ]], + texture2d dest [[ texture(2) ]], + constant mtlBlendParams& params [[ buffer(3) ]], + uint2 gid [[ thread_position_in_grid ]]) +{ + const auto mask_size = uint2(mask.get_width(), + mask.get_height()); + const auto loc_overlay = uint2(params.x_position, params.y_position); + if (gid.x < loc_overlay.x || + gid.y < loc_overlay.y || + gid.x >= mask_size.x + loc_overlay.x || + gid.y >= mask_size.y + loc_overlay.y) + { + float4 source_color = source.read(gid); + dest.write(source_color, gid); + } else { + float4 source_color = source.read(gid); + float4 mask_color = mask.read((gid - loc_overlay)); + float4 result_color = source_color * (1.0f - mask_color.w) + (mask_color * mask_color.w); + dest.write(result_color, gid); + } +} diff --git a/libavfilter/vf_overlay_videotoolbox.m b/libavfilter/vf_overlay_videotoolbox.m new file mode 100644 index 0000000000..a6614c1f55 --- /dev/null +++ b/libavfilter/vf_overlay_videotoolbox.m @@ -0,0 +1,609 @@ +/* + * Copyright (C) 2024 Gnattu OC + * + * This file is part of FFmpeg. + * + * FFmpeg is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * FFmpeg is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with FFmpeg; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + */ + +#include +#include +#include "internal.h" +#include "metal/utils.h" +#include "framesync.h" +#include "libavutil/hwcontext.h" +#include "libavutil/pixdesc.h" +#include "libavutil/opt.h" +#include "libavutil/objc.h" +#include "video.h" + +#include + +extern char ff_vf_overlay_videotoolbox_metallib_data[]; +extern unsigned int ff_vf_overlay_videotoolbox_metallib_len; + +// Although iOS 8.0 introduced basic Metal support, its feature set is not complete and does not have CoreImage compatability. +// We have to set the minimum iOS version to 9.0. +typedef struct API_AVAILABLE(macos(10.11), ios(9.0)) OverlayVideoToolboxContext { + AVBufferRef *device_ref; + FFFrameSync fs; + + CVMetalTextureCacheRef texture_cache; + CVPixelBufferRef input_main_pixel_buffer_cache; + CVPixelBufferRef input_overlay_pixel_buffer_cache; + CVPixelBufferRef output_pixel_buffer_cache; + CIContext *ci_ctx; + VTPixelTransferSessionRef vt_session; + + id mtl_device; + id mtl_library; + id mtl_queue; + id mtl_pipeline; + id mtl_function; + id mtl_params_buffer; + + uint x_position; + uint y_position; + uint hwframe_ctx_allocated; +} OverlayVideoToolboxContext API_AVAILABLE(macos(10.11), ios(9.0)); + +typedef struct MtlBlendParams { + uint x_position; + uint y_position; +} MtlBlendParams; + +// Using sizeof(OverlayVideoToolboxContext) without an availability check will error +// if we're targeting an older OS version, so we need to calculate the size ourselves +// (we'll statically verify it's correct in overlay_videotoolbox_init behind a check) +#define OVERLAY_VT_CTX_SIZE (sizeof(FFFrameSync) + sizeof(uint) * 3 + sizeof(void*) * 13 + 4) + +// Neither VideoToolbox nor CoreImage can convert YUV420P frames into 16-bit depth color formats. +// Additionally, the only hardware formats that support an Alpha channel are AYUV64 and BGRA. +// However, neither can be directly manipulated with YUV420P frames. +// In such cases, the user will have to use NV12 instead. +static const enum AVPixelFormat supported_main_formats[] = { + AV_PIX_FMT_NV12, + AV_PIX_FMT_P010, + AV_PIX_FMT_NONE, +}; + +static const enum AVPixelFormat supported_overlay_formats[] = { + AV_PIX_FMT_NV12, + AV_PIX_FMT_P010, + AV_PIX_FMT_AYUV64, + AV_PIX_FMT_BGRA, + AV_PIX_FMT_NONE, +}; + +/** + * Helper to find out if provided format is supported by filter + */ +static int format_is_supported(const enum AVPixelFormat formats[], enum AVPixelFormat fmt) +{ + for (int i = 0; formats[i] != AV_PIX_FMT_NONE; i++) + if (formats[i] == fmt) + return 1; + return 0; +} + +static void call_kernel(AVFilterContext *avctx, + id dst, + id main, + id overlay, + uint x_position, + uint y_position) API_AVAILABLE(macos(10.11), ios(9.0)) +{ + OverlayVideoToolboxContext *ctx = avctx->priv; + // Both the command buffer and encoder are auto-released by objc on default. + // Use CFBridgingRetain to get a more C-like behavior. + id buffer = CFBridgingRetain(ctx->mtl_queue.commandBuffer); + id encoder = CFBridgingRetain((__bridge id)buffer.computeCommandEncoder); + + MtlBlendParams *params = (MtlBlendParams *)ctx->mtl_params_buffer.contents; + *params = (MtlBlendParams){ + .x_position = x_position, + .y_position = y_position, + }; + + [(__bridge id)encoder setTexture: main atIndex: 0]; + [(__bridge id)encoder setTexture: overlay atIndex: 1]; + [(__bridge id)encoder setTexture: dst atIndex: 2]; + [(__bridge id)encoder setBuffer: ctx->mtl_params_buffer offset: 0 atIndex: 3]; + ff_metal_compute_encoder_dispatch(ctx->mtl_device, ctx->mtl_pipeline, (__bridge id)encoder, dst.width, dst.height); + [(__bridge id)encoder endEncoding]; + + [(__bridge id)buffer commit]; + [(__bridge id)buffer waitUntilCompleted]; + + ff_objc_release(&encoder); + ff_objc_release(&buffer); +} + +// Copies and/or converts one pixel buffer to another. +// This transparently handles pixel format and color spaces, and will do a conversion if needed. +static int transfer_pixel_buffer(OverlayVideoToolboxContext *ctx, CVPixelBufferRef source, CVPixelBufferRef destination) +{ + if (@available(macOS 10.8, iOS 16.0, *)) { + int ret = 0; + ret = VTPixelTransferSessionTransferImage(ctx->vt_session, source, destination); + if (ret < 0) + return ret; + } else { + CIImage *temp_image = NULL; + temp_image = CFBridgingRetain([CIImage imageWithCVPixelBuffer: source]); + [(__bridge CIContext*)ctx->ci_ctx render: (__bridge CIImage*)temp_image toCVPixelBuffer: destination]; + CFRelease(temp_image); + CVBufferPropagateAttachments(source, destination); + } + return 0; +} + +static int overlay_vt_blend(FFFrameSync *fs) API_AVAILABLE(macos(10.11), ios(9.0)) +{ + AVFilterContext *avctx = fs->parent; + OverlayVideoToolboxContext *ctx = avctx->priv; + AVFilterLink *outlink = avctx->outputs[0]; + AVFilterLink *inlink_main = avctx->inputs[0]; + AVFilterLink *inlink_overlay = avctx->inputs[1]; + AVFrame *input_main, *input_overlay; + AVFrame *output; + AVHWFramesContext *frames_ctx = (AVHWFramesContext*)inlink_main->hw_frames_ctx->data; + AVHWFramesContext *frames_ctx_overlay = (AVHWFramesContext*)inlink_overlay->hw_frames_ctx->data; + const AVPixFmtDescriptor *in_overlay_desc; + + CVMetalTextureRef main, dst, overlay; + id tex_main, tex_overlay, tex_dst; + + MTLPixelFormat mtl_format; + OSType cv_format; + int ret; + int i, overlay_planes = 0; + + in_overlay_desc = av_pix_fmt_desc_get(frames_ctx_overlay->sw_format); + if (@available(macOS 11.3, iOS 14.2, *)) { + mtl_format = MTLPixelFormatRGBA16Unorm; + cv_format = kCVPixelFormatType_64RGBALE; + } else { + // On older OS versions, 64-bit RGBA with 16-bit little-endian full-range samples is not supported. + // To handle inputs with color depth greater than 8, convert colors to float type during filtering on these versions. + mtl_format = MTLPixelFormatRGBA16Float; + cv_format = kCVPixelFormatType_64RGBAHalf; + } + + // read main and overlay frames from inputs + ret = ff_framesync_get_frame(fs, 0, &input_main, 0); + if (ret < 0) + return ret; + ret = ff_framesync_get_frame(fs, 1, &input_overlay, 0); + if (ret < 0) + return ret; + + if (!input_main) + return AVERROR_BUG; + + output = ff_get_video_buffer(outlink, outlink->w, outlink->h); + if (!output) + return AVERROR(ENOMEM); + + ret = av_frame_copy_props(output, input_main); + if (ret < 0) + return ret; + + if (!input_overlay) { + ret = transfer_pixel_buffer(ctx, (CVPixelBufferRef)input_main->data[3], (CVPixelBufferRef)output->data[3]); + if (ret < 0) + return ret; + return ff_filter_frame(outlink, output); + } + + for (i = 0; i < in_overlay_desc->nb_components; i++) + overlay_planes = FFMAX(overlay_planes, + in_overlay_desc->comp[i].plane + 1); + + if (!ctx->input_overlay_pixel_buffer_cache) { + ret = CVPixelBufferCreate(kCFAllocatorDefault, + CVPixelBufferGetWidthOfPlane((CVPixelBufferRef)input_overlay->data[3], 0), + CVPixelBufferGetHeightOfPlane((CVPixelBufferRef)input_overlay->data[3], 0), + cv_format, + (__bridge CFDictionaryRef)@{ + (NSString *)kCVPixelBufferCGImageCompatibilityKey: @(YES), + (NSString *)kCVPixelBufferMetalCompatibilityKey: @(YES) + }, + &ctx->input_overlay_pixel_buffer_cache); + if (ret < 0) + return ret; + } + + if (!ctx->input_main_pixel_buffer_cache) { + ret = CVPixelBufferCreate(kCFAllocatorDefault, + CVPixelBufferGetWidthOfPlane((CVPixelBufferRef)input_main->data[3], 0), + CVPixelBufferGetHeightOfPlane((CVPixelBufferRef)input_main->data[3], 0), + cv_format, + (__bridge CFDictionaryRef)@{ + (NSString *)kCVPixelBufferCGImageCompatibilityKey: @(YES), + (NSString *)kCVPixelBufferMetalCompatibilityKey: @(YES) + }, + &ctx->input_main_pixel_buffer_cache); + if (ret < 0) + return ret; + } + + if (!ctx->output_pixel_buffer_cache) { + ret = CVPixelBufferCreate(kCFAllocatorDefault, + CVPixelBufferGetWidthOfPlane((CVPixelBufferRef)input_main->data[3], 0), + CVPixelBufferGetHeightOfPlane((CVPixelBufferRef)input_main->data[3], 0), + cv_format, + (__bridge CFDictionaryRef)@{ + (NSString *)kCVPixelBufferCGImageCompatibilityKey: @(YES), + (NSString *)kCVPixelBufferMetalCompatibilityKey: @(YES) + }, + &ctx->output_pixel_buffer_cache); + if (ret < 0) + return ret; + } + + ret = transfer_pixel_buffer(ctx, (CVPixelBufferRef)input_main->data[3], ctx->input_main_pixel_buffer_cache); + if (ret < 0) + return ret; + + ret = transfer_pixel_buffer(ctx, (CVPixelBufferRef)input_overlay->data[3], ctx->input_overlay_pixel_buffer_cache); + if (ret < 0) + return ret; + + overlay = ff_metal_texture_from_pixbuf(avctx, ctx->texture_cache, ctx->input_overlay_pixel_buffer_cache, 0, mtl_format); + main = ff_metal_texture_from_pixbuf(avctx, ctx->texture_cache, ctx->input_main_pixel_buffer_cache, 0, mtl_format); + dst = ff_metal_texture_from_pixbuf(avctx, ctx->texture_cache, ctx->output_pixel_buffer_cache, 0, mtl_format); + + if (!overlay || !main || !dst) { + return AVERROR(ENOSYS); + } + + tex_main = CVMetalTextureGetTexture(main); + tex_overlay = CVMetalTextureGetTexture(overlay); + tex_dst = CVMetalTextureGetTexture(dst); + + call_kernel(avctx, tex_dst, tex_main, tex_overlay, ctx->x_position, ctx->y_position); + + ret = transfer_pixel_buffer(ctx, ctx->output_pixel_buffer_cache, (CVPixelBufferRef)output->data[3]); + if (ret < 0) { + CFRelease(main); + CFRelease(overlay); + CFRelease(dst); + return ret; + } + + CFRelease(main); + CFRelease(overlay); + CFRelease(dst); + + return ff_filter_frame(outlink, output); +} + +static av_cold void do_uninit(AVFilterContext *avctx) API_AVAILABLE(macos(10.11), ios(9.0)) +{ + OverlayVideoToolboxContext *ctx = avctx->priv; + + if (ctx->hwframe_ctx_allocated) { + av_buffer_unref(&ctx->device_ref); + ctx->hwframe_ctx_allocated = 0; + } + ff_framesync_uninit(&ctx->fs); + + if (ctx->ci_ctx) { + CFRelease(ctx->ci_ctx); + ctx->ci_ctx = NULL; + } + + ff_objc_release(&ctx->mtl_params_buffer); + ff_objc_release(&ctx->mtl_function); + ff_objc_release(&ctx->mtl_pipeline); + ff_objc_release(&ctx->mtl_queue); + ff_objc_release(&ctx->mtl_library); + ff_objc_release(&ctx->mtl_device); + + if (ctx->texture_cache) { + CFRelease(ctx->texture_cache); + ctx->texture_cache = NULL; + } + if (ctx->input_main_pixel_buffer_cache) { + CFRelease(ctx->input_main_pixel_buffer_cache); + ctx->input_main_pixel_buffer_cache = NULL; + } + if (ctx->input_overlay_pixel_buffer_cache) { + CFRelease(ctx->input_overlay_pixel_buffer_cache); + ctx->input_overlay_pixel_buffer_cache = NULL; + } + if (ctx->output_pixel_buffer_cache) { + CFRelease(ctx->output_pixel_buffer_cache); + ctx->output_pixel_buffer_cache = NULL; + } + if (ctx->vt_session) { + VTPixelTransferSessionInvalidate(ctx->vt_session); + CFRelease(ctx->vt_session); + ctx->vt_session = NULL; + } +} + +static av_cold void overlay_videotoolbox_uninit(AVFilterContext *ctx) +{ + if (@available(macOS 10.11, iOS 9.0, *)) { + do_uninit(ctx); + } +} + +static av_cold int do_init(AVFilterContext *avctx) API_AVAILABLE(macos(10.11), ios(9.0)) +{ + OverlayVideoToolboxContext *ctx = avctx->priv; + NSError *err = nil; + CVReturn ret; + dispatch_data_t libData; + + ctx->mtl_device = MTLCreateSystemDefaultDevice(); + if (!ctx->mtl_device) { + av_log(avctx, AV_LOG_ERROR, "Unable to find Metal device\n"); + goto fail; + } + + av_log(ctx, AV_LOG_INFO, "Using Metal device: %s\n", ctx->mtl_device.name.UTF8String); + + libData = dispatch_data_create( + ff_vf_overlay_videotoolbox_metallib_data, + ff_vf_overlay_videotoolbox_metallib_len, + nil, + nil); + + ctx->mtl_library = [ctx->mtl_device newLibraryWithData: libData error: &err]; + dispatch_release(libData); + libData = nil; + ctx->mtl_function = [ctx->mtl_library newFunctionWithName: @"blend_shader"]; + if (!ctx->mtl_function) { + av_log(avctx, AV_LOG_ERROR, "Failed to create Metal function!\n"); + goto fail; + } + + ctx->mtl_queue = ctx->mtl_device.newCommandQueue; + if (!ctx->mtl_queue) { + av_log(avctx, AV_LOG_ERROR, "Failed to create Metal command queue!\n"); + goto fail; + } + + ctx->mtl_pipeline = [ctx->mtl_device + newComputePipelineStateWithFunction: ctx->mtl_function + error: &err]; + if (err) { + av_log(ctx, AV_LOG_ERROR, "Failed to create Metal compute pipeline: %s\n", err.description.UTF8String); + goto fail; + } + + ctx->mtl_params_buffer = [ctx->mtl_device + newBufferWithLength: sizeof(MtlBlendParams) + options: MTLResourceStorageModeShared]; + if (!ctx->mtl_params_buffer) { + av_log(avctx, AV_LOG_ERROR, "Failed to create Metal buffer for parameters\n"); + goto fail; + } + + ret = CVMetalTextureCacheCreate( + NULL, + NULL, + ctx->mtl_device, + NULL, + &ctx->texture_cache + ); + if (ret != kCVReturnSuccess) { + av_log(avctx, AV_LOG_ERROR, "Failed to create CVMetalTextureCache: %d\n", ret); + goto fail; + } + + if (@available(macOS 10.8, iOS 16.0, *)) { + ret = VTPixelTransferSessionCreate(NULL, &ctx->vt_session); + if (ret != kCVReturnSuccess) { + av_log(avctx, AV_LOG_ERROR, "Failed to create VTPixelTransferSession: %d\n", ret); + goto fail; + } + } else { + // Use CoreImage as fallback for old OS. + // CoreImage has comparable performance to VTPixelTransferSession, but it supports less pixel formats than VTPixelTransferSession. + // Warn user about possible incorrect results. + av_log(avctx, AV_LOG_WARNING, "VTPixelTransferSessionTransferImage is not available on this OS version, fallback using CoreImage\n"); + av_log(avctx, AV_LOG_WARNING, "Try an overlay with BGRA format if you see no overlay\n"); + if (@available(macOS 10.15, iOS 13.0, *)) { + ctx->ci_ctx = CFBridgingRetain([CIContext contextWithMTLCommandQueue: ctx->mtl_queue]); + } else { + ctx->ci_ctx = CFBridgingRetain([CIContext contextWithMTLDevice: ctx->mtl_device]); + } + } + + ctx->fs.on_event = &overlay_vt_blend; + + return 0; +fail: + overlay_videotoolbox_uninit(avctx); + return AVERROR_EXTERNAL; +} + +static av_cold int overlay_videotoolbox_init(AVFilterContext *ctx) +{ + if (@available(macOS 10.11, iOS 9.0, *)) { + // Ensure we calculated OVERLAY_VT_CTX_SIZE correctly + static_assert(OVERLAY_VT_CTX_SIZE == sizeof(OverlayVideoToolboxContext), "Incorrect OVERLAY_VT_CTX_SIZE value!"); + return do_init(ctx); + } else { + av_log(ctx, AV_LOG_ERROR, "Metal is not available on this OS version\n"); + return AVERROR(ENOSYS); + } +} + +static int do_config_output(AVFilterLink *link) API_AVAILABLE(macos(10.11), ios(9.0)) +{ + AVFilterContext *avctx = link->src; + AVFilterLink *inlink_main = avctx->inputs[0]; + AVFilterLink *inlink_overlay = avctx->inputs[1]; + OverlayVideoToolboxContext *ctx = avctx->priv; + AVHWFramesContext *main_frames, *output_frames, *overlay_frames; + AVBufferRef *input_ref, *overlay_ref; + int ret = 0; + + if (!inlink_main->hw_frames_ctx || + !inlink_overlay->hw_frames_ctx) { + av_log(avctx, AV_LOG_ERROR, "An input HW frames reference is " + "required to associate the processing device.\n"); + return AVERROR(EINVAL); + } + + input_ref = inlink_main->hw_frames_ctx; + overlay_ref = inlink_overlay->hw_frames_ctx; + main_frames = (AVHWFramesContext*)input_ref->data; + overlay_frames = (AVHWFramesContext*)overlay_ref->data; + av_assert0(main_frames); + av_assert0(overlay_frames); + + if (!format_is_supported(supported_main_formats, main_frames->sw_format)) { + av_log(ctx, AV_LOG_ERROR, "Unsupported main input format: %s.\n", + av_get_pix_fmt_name(main_frames->sw_format)); + if (main_frames->sw_format == AV_PIX_FMT_YUV420P) { + av_log(ctx, AV_LOG_WARNING, "Hint: Use %s instead of %s.\n", + av_get_pix_fmt_name(AV_PIX_FMT_NV12), + av_get_pix_fmt_name(AV_PIX_FMT_YUV420P)); + } + return AVERROR(ENOSYS); + } + + if (!format_is_supported(supported_overlay_formats, overlay_frames->sw_format)) { + av_log(ctx, AV_LOG_ERROR, "Unsupported overlay input format: %s.\n", + av_get_pix_fmt_name(overlay_frames->sw_format)); + if (overlay_frames->sw_format == AV_PIX_FMT_YUV420P) { + av_log(ctx, AV_LOG_WARNING, "Hint: Use %s instead of %s.\n", + av_get_pix_fmt_name(AV_PIX_FMT_NV12), + av_get_pix_fmt_name(AV_PIX_FMT_YUV420P)); + } + return AVERROR(ENOSYS); + } + + ctx->device_ref = av_buffer_ref(main_frames->device_ref); + if (!ctx->device_ref) { + av_log(ctx, AV_LOG_ERROR, "A device reference create failed.\n"); + return AVERROR(ENOMEM); + } + + link->hw_frames_ctx = av_hwframe_ctx_alloc(ctx->device_ref); + if (!link->hw_frames_ctx) { + av_log(avctx, AV_LOG_ERROR, "Failed to create HW frame context " + "for output.\n"); + ret = AVERROR(ENOMEM); + return ret; + } + ctx->hwframe_ctx_allocated = 1; + + output_frames = (AVHWFramesContext*)link->hw_frames_ctx->data; + + output_frames->format = AV_PIX_FMT_VIDEOTOOLBOX; + output_frames->sw_format = main_frames->sw_format; + output_frames->width = inlink_main->w; + output_frames->height = inlink_main->h; + + ret = ff_filter_init_hw_frames(avctx, link, 1); + if (ret < 0) + return ret; + + ret = av_hwframe_ctx_init(link->hw_frames_ctx); + if (ret < 0) { + av_log(avctx, AV_LOG_ERROR, "Failed to initialise VideoToolbox frame " + "context for output: %d\n", ret); + return ret; + } + + link->time_base = inlink_main->time_base; + ctx->fs.time_base = link->time_base; + + ret = ff_framesync_init_dualinput(&ctx->fs, avctx); + if (ret < 0) + return ret; + + ret = ff_framesync_configure(&ctx->fs); + return ret; +} + +static int config_output(AVFilterLink *link) +{ + AVFilterContext *ctx = link->src; + if (@available(macOS 10.11, iOS 9.0, *)) { + return do_config_output(link); + } else { + av_log(ctx, AV_LOG_ERROR, "Metal is not available on this OS version\n"); + return AVERROR(ENOSYS); + } +} + +static int overlay_videotoolbox_activate(AVFilterContext *avctx) +{ + OverlayVideoToolboxContext *ctx = avctx->priv; + return ff_framesync_activate(&ctx->fs); +} + +#define FLAGS (AV_OPT_FLAG_FILTERING_PARAM | AV_OPT_FLAG_VIDEO_PARAM) +#define OFFSET(x) offsetof(OverlayVideoToolboxContext, x) + +static const AVOption overlay_videotoolbox_options[] = { + { "x", "Overlay x position", + OFFSET(x_position), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = FLAGS }, + { "y", "Overlay y position", + OFFSET(y_position), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, .flags = FLAGS }, + { "eof_action", "Action to take when encountering EOF from secondary input ", + OFFSET(fs.opt_eof_action), AV_OPT_TYPE_INT, { .i64 = EOF_ACTION_REPEAT }, + EOF_ACTION_REPEAT, EOF_ACTION_PASS, .flags = FLAGS, .unit = "eof_action" }, + { "repeat", "Repeat the previous frame.", 0, AV_OPT_TYPE_CONST, { .i64 = EOF_ACTION_REPEAT }, .flags = FLAGS, .unit = "eof_action" }, + { "endall", "End both streams.", 0, AV_OPT_TYPE_CONST, { .i64 = EOF_ACTION_ENDALL }, .flags = FLAGS, .unit = "eof_action" }, + { "pass", "Pass through the main input.", 0, AV_OPT_TYPE_CONST, { .i64 = EOF_ACTION_PASS }, .flags = FLAGS, .unit = "eof_action" }, + { "shortest", "force termination when the shortest input terminates", OFFSET(fs.opt_shortest), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, FLAGS }, + { "repeatlast", "repeat overlay of the last overlay frame", OFFSET(fs.opt_repeatlast), AV_OPT_TYPE_BOOL, { .i64 = 1 }, 0, 1, FLAGS }, + { NULL }, +}; + +FRAMESYNC_DEFINE_CLASS(overlay_videotoolbox, OverlayVideoToolboxContext, fs); + +static const AVFilterPad overlay_videotoolbox_inputs[] = { + { + .name = "main", + .type = AVMEDIA_TYPE_VIDEO, + }, + { + .name = "overlay", + .type = AVMEDIA_TYPE_VIDEO, + }, +}; + +static const AVFilterPad overlay_videotoolbox_outputs[] = { + { + .name = "default", + .type = AVMEDIA_TYPE_VIDEO, + .config_props = config_output, + }, +}; + +const AVFilter ff_vf_overlay_videotoolbox = { + .name = "overlay_videotoolbox", + .description = NULL_IF_CONFIG_SMALL("Overlay filter for VideoToolbox frames using Metal compute"), + .priv_size = OVERLAY_VT_CTX_SIZE, + .priv_class = &overlay_videotoolbox_class, + .init = overlay_videotoolbox_init, + .uninit = overlay_videotoolbox_uninit, + .activate = overlay_videotoolbox_activate, + .preinit = overlay_videotoolbox_framesync_preinit, + FILTER_SINGLE_PIXFMT(AV_PIX_FMT_VIDEOTOOLBOX), + FILTER_INPUTS(overlay_videotoolbox_inputs), + FILTER_OUTPUTS(overlay_videotoolbox_outputs), + .flags_internal = FF_FILTER_FLAG_HWFRAME_AWARE, +};