Message ID | tencent_2037D9B284AD942A73571E1B540CF335A606@qq.com |
---|---|
State | Accepted |
Headers | show |
Series | [FFmpeg-devel,v3] GSoC: Add guided filter | expand |
Context | Check | Description |
---|---|---|
andriy/x86_make | success | Make finished |
andriy/x86_make_fate | success | Make fate finished |
andriy/PPC64_make | success | Make finished |
andriy/PPC64_make_fate | success | Make fate finished |
> 2021年5月8日 下午6:06,Xuewei Meng <928826483@qq.com> 写道: > > V3: Add examples on how to use this filter, and improve the code style. > V2: Implement the slice-level parallelism for guided filter. > V1: Add the basic version of guided filter. > > Signed-off-by: Xuewei Meng <xwmeng96@gmail.com> > --- > doc/filters.texi | 38 +++++ > libavfilter/Makefile | 1 + > libavfilter/allfilters.c | 1 + > libavfilter/vf_guided.c | 429 +++++++++++++++++++++++++++++++++++++++++++++++ > 4 files changed, 469 insertions(+) > create mode 100644 libavfilter/vf_guided.c > > diff --git a/doc/filters.texi b/doc/filters.texi > index 36e35a1..515c655 100644 > --- a/doc/filters.texi > +++ b/doc/filters.texi > @@ -12918,6 +12918,44 @@ greyedge=difford=1:minknorm=0:sigma=2 > > @end itemize > > +@section guided filter > +Apply guided filter for edge-preserving smoothing, dehazing and so on. > + > +The filter accepts the following options: > +@table @option > +@item radius > +Set the radius in pixels. > +Allowed range is 1 to 20. Default is 3. > + > +@item eps > +Set regularization parameter. > +Allowed range is 0 to 1. Default is 0.1. > + > +@item planes > +Set planes to filter. Default is first only. > +@end table > + > +@subsection Commands > +This filter supports the all above options as @ref{commands}. > + > +@subsection Examples > +@itemize > +@item > +Edge-preserving smoothing with guided filter: > +@example > +ffmpeg -i in.png -i in.png -filter_complex guided out.png > +@end example > + > +@item > +Dehazing, structure-transferring filtering, detail enhancement with guided filter. > +For the generation of guidance image, > +see @url{http://kaiminghe.com/publications/pami12guidedfilter.pdf}. > +@example > +ffmpeg -i in.png -i guidance.png -filter_complex guided out.png > +@end example > + > +@end itemize > + > @anchor{haldclut} > @section haldclut > > diff --git a/libavfilter/Makefile b/libavfilter/Makefile > index 5a28736..60a97e1 100644 > --- a/libavfilter/Makefile > +++ b/libavfilter/Makefile > @@ -285,6 +285,7 @@ OBJS-$(CONFIG_GEQ_FILTER) += vf_geq.o > OBJS-$(CONFIG_GRADFUN_FILTER) += vf_gradfun.o > OBJS-$(CONFIG_GRAPHMONITOR_FILTER) += f_graphmonitor.o > OBJS-$(CONFIG_GREYEDGE_FILTER) += vf_colorconstancy.o > +OBJS-$(CONFIG_GUIDED_FILTER) += vf_guided.o > OBJS-$(CONFIG_HALDCLUT_FILTER) += vf_lut3d.o framesync.o > OBJS-$(CONFIG_HFLIP_FILTER) += vf_hflip.o > OBJS-$(CONFIG_HISTEQ_FILTER) += vf_histeq.o > diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c > index 931d7db..962f656 100644 > --- a/libavfilter/allfilters.c > +++ b/libavfilter/allfilters.c > @@ -270,6 +270,7 @@ extern const AVFilter ff_vf_geq; > extern const AVFilter ff_vf_gradfun; > extern const AVFilter ff_vf_graphmonitor; > extern const AVFilter ff_vf_greyedge; > +extern const AVFilter ff_vf_guided; > extern const AVFilter ff_vf_haldclut; > extern const AVFilter ff_vf_hflip; > extern const AVFilter ff_vf_histeq; > diff --git a/libavfilter/vf_guided.c b/libavfilter/vf_guided.c > new file mode 100644 > index 0000000..bd706fb > --- /dev/null > +++ b/libavfilter/vf_guided.c > @@ -0,0 +1,429 @@ > +/* > + * Copyright (c) 2021 Xuewei Meng > + * > + * This file is part of FFmpeg. > + * > + * FFmpeg is free software; you can redistribute it and/or > + * modify it under the terms of the GNU Lesser General Public > + * License as published by the Free Software Foundation; either > + * version 2.1 of the License, or (at your option) any later version. > + * > + * FFmpeg is distributed in the hope that it will be useful, > + * but WITHOUT ANY WARRANTY; without even the implied warranty of > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU > + * Lesser General Public License for more details. > + * > + * You should have received a copy of the GNU Lesser General Public > + * License along with FFmpeg; if not, write to the Free Software > + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA > + */ > + > +#include "libavutil/imgutils.h" > +#include "libavutil/opt.h" > +#include "libavutil/pixdesc.h" > +#include "avfilter.h" > +#include "formats.h" > +#include "framesync.h" > +#include "internal.h" > +#include "video.h" > + > +typedef struct GuidedContext { > + const AVClass *class; > + FFFrameSync fs; > + > + int radius; > + float eps; > + > + int planes; > + > + int width; > + int height; > + > + int nb_planes; > + int depth; > + int planewidth[4]; > + int planeheight[4]; > + > + int (*box_slice)(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs); > +} GuidedContext; > + > +#define OFFSET(x) offsetof(GuidedContext, x) > +#define FLAGS AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_RUNTIME_PARAM > + > +static const AVOption guided_options[] = { > + { "radius", "set the box radius", OFFSET(radius), AV_OPT_TYPE_INT, {.i64=3 }, 1, 20, FLAGS }, > + { "eps", "set the regularization parameter (with square)", OFFSET(eps), AV_OPT_TYPE_FLOAT, {.dbl=0.01 }, 0.0, 1, FLAGS }, > + { "planes", "set planes to filter", OFFSET(planes), AV_OPT_TYPE_INT, {.i64=1 }, 0, 0xF, FLAGS }, > + { NULL } > +}; > + > +AVFILTER_DEFINE_CLASS(guided); > + > +typedef struct ThreadData { > + int width; > + int height; > + float *src; > + float *dst; > + int srcStride; > + int dstStride; > +} ThreadData; > + > +static int box_slice(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs) > +{ > + GuidedContext *s = ctx->priv; > + ThreadData *t = arg; > + > + const int width = t->width; > + const int height = t->height; > + const int src_stride = t->srcStride; > + const int dst_stride = t->dstStride; > + const int slice_start = (height * jobnr) / nb_jobs; > + const int slice_end = (height * (jobnr + 1)) / nb_jobs; > + const int radius = s->radius; > + const float *src = t->src; > + float *dst = t->dst; > + > + int w; > + int numPix; > + w = (radius << 1) + 1; > + numPix = w * w; > + for (int i = slice_start;i < slice_end;i++) { > + for (int j = 0;j < width;j++) { > + float temp = 0.0; > + for (int row = -radius;row <= radius;row++) { > + for (int col = -radius;col <= radius;col++) { > + int x = i + row; > + int y = j + col; > + x = (x < 0) ? 0 : (x >= height ? height - 1 : x); > + y = (y < 0) ? 0 : (y >= width ? width - 1 : y); > + temp += src[x * src_stride + y]; > + } > + } > + dst[i * dst_stride + j] = temp / numPix; > + } > + } > + return 0; > +} > + > +static int query_formats(AVFilterContext *ctx) > +{ > + static const enum AVPixelFormat pix_fmts[] = { > + AV_PIX_FMT_YUVA444P, AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV440P, > + AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ440P, > + AV_PIX_FMT_YUVA422P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUVA420P, AV_PIX_FMT_YUV420P, > + AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUVJ420P, > + AV_PIX_FMT_YUVJ411P, AV_PIX_FMT_YUV411P, AV_PIX_FMT_YUV410P, > + AV_PIX_FMT_YUV420P9, AV_PIX_FMT_YUV422P9, AV_PIX_FMT_YUV444P9, > + AV_PIX_FMT_YUV420P10, AV_PIX_FMT_YUV422P10, AV_PIX_FMT_YUV444P10, > + AV_PIX_FMT_YUV420P12, AV_PIX_FMT_YUV422P12, AV_PIX_FMT_YUV444P12, AV_PIX_FMT_YUV440P12, > + AV_PIX_FMT_YUV420P14, AV_PIX_FMT_YUV422P14, AV_PIX_FMT_YUV444P14, > + AV_PIX_FMT_YUV420P16, AV_PIX_FMT_YUV422P16, AV_PIX_FMT_YUV444P16, > + AV_PIX_FMT_YUVA420P9, AV_PIX_FMT_YUVA422P9, AV_PIX_FMT_YUVA444P9, > + AV_PIX_FMT_YUVA420P10, AV_PIX_FMT_YUVA422P10, AV_PIX_FMT_YUVA444P10, > + AV_PIX_FMT_YUVA420P16, AV_PIX_FMT_YUVA422P16, AV_PIX_FMT_YUVA444P16, > + AV_PIX_FMT_GBRP, AV_PIX_FMT_GBRP9, AV_PIX_FMT_GBRP10, > + AV_PIX_FMT_GBRP12, AV_PIX_FMT_GBRP14, AV_PIX_FMT_GBRP16, > + AV_PIX_FMT_GBRAP, AV_PIX_FMT_GBRAP10, AV_PIX_FMT_GBRAP12, AV_PIX_FMT_GBRAP16, > + AV_PIX_FMT_GRAY8, AV_PIX_FMT_GRAY9, AV_PIX_FMT_GRAY10, AV_PIX_FMT_GRAY12, AV_PIX_FMT_GRAY14, AV_PIX_FMT_GRAY16, > + AV_PIX_FMT_NONE > + }; > + > + return ff_set_common_formats(ctx, ff_make_format_list(pix_fmts)); > +} > + > +static int config_input(AVFilterLink *inlink) > +{ > + AVFilterContext *ctx = inlink->dst; > + GuidedContext *s = ctx->priv; > + const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(inlink->format); > + > + if (ctx->inputs[0]->w != ctx->inputs[1]->w || > + ctx->inputs[0]->h != ctx->inputs[1]->h) { > + av_log(ctx, AV_LOG_ERROR, "Width and height of input videos must be same.\n"); > + return AVERROR(EINVAL); > + } > + if (ctx->inputs[0]->format != ctx->inputs[1]->format) { > + av_log(ctx, AV_LOG_ERROR, "Inputs must be of same pixel format.\n"); > + return AVERROR(EINVAL); > + } > + > + > + s->depth = desc->comp[0].depth; > + s->width = ctx->inputs[0]->w; > + s->height = ctx->inputs[0]->h; > + > + s->planewidth[1] = s->planewidth[2] = AV_CEIL_RSHIFT(inlink->w, desc->log2_chroma_w); > + s->planewidth[0] = s->planewidth[3] = inlink->w; > + s->planeheight[1] = s->planeheight[2] = AV_CEIL_RSHIFT(inlink->h, desc->log2_chroma_h); > + s->planeheight[0] = s->planeheight[3] = inlink->h; > + > + s->nb_planes = av_pix_fmt_count_planes(inlink->format); > + s->box_slice = box_slice; > + return 0; > +} > + > +#define GUIDED(type, name) \ > +static int guided_##name(AVFilterContext *ctx, GuidedContext *s, \ > + const uint8_t *ssrc, const uint8_t *ssrcRef, \ > + uint8_t *ddst, int radius, float eps, int width, int height, \ > + int src_stride, int src_ref_stride, int dst_stride, \ > + float maxval) \ > +{ \ > + int ret = 0; \ > + type *dst = (type *)ddst; \ > + const type *src = (const type *)ssrc; \ > + const type *srcRef = (const type *)ssrcRef; \ > + \ > + ThreadData t; \ > + const int nb_threads = ff_filter_get_nb_threads(ctx); \ > + float *I; \ > + float *II; \ > + float *P; \ > + float *IP; \ > + float *meanI; \ > + float *meanII; \ > + float *meanP; \ > + float *meanIP; \ > + float *A; \ > + float *B; \ > + float *meanA; \ > + float *meanB; \ > + \ > + I = av_calloc(width * height, sizeof(float)); \ > + II = av_calloc(width * height, sizeof(float)); \ > + P = av_calloc(width * height, sizeof(float)); \ > + IP = av_calloc(width * height, sizeof(float)); \ > + meanI = av_calloc(width * height, sizeof(float)); \ > + meanII = av_calloc(width * height, sizeof(float)); \ > + meanP = av_calloc(width * height, sizeof(float)); \ > + meanIP = av_calloc(width * height, sizeof(float)); \ > + \ > + A = av_calloc(width * height, sizeof(float)); \ > + B = av_calloc(width * height, sizeof(float)); \ > + meanA = av_calloc(width * height, sizeof(float)); \ > + meanB = av_calloc(width * height, sizeof(float)); \ > + \ > + if (!I || !II || !P || !IP || !meanI || !meanII || !meanP || \ > + !meanIP || !A || !B || !meanA || !meanB){ \ > + ret = AVERROR(ENOMEM); \ > + goto end; \ > + } \ > + for (int i = 0;i < height;i++) { \ > + for (int j = 0;j < width;j++) { \ > + int x = i * width + j; \ > + I[x] = src[i * src_stride + j] / maxval; \ > + II[x] = I[x] * I[x]; \ > + P[x] = srcRef[i * src_ref_stride + j] / maxval; \ > + IP[x] = I[x] * P[x]; \ > + } \ > + } \ > + \ > + t.width = width; \ > + t.height = height; \ > + t.srcStride = width; \ > + t.dstStride = width; \ > + t.src = I; \ > + t.dst = meanI; \ > + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ > + t.src = II; \ > + t.dst = meanII; \ > + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ > + t.src = P; \ > + t.dst = meanP; \ > + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ > + t.src = IP; \ > + t.dst = meanIP; \ > + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ > + \ > + for (int i = 0;i < height;i++) { \ > + for (int j = 0;j < width;j++) { \ > + int x = i * width + j; \ > + float varI = meanII[x] - (meanI[x] * meanI[x]); \ > + float covIP = meanIP[x] - (meanI[x] * meanP[x]); \ > + A[x] = covIP / (varI + eps); \ > + B[x] = meanP[x] - A[x] * meanI[x]; \ > + } \ > + } \ > + \ > + t.src = A; \ > + t.dst = meanA; \ > + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ > + t.src = B; \ > + t.dst = meanB; \ > + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ > + \ > + for (int i = 0;i < height;i++) { \ > + for (int j = 0;j < width;j++) { \ > + int x = i * width + j; \ > + dst[i * dst_stride + j] = meanA[x] * src[i * src_stride + j] + \ > + meanB[x] * maxval; \ > + } \ > + } \ > +end: \ > + av_freep(&I); \ > + av_freep(&II); \ > + av_freep(&P); \ > + av_freep(&IP); \ > + av_freep(&meanI); \ > + av_freep(&meanII); \ > + av_freep(&meanP); \ > + av_freep(&meanIP); \ > + av_freep(&A); \ > + av_freep(&B); \ > + av_freep(&meanA); \ > + av_freep(&meanB); \ > + return ret; \ > +} > + > +GUIDED(uint8_t, byte) > +GUIDED(uint16_t, word) > + > +static int process_frame(FFFrameSync *fs) > +{ > + AVFilterContext *ctx = fs->parent; > + GuidedContext *s = fs->opaque; > + AVFilterLink *outlink = ctx->outputs[0]; > + AVFrame *out_frame = NULL, *main_frame = NULL, *ref_frame = NULL; > + int ret; > + > + ret = ff_framesync_dualinput_get(fs, &main_frame, &ref_frame); > + if (ret < 0) > + return ret; > + > + out_frame = ff_get_video_buffer(outlink, outlink->w, outlink->h); > + if (!out_frame) { > + av_frame_free(&main_frame); > + av_frame_free(&ref_frame); > + return AVERROR(ENOMEM); > + } > + av_frame_copy_props(out_frame, main_frame); > + > + if (ctx->is_disabled || !ref_frame) { > + av_frame_copy_props(ref_frame, main_frame); > + } > + > + for (int plane = 0; plane < s->nb_planes; plane++) { > + if (!(s->planes & (1 << plane))) { > + av_image_copy_plane(out_frame->data[plane], out_frame->linesize[plane], > + main_frame->data[plane], main_frame->linesize[plane], > + s->planewidth[plane] * ((s->depth + 7) / 8), s->planeheight[plane]); > + continue; > + } > + if (s->depth <= 8) > + guided_byte(ctx, s, main_frame->data[plane], ref_frame->data[plane], out_frame->data[plane], s->radius, s->eps, > + s->planewidth[plane], s->planeheight[plane], > + main_frame->linesize[plane], ref_frame->linesize[plane], out_frame->linesize[plane], (1 << s->depth) - 1.f); > + else > + guided_word(ctx, s, main_frame->data[plane], ref_frame->data[plane], out_frame->data[plane], s->radius, s->eps, > + s->planewidth[plane], s->planeheight[plane], > + main_frame->linesize[plane] / 2, ref_frame->linesize[plane] / 2, out_frame->linesize[plane] / 2, (1 << s->depth) - 1.f); > + } > + > + return ff_filter_frame(outlink, out_frame); > +} > + > + > +static int config_output(AVFilterLink *outlink) > +{ > + AVFilterContext *ctx = outlink->src; > + > + GuidedContext *s = ctx->priv; > + AVFilterLink *mainlink = ctx->inputs[0]; > + FFFrameSyncIn *in; > + int ret; > + > + > + outlink->w = mainlink->w; > + outlink->h = mainlink->h; > + outlink->time_base = mainlink->time_base; > + outlink->sample_aspect_ratio = mainlink->sample_aspect_ratio; > + outlink->frame_rate = mainlink->frame_rate; > + if ((ret = ff_framesync_init(&s->fs, ctx, 2)) < 0) > + return ret; > + > + outlink->time_base = s->fs.time_base; > + > + in = s->fs.in; > + in[0].time_base = mainlink->time_base; > + in[1].time_base = ctx->inputs[1]->time_base; > + in[0].sync = 2; > + in[0].before = EXT_INFINITY; > + in[0].after = EXT_INFINITY; > + in[1].sync = 1; > + in[1].before = EXT_INFINITY; > + in[1].after = EXT_INFINITY; > + s->fs.opaque = s; > + s->fs.on_event = process_frame; > + > + return ff_framesync_configure(&s->fs); > +} > + > +static int activate(AVFilterContext *ctx) > +{ > + GuidedContext *s = ctx->priv; > + return ff_framesync_activate(&s->fs); > +} > + > +static av_cold int init(AVFilterContext *ctx) > +{ > + return 0; > +} > + > +static av_cold void uninit(AVFilterContext *ctx) > +{ > + GuidedContext *s = ctx->priv; > + ff_framesync_uninit(&s->fs); > + return; > +} > + > + > +static int process_command(AVFilterContext *ctx, > + const char *cmd, > + const char *arg, > + char *res, > + int res_len, > + int flags) > +{ > + int ret = ff_filter_process_command(ctx, cmd, arg, res, res_len, flags); > + > + if (ret < 0) > + return ret; > + > + return 0; > +} > + > +static const AVFilterPad guided_inputs[] = { > + { > + .name = "main", > + .type = AVMEDIA_TYPE_VIDEO, > + },{ > + .name = "reference", > + .type = AVMEDIA_TYPE_VIDEO, > + .config_props = config_input, > + }, > + { NULL } > +}; > + > +static const AVFilterPad guided_outputs[] = { > + { > + .name = "default", > + .type = AVMEDIA_TYPE_VIDEO, > + .config_props = config_output, > + }, > + { NULL } > +}; > + > +AVFilter ff_vf_guided = { > + .name = "guided", > + .description = NULL_IF_CONFIG_SMALL("Apply Guided filter."), > + .init = init, > + .uninit = uninit, > + .query_formats = query_formats, > + .priv_size = sizeof(GuidedContext), > + .priv_class = &guided_class, > + .activate = activate, > + .inputs = guided_inputs, > + .outputs = guided_outputs, > + .flags = AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC | AVFILTER_FLAG_SLICE_THREADS, > + .process_command = process_command, > +}; > -- > 1.9.1 > > _______________________________________________ > ffmpeg-devel mailing list > ffmpeg-devel@ffmpeg.org > https://ffmpeg.org/mailman/listinfo/ffmpeg-devel > > To unsubscribe, visit link above, or email > ffmpeg-devel-request@ffmpeg.org with subject "unsubscribe". > Looks ok to me, waiting for other reviewers, Thanks Steven Liu
> 2021年5月8日 下午6:20,Steven Liu <lq@chinaffmpeg.org> 写道: > > > >> 2021年5月8日 下午6:06,Xuewei Meng <928826483@qq.com> 写道: >> >> V3: Add examples on how to use this filter, and improve the code style. >> V2: Implement the slice-level parallelism for guided filter. >> V1: Add the basic version of guided filter. >> >> Signed-off-by: Xuewei Meng <xwmeng96@gmail.com> >> --- >> doc/filters.texi | 38 +++++ >> libavfilter/Makefile | 1 + >> libavfilter/allfilters.c | 1 + >> libavfilter/vf_guided.c | 429 +++++++++++++++++++++++++++++++++++++++++++++++ >> 4 files changed, 469 insertions(+) >> create mode 100644 libavfilter/vf_guided.c >> >> diff --git a/doc/filters.texi b/doc/filters.texi >> index 36e35a1..515c655 100644 >> --- a/doc/filters.texi >> +++ b/doc/filters.texi >> @@ -12918,6 +12918,44 @@ greyedge=difford=1:minknorm=0:sigma=2 >> >> @end itemize >> >> +@section guided filter >> +Apply guided filter for edge-preserving smoothing, dehazing and so on. >> + >> +The filter accepts the following options: >> +@table @option >> +@item radius >> +Set the radius in pixels. >> +Allowed range is 1 to 20. Default is 3. >> + >> +@item eps >> +Set regularization parameter. >> +Allowed range is 0 to 1. Default is 0.1. >> + >> +@item planes >> +Set planes to filter. Default is first only. >> +@end table >> + >> +@subsection Commands >> +This filter supports the all above options as @ref{commands}. >> + >> +@subsection Examples >> +@itemize >> +@item >> +Edge-preserving smoothing with guided filter: >> +@example >> +ffmpeg -i in.png -i in.png -filter_complex guided out.png >> +@end example >> + >> +@item >> +Dehazing, structure-transferring filtering, detail enhancement with guided filter. >> +For the generation of guidance image, >> +see @url{http://kaiminghe.com/publications/pami12guidedfilter.pdf}. >> +@example >> +ffmpeg -i in.png -i guidance.png -filter_complex guided out.png >> +@end example >> + >> +@end itemize >> + >> @anchor{haldclut} >> @section haldclut >> >> diff --git a/libavfilter/Makefile b/libavfilter/Makefile >> index 5a28736..60a97e1 100644 >> --- a/libavfilter/Makefile >> +++ b/libavfilter/Makefile >> @@ -285,6 +285,7 @@ OBJS-$(CONFIG_GEQ_FILTER) += vf_geq.o >> OBJS-$(CONFIG_GRADFUN_FILTER) += vf_gradfun.o >> OBJS-$(CONFIG_GRAPHMONITOR_FILTER) += f_graphmonitor.o >> OBJS-$(CONFIG_GREYEDGE_FILTER) += vf_colorconstancy.o >> +OBJS-$(CONFIG_GUIDED_FILTER) += vf_guided.o >> OBJS-$(CONFIG_HALDCLUT_FILTER) += vf_lut3d.o framesync.o >> OBJS-$(CONFIG_HFLIP_FILTER) += vf_hflip.o >> OBJS-$(CONFIG_HISTEQ_FILTER) += vf_histeq.o >> diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c >> index 931d7db..962f656 100644 >> --- a/libavfilter/allfilters.c >> +++ b/libavfilter/allfilters.c >> @@ -270,6 +270,7 @@ extern const AVFilter ff_vf_geq; >> extern const AVFilter ff_vf_gradfun; >> extern const AVFilter ff_vf_graphmonitor; >> extern const AVFilter ff_vf_greyedge; >> +extern const AVFilter ff_vf_guided; >> extern const AVFilter ff_vf_haldclut; >> extern const AVFilter ff_vf_hflip; >> extern const AVFilter ff_vf_histeq; >> diff --git a/libavfilter/vf_guided.c b/libavfilter/vf_guided.c >> new file mode 100644 >> index 0000000..bd706fb >> --- /dev/null >> +++ b/libavfilter/vf_guided.c >> @@ -0,0 +1,429 @@ >> +/* >> + * Copyright (c) 2021 Xuewei Meng >> + * >> + * This file is part of FFmpeg. >> + * >> + * FFmpeg is free software; you can redistribute it and/or >> + * modify it under the terms of the GNU Lesser General Public >> + * License as published by the Free Software Foundation; either >> + * version 2.1 of the License, or (at your option) any later version. >> + * >> + * FFmpeg is distributed in the hope that it will be useful, >> + * but WITHOUT ANY WARRANTY; without even the implied warranty of >> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU >> + * Lesser General Public License for more details. >> + * >> + * You should have received a copy of the GNU Lesser General Public >> + * License along with FFmpeg; if not, write to the Free Software >> + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA >> + */ >> + >> +#include "libavutil/imgutils.h" >> +#include "libavutil/opt.h" >> +#include "libavutil/pixdesc.h" >> +#include "avfilter.h" >> +#include "formats.h" >> +#include "framesync.h" >> +#include "internal.h" >> +#include "video.h" >> + >> +typedef struct GuidedContext { >> + const AVClass *class; >> + FFFrameSync fs; >> + >> + int radius; >> + float eps; >> + >> + int planes; >> + >> + int width; >> + int height; >> + >> + int nb_planes; >> + int depth; >> + int planewidth[4]; >> + int planeheight[4]; >> + >> + int (*box_slice)(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs); >> +} GuidedContext; >> + >> +#define OFFSET(x) offsetof(GuidedContext, x) >> +#define FLAGS AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_RUNTIME_PARAM >> + >> +static const AVOption guided_options[] = { >> + { "radius", "set the box radius", OFFSET(radius), AV_OPT_TYPE_INT, {.i64=3 }, 1, 20, FLAGS }, >> + { "eps", "set the regularization parameter (with square)", OFFSET(eps), AV_OPT_TYPE_FLOAT, {.dbl=0.01 }, 0.0, 1, FLAGS }, >> + { "planes", "set planes to filter", OFFSET(planes), AV_OPT_TYPE_INT, {.i64=1 }, 0, 0xF, FLAGS }, >> + { NULL } >> +}; >> + >> +AVFILTER_DEFINE_CLASS(guided); >> + >> +typedef struct ThreadData { >> + int width; >> + int height; >> + float *src; >> + float *dst; >> + int srcStride; >> + int dstStride; >> +} ThreadData; >> + >> +static int box_slice(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs) >> +{ >> + GuidedContext *s = ctx->priv; >> + ThreadData *t = arg; >> + >> + const int width = t->width; >> + const int height = t->height; >> + const int src_stride = t->srcStride; >> + const int dst_stride = t->dstStride; >> + const int slice_start = (height * jobnr) / nb_jobs; >> + const int slice_end = (height * (jobnr + 1)) / nb_jobs; >> + const int radius = s->radius; >> + const float *src = t->src; >> + float *dst = t->dst; >> + >> + int w; >> + int numPix; >> + w = (radius << 1) + 1; >> + numPix = w * w; >> + for (int i = slice_start;i < slice_end;i++) { >> + for (int j = 0;j < width;j++) { >> + float temp = 0.0; >> + for (int row = -radius;row <= radius;row++) { >> + for (int col = -radius;col <= radius;col++) { >> + int x = i + row; >> + int y = j + col; >> + x = (x < 0) ? 0 : (x >= height ? height - 1 : x); >> + y = (y < 0) ? 0 : (y >= width ? width - 1 : y); >> + temp += src[x * src_stride + y]; >> + } >> + } >> + dst[i * dst_stride + j] = temp / numPix; >> + } >> + } >> + return 0; >> +} >> + >> +static int query_formats(AVFilterContext *ctx) >> +{ >> + static const enum AVPixelFormat pix_fmts[] = { >> + AV_PIX_FMT_YUVA444P, AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV440P, >> + AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ440P, >> + AV_PIX_FMT_YUVA422P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUVA420P, AV_PIX_FMT_YUV420P, >> + AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUVJ420P, >> + AV_PIX_FMT_YUVJ411P, AV_PIX_FMT_YUV411P, AV_PIX_FMT_YUV410P, >> + AV_PIX_FMT_YUV420P9, AV_PIX_FMT_YUV422P9, AV_PIX_FMT_YUV444P9, >> + AV_PIX_FMT_YUV420P10, AV_PIX_FMT_YUV422P10, AV_PIX_FMT_YUV444P10, >> + AV_PIX_FMT_YUV420P12, AV_PIX_FMT_YUV422P12, AV_PIX_FMT_YUV444P12, AV_PIX_FMT_YUV440P12, >> + AV_PIX_FMT_YUV420P14, AV_PIX_FMT_YUV422P14, AV_PIX_FMT_YUV444P14, >> + AV_PIX_FMT_YUV420P16, AV_PIX_FMT_YUV422P16, AV_PIX_FMT_YUV444P16, >> + AV_PIX_FMT_YUVA420P9, AV_PIX_FMT_YUVA422P9, AV_PIX_FMT_YUVA444P9, >> + AV_PIX_FMT_YUVA420P10, AV_PIX_FMT_YUVA422P10, AV_PIX_FMT_YUVA444P10, >> + AV_PIX_FMT_YUVA420P16, AV_PIX_FMT_YUVA422P16, AV_PIX_FMT_YUVA444P16, >> + AV_PIX_FMT_GBRP, AV_PIX_FMT_GBRP9, AV_PIX_FMT_GBRP10, >> + AV_PIX_FMT_GBRP12, AV_PIX_FMT_GBRP14, AV_PIX_FMT_GBRP16, >> + AV_PIX_FMT_GBRAP, AV_PIX_FMT_GBRAP10, AV_PIX_FMT_GBRAP12, AV_PIX_FMT_GBRAP16, >> + AV_PIX_FMT_GRAY8, AV_PIX_FMT_GRAY9, AV_PIX_FMT_GRAY10, AV_PIX_FMT_GRAY12, AV_PIX_FMT_GRAY14, AV_PIX_FMT_GRAY16, >> + AV_PIX_FMT_NONE >> + }; >> + >> + return ff_set_common_formats(ctx, ff_make_format_list(pix_fmts)); >> +} >> + >> +static int config_input(AVFilterLink *inlink) >> +{ >> + AVFilterContext *ctx = inlink->dst; >> + GuidedContext *s = ctx->priv; >> + const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(inlink->format); >> + >> + if (ctx->inputs[0]->w != ctx->inputs[1]->w || >> + ctx->inputs[0]->h != ctx->inputs[1]->h) { >> + av_log(ctx, AV_LOG_ERROR, "Width and height of input videos must be same.\n"); >> + return AVERROR(EINVAL); >> + } >> + if (ctx->inputs[0]->format != ctx->inputs[1]->format) { >> + av_log(ctx, AV_LOG_ERROR, "Inputs must be of same pixel format.\n"); >> + return AVERROR(EINVAL); >> + } >> + >> + >> + s->depth = desc->comp[0].depth; >> + s->width = ctx->inputs[0]->w; >> + s->height = ctx->inputs[0]->h; >> + >> + s->planewidth[1] = s->planewidth[2] = AV_CEIL_RSHIFT(inlink->w, desc->log2_chroma_w); >> + s->planewidth[0] = s->planewidth[3] = inlink->w; >> + s->planeheight[1] = s->planeheight[2] = AV_CEIL_RSHIFT(inlink->h, desc->log2_chroma_h); >> + s->planeheight[0] = s->planeheight[3] = inlink->h; >> + >> + s->nb_planes = av_pix_fmt_count_planes(inlink->format); >> + s->box_slice = box_slice; >> + return 0; >> +} >> + >> +#define GUIDED(type, name) \ >> +static int guided_##name(AVFilterContext *ctx, GuidedContext *s, \ >> + const uint8_t *ssrc, const uint8_t *ssrcRef, \ >> + uint8_t *ddst, int radius, float eps, int width, int height, \ >> + int src_stride, int src_ref_stride, int dst_stride, \ >> + float maxval) \ >> +{ \ >> + int ret = 0; \ >> + type *dst = (type *)ddst; \ >> + const type *src = (const type *)ssrc; \ >> + const type *srcRef = (const type *)ssrcRef; \ >> + \ >> + ThreadData t; \ >> + const int nb_threads = ff_filter_get_nb_threads(ctx); \ >> + float *I; \ >> + float *II; \ >> + float *P; \ >> + float *IP; \ >> + float *meanI; \ >> + float *meanII; \ >> + float *meanP; \ >> + float *meanIP; \ >> + float *A; \ >> + float *B; \ >> + float *meanA; \ >> + float *meanB; \ >> + \ >> + I = av_calloc(width * height, sizeof(float)); \ >> + II = av_calloc(width * height, sizeof(float)); \ >> + P = av_calloc(width * height, sizeof(float)); \ >> + IP = av_calloc(width * height, sizeof(float)); \ >> + meanI = av_calloc(width * height, sizeof(float)); \ >> + meanII = av_calloc(width * height, sizeof(float)); \ >> + meanP = av_calloc(width * height, sizeof(float)); \ >> + meanIP = av_calloc(width * height, sizeof(float)); \ >> + \ >> + A = av_calloc(width * height, sizeof(float)); \ >> + B = av_calloc(width * height, sizeof(float)); \ >> + meanA = av_calloc(width * height, sizeof(float)); \ >> + meanB = av_calloc(width * height, sizeof(float)); \ >> + \ >> + if (!I || !II || !P || !IP || !meanI || !meanII || !meanP || \ >> + !meanIP || !A || !B || !meanA || !meanB){ \ >> + ret = AVERROR(ENOMEM); \ >> + goto end; \ >> + } \ >> + for (int i = 0;i < height;i++) { \ >> + for (int j = 0;j < width;j++) { \ >> + int x = i * width + j; \ >> + I[x] = src[i * src_stride + j] / maxval; \ >> + II[x] = I[x] * I[x]; \ >> + P[x] = srcRef[i * src_ref_stride + j] / maxval; \ >> + IP[x] = I[x] * P[x]; \ >> + } \ >> + } \ >> + \ >> + t.width = width; \ >> + t.height = height; \ >> + t.srcStride = width; \ >> + t.dstStride = width; \ >> + t.src = I; \ >> + t.dst = meanI; \ >> + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ >> + t.src = II; \ >> + t.dst = meanII; \ >> + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ >> + t.src = P; \ >> + t.dst = meanP; \ >> + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ >> + t.src = IP; \ >> + t.dst = meanIP; \ >> + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ >> + \ >> + for (int i = 0;i < height;i++) { \ >> + for (int j = 0;j < width;j++) { \ >> + int x = i * width + j; \ >> + float varI = meanII[x] - (meanI[x] * meanI[x]); \ >> + float covIP = meanIP[x] - (meanI[x] * meanP[x]); \ >> + A[x] = covIP / (varI + eps); \ >> + B[x] = meanP[x] - A[x] * meanI[x]; \ >> + } \ >> + } \ >> + \ >> + t.src = A; \ >> + t.dst = meanA; \ >> + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ >> + t.src = B; \ >> + t.dst = meanB; \ >> + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ >> + \ >> + for (int i = 0;i < height;i++) { \ >> + for (int j = 0;j < width;j++) { \ >> + int x = i * width + j; \ >> + dst[i * dst_stride + j] = meanA[x] * src[i * src_stride + j] + \ >> + meanB[x] * maxval; \ >> + } \ >> + } \ >> +end: \ >> + av_freep(&I); \ >> + av_freep(&II); \ >> + av_freep(&P); \ >> + av_freep(&IP); \ >> + av_freep(&meanI); \ >> + av_freep(&meanII); \ >> + av_freep(&meanP); \ >> + av_freep(&meanIP); \ >> + av_freep(&A); \ >> + av_freep(&B); \ >> + av_freep(&meanA); \ >> + av_freep(&meanB); \ >> + return ret; \ >> +} >> + >> +GUIDED(uint8_t, byte) >> +GUIDED(uint16_t, word) >> + >> +static int process_frame(FFFrameSync *fs) >> +{ >> + AVFilterContext *ctx = fs->parent; >> + GuidedContext *s = fs->opaque; >> + AVFilterLink *outlink = ctx->outputs[0]; >> + AVFrame *out_frame = NULL, *main_frame = NULL, *ref_frame = NULL; >> + int ret; >> + >> + ret = ff_framesync_dualinput_get(fs, &main_frame, &ref_frame); >> + if (ret < 0) >> + return ret; >> + >> + out_frame = ff_get_video_buffer(outlink, outlink->w, outlink->h); >> + if (!out_frame) { >> + av_frame_free(&main_frame); >> + av_frame_free(&ref_frame); >> + return AVERROR(ENOMEM); >> + } >> + av_frame_copy_props(out_frame, main_frame); >> + >> + if (ctx->is_disabled || !ref_frame) { >> + av_frame_copy_props(ref_frame, main_frame); >> + } >> + >> + for (int plane = 0; plane < s->nb_planes; plane++) { >> + if (!(s->planes & (1 << plane))) { >> + av_image_copy_plane(out_frame->data[plane], out_frame->linesize[plane], >> + main_frame->data[plane], main_frame->linesize[plane], >> + s->planewidth[plane] * ((s->depth + 7) / 8), s->planeheight[plane]); >> + continue; >> + } >> + if (s->depth <= 8) >> + guided_byte(ctx, s, main_frame->data[plane], ref_frame->data[plane], out_frame->data[plane], s->radius, s->eps, >> + s->planewidth[plane], s->planeheight[plane], >> + main_frame->linesize[plane], ref_frame->linesize[plane], out_frame->linesize[plane], (1 << s->depth) - 1.f); >> + else >> + guided_word(ctx, s, main_frame->data[plane], ref_frame->data[plane], out_frame->data[plane], s->radius, s->eps, >> + s->planewidth[plane], s->planeheight[plane], >> + main_frame->linesize[plane] / 2, ref_frame->linesize[plane] / 2, out_frame->linesize[plane] / 2, (1 << s->depth) - 1.f); >> + } >> + >> + return ff_filter_frame(outlink, out_frame); >> +} >> + >> + >> +static int config_output(AVFilterLink *outlink) >> +{ >> + AVFilterContext *ctx = outlink->src; >> + >> + GuidedContext *s = ctx->priv; >> + AVFilterLink *mainlink = ctx->inputs[0]; >> + FFFrameSyncIn *in; >> + int ret; >> + >> + >> + outlink->w = mainlink->w; >> + outlink->h = mainlink->h; >> + outlink->time_base = mainlink->time_base; >> + outlink->sample_aspect_ratio = mainlink->sample_aspect_ratio; >> + outlink->frame_rate = mainlink->frame_rate; >> + if ((ret = ff_framesync_init(&s->fs, ctx, 2)) < 0) >> + return ret; >> + >> + outlink->time_base = s->fs.time_base; >> + >> + in = s->fs.in; >> + in[0].time_base = mainlink->time_base; >> + in[1].time_base = ctx->inputs[1]->time_base; >> + in[0].sync = 2; >> + in[0].before = EXT_INFINITY; >> + in[0].after = EXT_INFINITY; >> + in[1].sync = 1; >> + in[1].before = EXT_INFINITY; >> + in[1].after = EXT_INFINITY; >> + s->fs.opaque = s; >> + s->fs.on_event = process_frame; >> + >> + return ff_framesync_configure(&s->fs); >> +} >> + >> +static int activate(AVFilterContext *ctx) >> +{ >> + GuidedContext *s = ctx->priv; >> + return ff_framesync_activate(&s->fs); >> +} >> + >> +static av_cold int init(AVFilterContext *ctx) >> +{ >> + return 0; >> +} >> + >> +static av_cold void uninit(AVFilterContext *ctx) >> +{ >> + GuidedContext *s = ctx->priv; >> + ff_framesync_uninit(&s->fs); >> + return; >> +} >> + >> + >> +static int process_command(AVFilterContext *ctx, >> + const char *cmd, >> + const char *arg, >> + char *res, >> + int res_len, >> + int flags) >> +{ >> + int ret = ff_filter_process_command(ctx, cmd, arg, res, res_len, flags); >> + >> + if (ret < 0) >> + return ret; >> + >> + return 0; >> +} >> + >> +static const AVFilterPad guided_inputs[] = { >> + { >> + .name = "main", >> + .type = AVMEDIA_TYPE_VIDEO, >> + },{ >> + .name = "reference", >> + .type = AVMEDIA_TYPE_VIDEO, >> + .config_props = config_input, >> + }, >> + { NULL } >> +}; >> + >> +static const AVFilterPad guided_outputs[] = { >> + { >> + .name = "default", >> + .type = AVMEDIA_TYPE_VIDEO, >> + .config_props = config_output, >> + }, >> + { NULL } >> +}; >> + >> +AVFilter ff_vf_guided = { >> + .name = "guided", >> + .description = NULL_IF_CONFIG_SMALL("Apply Guided filter."), >> + .init = init, >> + .uninit = uninit, >> + .query_formats = query_formats, >> + .priv_size = sizeof(GuidedContext), >> + .priv_class = &guided_class, >> + .activate = activate, >> + .inputs = guided_inputs, >> + .outputs = guided_outputs, >> + .flags = AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC | AVFILTER_FLAG_SLICE_THREADS, >> + .process_command = process_command, >> +}; >> -- >> 1.9.1 >> >> _______________________________________________ >> ffmpeg-devel mailing list >> ffmpeg-devel@ffmpeg.org >> https://ffmpeg.org/mailman/listinfo/ffmpeg-devel >> >> To unsubscribe, visit link above, or email >> ffmpeg-devel-request@ffmpeg.org with subject "unsubscribe". >> > > Looks ok to me, waiting for other reviewers, Pushed Thanks Steven Liu
Xuewei Meng: > V3: Add examples on how to use this filter, and improve the code style. > V2: Implement the slice-level parallelism for guided filter. > V1: Add the basic version of guided filter. > > Signed-off-by: Xuewei Meng <xwmeng96@gmail.com> > --- > doc/filters.texi | 38 +++++ > libavfilter/Makefile | 1 + > libavfilter/allfilters.c | 1 + > libavfilter/vf_guided.c | 429 +++++++++++++++++++++++++++++++++++++++++++++++ > 4 files changed, 469 insertions(+) > create mode 100644 libavfilter/vf_guided.c > > diff --git a/doc/filters.texi b/doc/filters.texi > index 36e35a1..515c655 100644 > --- a/doc/filters.texi > +++ b/doc/filters.texi > @@ -12918,6 +12918,44 @@ greyedge=difford=1:minknorm=0:sigma=2 > > @end itemize > > +@section guided filter > +Apply guided filter for edge-preserving smoothing, dehazing and so on. > + > +The filter accepts the following options: > +@table @option > +@item radius > +Set the radius in pixels. > +Allowed range is 1 to 20. Default is 3. > + > +@item eps > +Set regularization parameter. > +Allowed range is 0 to 1. Default is 0.1. > + > +@item planes > +Set planes to filter. Default is first only. > +@end table > + > +@subsection Commands > +This filter supports the all above options as @ref{commands}. > + > +@subsection Examples > +@itemize > +@item > +Edge-preserving smoothing with guided filter: > +@example > +ffmpeg -i in.png -i in.png -filter_complex guided out.png > +@end example > + > +@item > +Dehazing, structure-transferring filtering, detail enhancement with guided filter. > +For the generation of guidance image, > +see @url{http://kaiminghe.com/publications/pami12guidedfilter.pdf}. > +@example > +ffmpeg -i in.png -i guidance.png -filter_complex guided out.png > +@end example > + > +@end itemize > + > @anchor{haldclut} > @section haldclut > > diff --git a/libavfilter/Makefile b/libavfilter/Makefile > index 5a28736..60a97e1 100644 > --- a/libavfilter/Makefile > +++ b/libavfilter/Makefile > @@ -285,6 +285,7 @@ OBJS-$(CONFIG_GEQ_FILTER) += vf_geq.o > OBJS-$(CONFIG_GRADFUN_FILTER) += vf_gradfun.o > OBJS-$(CONFIG_GRAPHMONITOR_FILTER) += f_graphmonitor.o > OBJS-$(CONFIG_GREYEDGE_FILTER) += vf_colorconstancy.o > +OBJS-$(CONFIG_GUIDED_FILTER) += vf_guided.o > OBJS-$(CONFIG_HALDCLUT_FILTER) += vf_lut3d.o framesync.o > OBJS-$(CONFIG_HFLIP_FILTER) += vf_hflip.o > OBJS-$(CONFIG_HISTEQ_FILTER) += vf_histeq.o > diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c > index 931d7db..962f656 100644 > --- a/libavfilter/allfilters.c > +++ b/libavfilter/allfilters.c > @@ -270,6 +270,7 @@ extern const AVFilter ff_vf_geq; > extern const AVFilter ff_vf_gradfun; > extern const AVFilter ff_vf_graphmonitor; > extern const AVFilter ff_vf_greyedge; > +extern const AVFilter ff_vf_guided; > extern const AVFilter ff_vf_haldclut; > extern const AVFilter ff_vf_hflip; > extern const AVFilter ff_vf_histeq; > diff --git a/libavfilter/vf_guided.c b/libavfilter/vf_guided.c > new file mode 100644 > index 0000000..bd706fb > --- /dev/null > +++ b/libavfilter/vf_guided.c > @@ -0,0 +1,429 @@ > +/* > + * Copyright (c) 2021 Xuewei Meng > + * > + * This file is part of FFmpeg. > + * > + * FFmpeg is free software; you can redistribute it and/or > + * modify it under the terms of the GNU Lesser General Public > + * License as published by the Free Software Foundation; either > + * version 2.1 of the License, or (at your option) any later version. > + * > + * FFmpeg is distributed in the hope that it will be useful, > + * but WITHOUT ANY WARRANTY; without even the implied warranty of > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU > + * Lesser General Public License for more details. > + * > + * You should have received a copy of the GNU Lesser General Public > + * License along with FFmpeg; if not, write to the Free Software > + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA > + */ > + > +#include "libavutil/imgutils.h" > +#include "libavutil/opt.h" > +#include "libavutil/pixdesc.h" > +#include "avfilter.h" > +#include "formats.h" > +#include "framesync.h" > +#include "internal.h" > +#include "video.h" > + > +typedef struct GuidedContext { > + const AVClass *class; > + FFFrameSync fs; > + > + int radius; > + float eps; > + > + int planes; > + > + int width; > + int height; > + > + int nb_planes; > + int depth; > + int planewidth[4]; > + int planeheight[4]; > + > + int (*box_slice)(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs); > +} GuidedContext; > + > +#define OFFSET(x) offsetof(GuidedContext, x) > +#define FLAGS AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_RUNTIME_PARAM > + > +static const AVOption guided_options[] = { > + { "radius", "set the box radius", OFFSET(radius), AV_OPT_TYPE_INT, {.i64=3 }, 1, 20, FLAGS }, > + { "eps", "set the regularization parameter (with square)", OFFSET(eps), AV_OPT_TYPE_FLOAT, {.dbl=0.01 }, 0.0, 1, FLAGS }, > + { "planes", "set planes to filter", OFFSET(planes), AV_OPT_TYPE_INT, {.i64=1 }, 0, 0xF, FLAGS }, > + { NULL } > +}; > + > +AVFILTER_DEFINE_CLASS(guided); > + > +typedef struct ThreadData { > + int width; > + int height; > + float *src; > + float *dst; > + int srcStride; > + int dstStride; > +} ThreadData; > + > +static int box_slice(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs) > +{ > + GuidedContext *s = ctx->priv; > + ThreadData *t = arg; > + > + const int width = t->width; > + const int height = t->height; > + const int src_stride = t->srcStride; > + const int dst_stride = t->dstStride; > + const int slice_start = (height * jobnr) / nb_jobs; > + const int slice_end = (height * (jobnr + 1)) / nb_jobs; > + const int radius = s->radius; > + const float *src = t->src; > + float *dst = t->dst; > + > + int w; > + int numPix; > + w = (radius << 1) + 1; > + numPix = w * w; > + for (int i = slice_start;i < slice_end;i++) { > + for (int j = 0;j < width;j++) { > + float temp = 0.0; > + for (int row = -radius;row <= radius;row++) { > + for (int col = -radius;col <= radius;col++) { > + int x = i + row; > + int y = j + col; > + x = (x < 0) ? 0 : (x >= height ? height - 1 : x); > + y = (y < 0) ? 0 : (y >= width ? width - 1 : y); > + temp += src[x * src_stride + y]; > + } > + } > + dst[i * dst_stride + j] = temp / numPix; > + } > + } > + return 0; > +} > + > +static int query_formats(AVFilterContext *ctx) > +{ > + static const enum AVPixelFormat pix_fmts[] = { > + AV_PIX_FMT_YUVA444P, AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV440P, > + AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ440P, > + AV_PIX_FMT_YUVA422P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUVA420P, AV_PIX_FMT_YUV420P, > + AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUVJ420P, > + AV_PIX_FMT_YUVJ411P, AV_PIX_FMT_YUV411P, AV_PIX_FMT_YUV410P, > + AV_PIX_FMT_YUV420P9, AV_PIX_FMT_YUV422P9, AV_PIX_FMT_YUV444P9, > + AV_PIX_FMT_YUV420P10, AV_PIX_FMT_YUV422P10, AV_PIX_FMT_YUV444P10, > + AV_PIX_FMT_YUV420P12, AV_PIX_FMT_YUV422P12, AV_PIX_FMT_YUV444P12, AV_PIX_FMT_YUV440P12, > + AV_PIX_FMT_YUV420P14, AV_PIX_FMT_YUV422P14, AV_PIX_FMT_YUV444P14, > + AV_PIX_FMT_YUV420P16, AV_PIX_FMT_YUV422P16, AV_PIX_FMT_YUV444P16, > + AV_PIX_FMT_YUVA420P9, AV_PIX_FMT_YUVA422P9, AV_PIX_FMT_YUVA444P9, > + AV_PIX_FMT_YUVA420P10, AV_PIX_FMT_YUVA422P10, AV_PIX_FMT_YUVA444P10, > + AV_PIX_FMT_YUVA420P16, AV_PIX_FMT_YUVA422P16, AV_PIX_FMT_YUVA444P16, > + AV_PIX_FMT_GBRP, AV_PIX_FMT_GBRP9, AV_PIX_FMT_GBRP10, > + AV_PIX_FMT_GBRP12, AV_PIX_FMT_GBRP14, AV_PIX_FMT_GBRP16, > + AV_PIX_FMT_GBRAP, AV_PIX_FMT_GBRAP10, AV_PIX_FMT_GBRAP12, AV_PIX_FMT_GBRAP16, > + AV_PIX_FMT_GRAY8, AV_PIX_FMT_GRAY9, AV_PIX_FMT_GRAY10, AV_PIX_FMT_GRAY12, AV_PIX_FMT_GRAY14, AV_PIX_FMT_GRAY16, > + AV_PIX_FMT_NONE > + }; > + > + return ff_set_common_formats(ctx, ff_make_format_list(pix_fmts)); > +} > + > +static int config_input(AVFilterLink *inlink) > +{ > + AVFilterContext *ctx = inlink->dst; > + GuidedContext *s = ctx->priv; > + const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(inlink->format); > + > + if (ctx->inputs[0]->w != ctx->inputs[1]->w || > + ctx->inputs[0]->h != ctx->inputs[1]->h) { > + av_log(ctx, AV_LOG_ERROR, "Width and height of input videos must be same.\n"); > + return AVERROR(EINVAL); > + } > + if (ctx->inputs[0]->format != ctx->inputs[1]->format) { > + av_log(ctx, AV_LOG_ERROR, "Inputs must be of same pixel format.\n"); > + return AVERROR(EINVAL); > + } > + > + > + s->depth = desc->comp[0].depth; > + s->width = ctx->inputs[0]->w; > + s->height = ctx->inputs[0]->h; > + > + s->planewidth[1] = s->planewidth[2] = AV_CEIL_RSHIFT(inlink->w, desc->log2_chroma_w); > + s->planewidth[0] = s->planewidth[3] = inlink->w; > + s->planeheight[1] = s->planeheight[2] = AV_CEIL_RSHIFT(inlink->h, desc->log2_chroma_h); > + s->planeheight[0] = s->planeheight[3] = inlink->h; > + > + s->nb_planes = av_pix_fmt_count_planes(inlink->format); > + s->box_slice = box_slice; > + return 0; > +} > + > +#define GUIDED(type, name) \ > +static int guided_##name(AVFilterContext *ctx, GuidedContext *s, \ > + const uint8_t *ssrc, const uint8_t *ssrcRef, \ > + uint8_t *ddst, int radius, float eps, int width, int height, \ > + int src_stride, int src_ref_stride, int dst_stride, \ > + float maxval) \ > +{ \ > + int ret = 0; \ > + type *dst = (type *)ddst; \ > + const type *src = (const type *)ssrc; \ > + const type *srcRef = (const type *)ssrcRef; \ > + \ > + ThreadData t; \ > + const int nb_threads = ff_filter_get_nb_threads(ctx); \ > + float *I; \ > + float *II; \ > + float *P; \ > + float *IP; \ > + float *meanI; \ > + float *meanII; \ > + float *meanP; \ > + float *meanIP; \ > + float *A; \ > + float *B; \ > + float *meanA; \ > + float *meanB; \ > + \ > + I = av_calloc(width * height, sizeof(float)); \ > + II = av_calloc(width * height, sizeof(float)); \ > + P = av_calloc(width * height, sizeof(float)); \ > + IP = av_calloc(width * height, sizeof(float)); \ > + meanI = av_calloc(width * height, sizeof(float)); \ > + meanII = av_calloc(width * height, sizeof(float)); \ > + meanP = av_calloc(width * height, sizeof(float)); \ > + meanIP = av_calloc(width * height, sizeof(float)); \ > + \ > + A = av_calloc(width * height, sizeof(float)); \ > + B = av_calloc(width * height, sizeof(float)); \ > + meanA = av_calloc(width * height, sizeof(float)); \ > + meanB = av_calloc(width * height, sizeof(float)); \ > + \ > + if (!I || !II || !P || !IP || !meanI || !meanII || !meanP || \ > + !meanIP || !A || !B || !meanA || !meanB){ \ > + ret = AVERROR(ENOMEM); \ > + goto end; \ > + } \ > + for (int i = 0;i < height;i++) { \ > + for (int j = 0;j < width;j++) { \ > + int x = i * width + j; \ > + I[x] = src[i * src_stride + j] / maxval; \ > + II[x] = I[x] * I[x]; \ > + P[x] = srcRef[i * src_ref_stride + j] / maxval; \ > + IP[x] = I[x] * P[x]; \ > + } \ > + } \ > + \ > + t.width = width; \ > + t.height = height; \ > + t.srcStride = width; \ > + t.dstStride = width; \ > + t.src = I; \ > + t.dst = meanI; \ > + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ > + t.src = II; \ > + t.dst = meanII; \ > + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ > + t.src = P; \ > + t.dst = meanP; \ > + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ > + t.src = IP; \ > + t.dst = meanIP; \ > + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ > + \ > + for (int i = 0;i < height;i++) { \ > + for (int j = 0;j < width;j++) { \ > + int x = i * width + j; \ > + float varI = meanII[x] - (meanI[x] * meanI[x]); \ > + float covIP = meanIP[x] - (meanI[x] * meanP[x]); \ > + A[x] = covIP / (varI + eps); \ > + B[x] = meanP[x] - A[x] * meanI[x]; \ > + } \ > + } \ > + \ > + t.src = A; \ > + t.dst = meanA; \ > + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ > + t.src = B; \ > + t.dst = meanB; \ > + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ > + \ > + for (int i = 0;i < height;i++) { \ > + for (int j = 0;j < width;j++) { \ > + int x = i * width + j; \ > + dst[i * dst_stride + j] = meanA[x] * src[i * src_stride + j] + \ > + meanB[x] * maxval; \ > + } \ > + } \ > +end: \ > + av_freep(&I); \ > + av_freep(&II); \ > + av_freep(&P); \ > + av_freep(&IP); \ > + av_freep(&meanI); \ > + av_freep(&meanII); \ > + av_freep(&meanP); \ > + av_freep(&meanIP); \ > + av_freep(&A); \ > + av_freep(&B); \ > + av_freep(&meanA); \ > + av_freep(&meanB); \ > + return ret; \ > +} > + > +GUIDED(uint8_t, byte) > +GUIDED(uint16_t, word) > + > +static int process_frame(FFFrameSync *fs) > +{ > + AVFilterContext *ctx = fs->parent; > + GuidedContext *s = fs->opaque; > + AVFilterLink *outlink = ctx->outputs[0]; > + AVFrame *out_frame = NULL, *main_frame = NULL, *ref_frame = NULL; > + int ret; > + > + ret = ff_framesync_dualinput_get(fs, &main_frame, &ref_frame); > + if (ret < 0) > + return ret; > + > + out_frame = ff_get_video_buffer(outlink, outlink->w, outlink->h); > + if (!out_frame) { > + av_frame_free(&main_frame); > + av_frame_free(&ref_frame); > + return AVERROR(ENOMEM); > + } > + av_frame_copy_props(out_frame, main_frame); > + > + if (ctx->is_disabled || !ref_frame) { The "|| !ref_frame" makes no sense at all, because you get a guaranteed segfault if ref is indeed NULL. Is this supposed to be "&& ref_frame"? Anyway, this is Coverity ID 1484785. > + av_frame_copy_props(ref_frame, main_frame); > + } > + > + for (int plane = 0; plane < s->nb_planes; plane++) { > + if (!(s->planes & (1 << plane))) { > + av_image_copy_plane(out_frame->data[plane], out_frame->linesize[plane], > + main_frame->data[plane], main_frame->linesize[plane], > + s->planewidth[plane] * ((s->depth + 7) / 8), s->planeheight[plane]); > + continue; > + } > + if (s->depth <= 8) > + guided_byte(ctx, s, main_frame->data[plane], ref_frame->data[plane], out_frame->data[plane], s->radius, s->eps, > + s->planewidth[plane], s->planeheight[plane], > + main_frame->linesize[plane], ref_frame->linesize[plane], out_frame->linesize[plane], (1 << s->depth) - 1.f); > + else > + guided_word(ctx, s, main_frame->data[plane], ref_frame->data[plane], out_frame->data[plane], s->radius, s->eps, > + s->planewidth[plane], s->planeheight[plane], > + main_frame->linesize[plane] / 2, ref_frame->linesize[plane] / 2, out_frame->linesize[plane] / 2, (1 << s->depth) - 1.f); > + } > + > + return ff_filter_frame(outlink, out_frame); > +} > + > + > +static int config_output(AVFilterLink *outlink) > +{ > + AVFilterContext *ctx = outlink->src; > + > + GuidedContext *s = ctx->priv; > + AVFilterLink *mainlink = ctx->inputs[0]; > + FFFrameSyncIn *in; > + int ret; > + > + > + outlink->w = mainlink->w; > + outlink->h = mainlink->h; > + outlink->time_base = mainlink->time_base; > + outlink->sample_aspect_ratio = mainlink->sample_aspect_ratio; > + outlink->frame_rate = mainlink->frame_rate; > + if ((ret = ff_framesync_init(&s->fs, ctx, 2)) < 0) > + return ret; > + > + outlink->time_base = s->fs.time_base; > + > + in = s->fs.in; > + in[0].time_base = mainlink->time_base; > + in[1].time_base = ctx->inputs[1]->time_base; > + in[0].sync = 2; > + in[0].before = EXT_INFINITY; > + in[0].after = EXT_INFINITY; > + in[1].sync = 1; > + in[1].before = EXT_INFINITY; > + in[1].after = EXT_INFINITY; > + s->fs.opaque = s; > + s->fs.on_event = process_frame; > + > + return ff_framesync_configure(&s->fs); > +} > + > +static int activate(AVFilterContext *ctx) > +{ > + GuidedContext *s = ctx->priv; > + return ff_framesync_activate(&s->fs); > +} > + > +static av_cold int init(AVFilterContext *ctx) > +{ > + return 0; > +} > + > +static av_cold void uninit(AVFilterContext *ctx) > +{ > + GuidedContext *s = ctx->priv; > + ff_framesync_uninit(&s->fs); > + return; > +} > + > + > +static int process_command(AVFilterContext *ctx, > + const char *cmd, > + const char *arg, > + char *res, > + int res_len, > + int flags) > +{ > + int ret = ff_filter_process_command(ctx, cmd, arg, res, res_len, flags); > + > + if (ret < 0) > + return ret; > + > + return 0; > +} > + > +static const AVFilterPad guided_inputs[] = { > + { > + .name = "main", > + .type = AVMEDIA_TYPE_VIDEO, > + },{ > + .name = "reference", > + .type = AVMEDIA_TYPE_VIDEO, > + .config_props = config_input, > + }, > + { NULL } > +}; > + > +static const AVFilterPad guided_outputs[] = { > + { > + .name = "default", > + .type = AVMEDIA_TYPE_VIDEO, > + .config_props = config_output, > + }, > + { NULL } > +}; > + > +AVFilter ff_vf_guided = { > + .name = "guided", > + .description = NULL_IF_CONFIG_SMALL("Apply Guided filter."), > + .init = init, > + .uninit = uninit, > + .query_formats = query_formats, > + .priv_size = sizeof(GuidedContext), > + .priv_class = &guided_class, > + .activate = activate, > + .inputs = guided_inputs, > + .outputs = guided_outputs, > + .flags = AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC | AVFILTER_FLAG_SLICE_THREADS, > + .process_command = process_command, > +}; >
diff --git a/doc/filters.texi b/doc/filters.texi index 36e35a1..515c655 100644 --- a/doc/filters.texi +++ b/doc/filters.texi @@ -12918,6 +12918,44 @@ greyedge=difford=1:minknorm=0:sigma=2 @end itemize +@section guided filter +Apply guided filter for edge-preserving smoothing, dehazing and so on. + +The filter accepts the following options: +@table @option +@item radius +Set the radius in pixels. +Allowed range is 1 to 20. Default is 3. + +@item eps +Set regularization parameter. +Allowed range is 0 to 1. Default is 0.1. + +@item planes +Set planes to filter. Default is first only. +@end table + +@subsection Commands +This filter supports the all above options as @ref{commands}. + +@subsection Examples +@itemize +@item +Edge-preserving smoothing with guided filter: +@example +ffmpeg -i in.png -i in.png -filter_complex guided out.png +@end example + +@item +Dehazing, structure-transferring filtering, detail enhancement with guided filter. +For the generation of guidance image, +see @url{http://kaiminghe.com/publications/pami12guidedfilter.pdf}. +@example +ffmpeg -i in.png -i guidance.png -filter_complex guided out.png +@end example + +@end itemize + @anchor{haldclut} @section haldclut diff --git a/libavfilter/Makefile b/libavfilter/Makefile index 5a28736..60a97e1 100644 --- a/libavfilter/Makefile +++ b/libavfilter/Makefile @@ -285,6 +285,7 @@ OBJS-$(CONFIG_GEQ_FILTER) += vf_geq.o OBJS-$(CONFIG_GRADFUN_FILTER) += vf_gradfun.o OBJS-$(CONFIG_GRAPHMONITOR_FILTER) += f_graphmonitor.o OBJS-$(CONFIG_GREYEDGE_FILTER) += vf_colorconstancy.o +OBJS-$(CONFIG_GUIDED_FILTER) += vf_guided.o OBJS-$(CONFIG_HALDCLUT_FILTER) += vf_lut3d.o framesync.o OBJS-$(CONFIG_HFLIP_FILTER) += vf_hflip.o OBJS-$(CONFIG_HISTEQ_FILTER) += vf_histeq.o diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c index 931d7db..962f656 100644 --- a/libavfilter/allfilters.c +++ b/libavfilter/allfilters.c @@ -270,6 +270,7 @@ extern const AVFilter ff_vf_geq; extern const AVFilter ff_vf_gradfun; extern const AVFilter ff_vf_graphmonitor; extern const AVFilter ff_vf_greyedge; +extern const AVFilter ff_vf_guided; extern const AVFilter ff_vf_haldclut; extern const AVFilter ff_vf_hflip; extern const AVFilter ff_vf_histeq; diff --git a/libavfilter/vf_guided.c b/libavfilter/vf_guided.c new file mode 100644 index 0000000..bd706fb --- /dev/null +++ b/libavfilter/vf_guided.c @@ -0,0 +1,429 @@ +/* + * Copyright (c) 2021 Xuewei Meng + * + * This file is part of FFmpeg. + * + * FFmpeg is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * FFmpeg is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with FFmpeg; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + */ + +#include "libavutil/imgutils.h" +#include "libavutil/opt.h" +#include "libavutil/pixdesc.h" +#include "avfilter.h" +#include "formats.h" +#include "framesync.h" +#include "internal.h" +#include "video.h" + +typedef struct GuidedContext { + const AVClass *class; + FFFrameSync fs; + + int radius; + float eps; + + int planes; + + int width; + int height; + + int nb_planes; + int depth; + int planewidth[4]; + int planeheight[4]; + + int (*box_slice)(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs); +} GuidedContext; + +#define OFFSET(x) offsetof(GuidedContext, x) +#define FLAGS AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_RUNTIME_PARAM + +static const AVOption guided_options[] = { + { "radius", "set the box radius", OFFSET(radius), AV_OPT_TYPE_INT, {.i64=3 }, 1, 20, FLAGS }, + { "eps", "set the regularization parameter (with square)", OFFSET(eps), AV_OPT_TYPE_FLOAT, {.dbl=0.01 }, 0.0, 1, FLAGS }, + { "planes", "set planes to filter", OFFSET(planes), AV_OPT_TYPE_INT, {.i64=1 }, 0, 0xF, FLAGS }, + { NULL } +}; + +AVFILTER_DEFINE_CLASS(guided); + +typedef struct ThreadData { + int width; + int height; + float *src; + float *dst; + int srcStride; + int dstStride; +} ThreadData; + +static int box_slice(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs) +{ + GuidedContext *s = ctx->priv; + ThreadData *t = arg; + + const int width = t->width; + const int height = t->height; + const int src_stride = t->srcStride; + const int dst_stride = t->dstStride; + const int slice_start = (height * jobnr) / nb_jobs; + const int slice_end = (height * (jobnr + 1)) / nb_jobs; + const int radius = s->radius; + const float *src = t->src; + float *dst = t->dst; + + int w; + int numPix; + w = (radius << 1) + 1; + numPix = w * w; + for (int i = slice_start;i < slice_end;i++) { + for (int j = 0;j < width;j++) { + float temp = 0.0; + for (int row = -radius;row <= radius;row++) { + for (int col = -radius;col <= radius;col++) { + int x = i + row; + int y = j + col; + x = (x < 0) ? 0 : (x >= height ? height - 1 : x); + y = (y < 0) ? 0 : (y >= width ? width - 1 : y); + temp += src[x * src_stride + y]; + } + } + dst[i * dst_stride + j] = temp / numPix; + } + } + return 0; +} + +static int query_formats(AVFilterContext *ctx) +{ + static const enum AVPixelFormat pix_fmts[] = { + AV_PIX_FMT_YUVA444P, AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV440P, + AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ440P, + AV_PIX_FMT_YUVA422P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUVA420P, AV_PIX_FMT_YUV420P, + AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUVJ420P, + AV_PIX_FMT_YUVJ411P, AV_PIX_FMT_YUV411P, AV_PIX_FMT_YUV410P, + AV_PIX_FMT_YUV420P9, AV_PIX_FMT_YUV422P9, AV_PIX_FMT_YUV444P9, + AV_PIX_FMT_YUV420P10, AV_PIX_FMT_YUV422P10, AV_PIX_FMT_YUV444P10, + AV_PIX_FMT_YUV420P12, AV_PIX_FMT_YUV422P12, AV_PIX_FMT_YUV444P12, AV_PIX_FMT_YUV440P12, + AV_PIX_FMT_YUV420P14, AV_PIX_FMT_YUV422P14, AV_PIX_FMT_YUV444P14, + AV_PIX_FMT_YUV420P16, AV_PIX_FMT_YUV422P16, AV_PIX_FMT_YUV444P16, + AV_PIX_FMT_YUVA420P9, AV_PIX_FMT_YUVA422P9, AV_PIX_FMT_YUVA444P9, + AV_PIX_FMT_YUVA420P10, AV_PIX_FMT_YUVA422P10, AV_PIX_FMT_YUVA444P10, + AV_PIX_FMT_YUVA420P16, AV_PIX_FMT_YUVA422P16, AV_PIX_FMT_YUVA444P16, + AV_PIX_FMT_GBRP, AV_PIX_FMT_GBRP9, AV_PIX_FMT_GBRP10, + AV_PIX_FMT_GBRP12, AV_PIX_FMT_GBRP14, AV_PIX_FMT_GBRP16, + AV_PIX_FMT_GBRAP, AV_PIX_FMT_GBRAP10, AV_PIX_FMT_GBRAP12, AV_PIX_FMT_GBRAP16, + AV_PIX_FMT_GRAY8, AV_PIX_FMT_GRAY9, AV_PIX_FMT_GRAY10, AV_PIX_FMT_GRAY12, AV_PIX_FMT_GRAY14, AV_PIX_FMT_GRAY16, + AV_PIX_FMT_NONE + }; + + return ff_set_common_formats(ctx, ff_make_format_list(pix_fmts)); +} + +static int config_input(AVFilterLink *inlink) +{ + AVFilterContext *ctx = inlink->dst; + GuidedContext *s = ctx->priv; + const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(inlink->format); + + if (ctx->inputs[0]->w != ctx->inputs[1]->w || + ctx->inputs[0]->h != ctx->inputs[1]->h) { + av_log(ctx, AV_LOG_ERROR, "Width and height of input videos must be same.\n"); + return AVERROR(EINVAL); + } + if (ctx->inputs[0]->format != ctx->inputs[1]->format) { + av_log(ctx, AV_LOG_ERROR, "Inputs must be of same pixel format.\n"); + return AVERROR(EINVAL); + } + + + s->depth = desc->comp[0].depth; + s->width = ctx->inputs[0]->w; + s->height = ctx->inputs[0]->h; + + s->planewidth[1] = s->planewidth[2] = AV_CEIL_RSHIFT(inlink->w, desc->log2_chroma_w); + s->planewidth[0] = s->planewidth[3] = inlink->w; + s->planeheight[1] = s->planeheight[2] = AV_CEIL_RSHIFT(inlink->h, desc->log2_chroma_h); + s->planeheight[0] = s->planeheight[3] = inlink->h; + + s->nb_planes = av_pix_fmt_count_planes(inlink->format); + s->box_slice = box_slice; + return 0; +} + +#define GUIDED(type, name) \ +static int guided_##name(AVFilterContext *ctx, GuidedContext *s, \ + const uint8_t *ssrc, const uint8_t *ssrcRef, \ + uint8_t *ddst, int radius, float eps, int width, int height, \ + int src_stride, int src_ref_stride, int dst_stride, \ + float maxval) \ +{ \ + int ret = 0; \ + type *dst = (type *)ddst; \ + const type *src = (const type *)ssrc; \ + const type *srcRef = (const type *)ssrcRef; \ + \ + ThreadData t; \ + const int nb_threads = ff_filter_get_nb_threads(ctx); \ + float *I; \ + float *II; \ + float *P; \ + float *IP; \ + float *meanI; \ + float *meanII; \ + float *meanP; \ + float *meanIP; \ + float *A; \ + float *B; \ + float *meanA; \ + float *meanB; \ + \ + I = av_calloc(width * height, sizeof(float)); \ + II = av_calloc(width * height, sizeof(float)); \ + P = av_calloc(width * height, sizeof(float)); \ + IP = av_calloc(width * height, sizeof(float)); \ + meanI = av_calloc(width * height, sizeof(float)); \ + meanII = av_calloc(width * height, sizeof(float)); \ + meanP = av_calloc(width * height, sizeof(float)); \ + meanIP = av_calloc(width * height, sizeof(float)); \ + \ + A = av_calloc(width * height, sizeof(float)); \ + B = av_calloc(width * height, sizeof(float)); \ + meanA = av_calloc(width * height, sizeof(float)); \ + meanB = av_calloc(width * height, sizeof(float)); \ + \ + if (!I || !II || !P || !IP || !meanI || !meanII || !meanP || \ + !meanIP || !A || !B || !meanA || !meanB){ \ + ret = AVERROR(ENOMEM); \ + goto end; \ + } \ + for (int i = 0;i < height;i++) { \ + for (int j = 0;j < width;j++) { \ + int x = i * width + j; \ + I[x] = src[i * src_stride + j] / maxval; \ + II[x] = I[x] * I[x]; \ + P[x] = srcRef[i * src_ref_stride + j] / maxval; \ + IP[x] = I[x] * P[x]; \ + } \ + } \ + \ + t.width = width; \ + t.height = height; \ + t.srcStride = width; \ + t.dstStride = width; \ + t.src = I; \ + t.dst = meanI; \ + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ + t.src = II; \ + t.dst = meanII; \ + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ + t.src = P; \ + t.dst = meanP; \ + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ + t.src = IP; \ + t.dst = meanIP; \ + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ + \ + for (int i = 0;i < height;i++) { \ + for (int j = 0;j < width;j++) { \ + int x = i * width + j; \ + float varI = meanII[x] - (meanI[x] * meanI[x]); \ + float covIP = meanIP[x] - (meanI[x] * meanP[x]); \ + A[x] = covIP / (varI + eps); \ + B[x] = meanP[x] - A[x] * meanI[x]; \ + } \ + } \ + \ + t.src = A; \ + t.dst = meanA; \ + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ + t.src = B; \ + t.dst = meanB; \ + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \ + \ + for (int i = 0;i < height;i++) { \ + for (int j = 0;j < width;j++) { \ + int x = i * width + j; \ + dst[i * dst_stride + j] = meanA[x] * src[i * src_stride + j] + \ + meanB[x] * maxval; \ + } \ + } \ +end: \ + av_freep(&I); \ + av_freep(&II); \ + av_freep(&P); \ + av_freep(&IP); \ + av_freep(&meanI); \ + av_freep(&meanII); \ + av_freep(&meanP); \ + av_freep(&meanIP); \ + av_freep(&A); \ + av_freep(&B); \ + av_freep(&meanA); \ + av_freep(&meanB); \ + return ret; \ +} + +GUIDED(uint8_t, byte) +GUIDED(uint16_t, word) + +static int process_frame(FFFrameSync *fs) +{ + AVFilterContext *ctx = fs->parent; + GuidedContext *s = fs->opaque; + AVFilterLink *outlink = ctx->outputs[0]; + AVFrame *out_frame = NULL, *main_frame = NULL, *ref_frame = NULL; + int ret; + + ret = ff_framesync_dualinput_get(fs, &main_frame, &ref_frame); + if (ret < 0) + return ret; + + out_frame = ff_get_video_buffer(outlink, outlink->w, outlink->h); + if (!out_frame) { + av_frame_free(&main_frame); + av_frame_free(&ref_frame); + return AVERROR(ENOMEM); + } + av_frame_copy_props(out_frame, main_frame); + + if (ctx->is_disabled || !ref_frame) { + av_frame_copy_props(ref_frame, main_frame); + } + + for (int plane = 0; plane < s->nb_planes; plane++) { + if (!(s->planes & (1 << plane))) { + av_image_copy_plane(out_frame->data[plane], out_frame->linesize[plane], + main_frame->data[plane], main_frame->linesize[plane], + s->planewidth[plane] * ((s->depth + 7) / 8), s->planeheight[plane]); + continue; + } + if (s->depth <= 8) + guided_byte(ctx, s, main_frame->data[plane], ref_frame->data[plane], out_frame->data[plane], s->radius, s->eps, + s->planewidth[plane], s->planeheight[plane], + main_frame->linesize[plane], ref_frame->linesize[plane], out_frame->linesize[plane], (1 << s->depth) - 1.f); + else + guided_word(ctx, s, main_frame->data[plane], ref_frame->data[plane], out_frame->data[plane], s->radius, s->eps, + s->planewidth[plane], s->planeheight[plane], + main_frame->linesize[plane] / 2, ref_frame->linesize[plane] / 2, out_frame->linesize[plane] / 2, (1 << s->depth) - 1.f); + } + + return ff_filter_frame(outlink, out_frame); +} + + +static int config_output(AVFilterLink *outlink) +{ + AVFilterContext *ctx = outlink->src; + + GuidedContext *s = ctx->priv; + AVFilterLink *mainlink = ctx->inputs[0]; + FFFrameSyncIn *in; + int ret; + + + outlink->w = mainlink->w; + outlink->h = mainlink->h; + outlink->time_base = mainlink->time_base; + outlink->sample_aspect_ratio = mainlink->sample_aspect_ratio; + outlink->frame_rate = mainlink->frame_rate; + if ((ret = ff_framesync_init(&s->fs, ctx, 2)) < 0) + return ret; + + outlink->time_base = s->fs.time_base; + + in = s->fs.in; + in[0].time_base = mainlink->time_base; + in[1].time_base = ctx->inputs[1]->time_base; + in[0].sync = 2; + in[0].before = EXT_INFINITY; + in[0].after = EXT_INFINITY; + in[1].sync = 1; + in[1].before = EXT_INFINITY; + in[1].after = EXT_INFINITY; + s->fs.opaque = s; + s->fs.on_event = process_frame; + + return ff_framesync_configure(&s->fs); +} + +static int activate(AVFilterContext *ctx) +{ + GuidedContext *s = ctx->priv; + return ff_framesync_activate(&s->fs); +} + +static av_cold int init(AVFilterContext *ctx) +{ + return 0; +} + +static av_cold void uninit(AVFilterContext *ctx) +{ + GuidedContext *s = ctx->priv; + ff_framesync_uninit(&s->fs); + return; +} + + +static int process_command(AVFilterContext *ctx, + const char *cmd, + const char *arg, + char *res, + int res_len, + int flags) +{ + int ret = ff_filter_process_command(ctx, cmd, arg, res, res_len, flags); + + if (ret < 0) + return ret; + + return 0; +} + +static const AVFilterPad guided_inputs[] = { + { + .name = "main", + .type = AVMEDIA_TYPE_VIDEO, + },{ + .name = "reference", + .type = AVMEDIA_TYPE_VIDEO, + .config_props = config_input, + }, + { NULL } +}; + +static const AVFilterPad guided_outputs[] = { + { + .name = "default", + .type = AVMEDIA_TYPE_VIDEO, + .config_props = config_output, + }, + { NULL } +}; + +AVFilter ff_vf_guided = { + .name = "guided", + .description = NULL_IF_CONFIG_SMALL("Apply Guided filter."), + .init = init, + .uninit = uninit, + .query_formats = query_formats, + .priv_size = sizeof(GuidedContext), + .priv_class = &guided_class, + .activate = activate, + .inputs = guided_inputs, + .outputs = guided_outputs, + .flags = AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC | AVFILTER_FLAG_SLICE_THREADS, + .process_command = process_command, +};
V3: Add examples on how to use this filter, and improve the code style. V2: Implement the slice-level parallelism for guided filter. V1: Add the basic version of guided filter. Signed-off-by: Xuewei Meng <xwmeng96@gmail.com> --- doc/filters.texi | 38 +++++ libavfilter/Makefile | 1 + libavfilter/allfilters.c | 1 + libavfilter/vf_guided.c | 429 +++++++++++++++++++++++++++++++++++++++++++++++ 4 files changed, 469 insertions(+) create mode 100644 libavfilter/vf_guided.c