[FFmpeg-devel] [PATCH v3] GSoC: Add guided filter
Steven Liu
lq at chinaffmpeg.org
Mon May 10 08:35:44 EEST 2021
> 2021年5月8日 下午6:20,Steven Liu <lq at chinaffmpeg.org> 写道:
>
>
>
>> 2021年5月8日 下午6:06,Xuewei Meng <928826483 at qq.com> 写道:
>>
>> V3: Add examples on how to use this filter, and improve the code style.
>> V2: Implement the slice-level parallelism for guided filter.
>> V1: Add the basic version of guided filter.
>>
>> Signed-off-by: Xuewei Meng <xwmeng96 at gmail.com>
>> ---
>> doc/filters.texi | 38 +++++
>> libavfilter/Makefile | 1 +
>> libavfilter/allfilters.c | 1 +
>> libavfilter/vf_guided.c | 429 +++++++++++++++++++++++++++++++++++++++++++++++
>> 4 files changed, 469 insertions(+)
>> create mode 100644 libavfilter/vf_guided.c
>>
>> diff --git a/doc/filters.texi b/doc/filters.texi
>> index 36e35a1..515c655 100644
>> --- a/doc/filters.texi
>> +++ b/doc/filters.texi
>> @@ -12918,6 +12918,44 @@ greyedge=difford=1:minknorm=0:sigma=2
>>
>> @end itemize
>>
>> + at section guided filter
>> +Apply guided filter for edge-preserving smoothing, dehazing and so on.
>> +
>> +The filter accepts the following options:
>> + at table @option
>> + at item radius
>> +Set the radius in pixels.
>> +Allowed range is 1 to 20. Default is 3.
>> +
>> + at item eps
>> +Set regularization parameter.
>> +Allowed range is 0 to 1. Default is 0.1.
>> +
>> + at item planes
>> +Set planes to filter. Default is first only.
>> + at end table
>> +
>> + at subsection Commands
>> +This filter supports the all above options as @ref{commands}.
>> +
>> + at subsection Examples
>> + at itemize
>> + at item
>> +Edge-preserving smoothing with guided filter:
>> + at example
>> +ffmpeg -i in.png -i in.png -filter_complex guided out.png
>> + at end example
>> +
>> + at item
>> +Dehazing, structure-transferring filtering, detail enhancement with guided filter.
>> +For the generation of guidance image,
>> +see @url{http://kaiminghe.com/publications/pami12guidedfilter.pdf}.
>> + at example
>> +ffmpeg -i in.png -i guidance.png -filter_complex guided out.png
>> + at end example
>> +
>> + at end itemize
>> +
>> @anchor{haldclut}
>> @section haldclut
>>
>> diff --git a/libavfilter/Makefile b/libavfilter/Makefile
>> index 5a28736..60a97e1 100644
>> --- a/libavfilter/Makefile
>> +++ b/libavfilter/Makefile
>> @@ -285,6 +285,7 @@ OBJS-$(CONFIG_GEQ_FILTER) += vf_geq.o
>> OBJS-$(CONFIG_GRADFUN_FILTER) += vf_gradfun.o
>> OBJS-$(CONFIG_GRAPHMONITOR_FILTER) += f_graphmonitor.o
>> OBJS-$(CONFIG_GREYEDGE_FILTER) += vf_colorconstancy.o
>> +OBJS-$(CONFIG_GUIDED_FILTER) += vf_guided.o
>> OBJS-$(CONFIG_HALDCLUT_FILTER) += vf_lut3d.o framesync.o
>> OBJS-$(CONFIG_HFLIP_FILTER) += vf_hflip.o
>> OBJS-$(CONFIG_HISTEQ_FILTER) += vf_histeq.o
>> diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c
>> index 931d7db..962f656 100644
>> --- a/libavfilter/allfilters.c
>> +++ b/libavfilter/allfilters.c
>> @@ -270,6 +270,7 @@ extern const AVFilter ff_vf_geq;
>> extern const AVFilter ff_vf_gradfun;
>> extern const AVFilter ff_vf_graphmonitor;
>> extern const AVFilter ff_vf_greyedge;
>> +extern const AVFilter ff_vf_guided;
>> extern const AVFilter ff_vf_haldclut;
>> extern const AVFilter ff_vf_hflip;
>> extern const AVFilter ff_vf_histeq;
>> diff --git a/libavfilter/vf_guided.c b/libavfilter/vf_guided.c
>> new file mode 100644
>> index 0000000..bd706fb
>> --- /dev/null
>> +++ b/libavfilter/vf_guided.c
>> @@ -0,0 +1,429 @@
>> +/*
>> + * Copyright (c) 2021 Xuewei Meng
>> + *
>> + * This file is part of FFmpeg.
>> + *
>> + * FFmpeg is free software; you can redistribute it and/or
>> + * modify it under the terms of the GNU Lesser General Public
>> + * License as published by the Free Software Foundation; either
>> + * version 2.1 of the License, or (at your option) any later version.
>> + *
>> + * FFmpeg is distributed in the hope that it will be useful,
>> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
>> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
>> + * Lesser General Public License for more details.
>> + *
>> + * You should have received a copy of the GNU Lesser General Public
>> + * License along with FFmpeg; if not, write to the Free Software
>> + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
>> + */
>> +
>> +#include "libavutil/imgutils.h"
>> +#include "libavutil/opt.h"
>> +#include "libavutil/pixdesc.h"
>> +#include "avfilter.h"
>> +#include "formats.h"
>> +#include "framesync.h"
>> +#include "internal.h"
>> +#include "video.h"
>> +
>> +typedef struct GuidedContext {
>> + const AVClass *class;
>> + FFFrameSync fs;
>> +
>> + int radius;
>> + float eps;
>> +
>> + int planes;
>> +
>> + int width;
>> + int height;
>> +
>> + int nb_planes;
>> + int depth;
>> + int planewidth[4];
>> + int planeheight[4];
>> +
>> + int (*box_slice)(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs);
>> +} GuidedContext;
>> +
>> +#define OFFSET(x) offsetof(GuidedContext, x)
>> +#define FLAGS AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_RUNTIME_PARAM
>> +
>> +static const AVOption guided_options[] = {
>> + { "radius", "set the box radius", OFFSET(radius), AV_OPT_TYPE_INT, {.i64=3 }, 1, 20, FLAGS },
>> + { "eps", "set the regularization parameter (with square)", OFFSET(eps), AV_OPT_TYPE_FLOAT, {.dbl=0.01 }, 0.0, 1, FLAGS },
>> + { "planes", "set planes to filter", OFFSET(planes), AV_OPT_TYPE_INT, {.i64=1 }, 0, 0xF, FLAGS },
>> + { NULL }
>> +};
>> +
>> +AVFILTER_DEFINE_CLASS(guided);
>> +
>> +typedef struct ThreadData {
>> + int width;
>> + int height;
>> + float *src;
>> + float *dst;
>> + int srcStride;
>> + int dstStride;
>> +} ThreadData;
>> +
>> +static int box_slice(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
>> +{
>> + GuidedContext *s = ctx->priv;
>> + ThreadData *t = arg;
>> +
>> + const int width = t->width;
>> + const int height = t->height;
>> + const int src_stride = t->srcStride;
>> + const int dst_stride = t->dstStride;
>> + const int slice_start = (height * jobnr) / nb_jobs;
>> + const int slice_end = (height * (jobnr + 1)) / nb_jobs;
>> + const int radius = s->radius;
>> + const float *src = t->src;
>> + float *dst = t->dst;
>> +
>> + int w;
>> + int numPix;
>> + w = (radius << 1) + 1;
>> + numPix = w * w;
>> + for (int i = slice_start;i < slice_end;i++) {
>> + for (int j = 0;j < width;j++) {
>> + float temp = 0.0;
>> + for (int row = -radius;row <= radius;row++) {
>> + for (int col = -radius;col <= radius;col++) {
>> + int x = i + row;
>> + int y = j + col;
>> + x = (x < 0) ? 0 : (x >= height ? height - 1 : x);
>> + y = (y < 0) ? 0 : (y >= width ? width - 1 : y);
>> + temp += src[x * src_stride + y];
>> + }
>> + }
>> + dst[i * dst_stride + j] = temp / numPix;
>> + }
>> + }
>> + return 0;
>> +}
>> +
>> +static int query_formats(AVFilterContext *ctx)
>> +{
>> + static const enum AVPixelFormat pix_fmts[] = {
>> + AV_PIX_FMT_YUVA444P, AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV440P,
>> + AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ440P,
>> + AV_PIX_FMT_YUVA422P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUVA420P, AV_PIX_FMT_YUV420P,
>> + AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUVJ420P,
>> + AV_PIX_FMT_YUVJ411P, AV_PIX_FMT_YUV411P, AV_PIX_FMT_YUV410P,
>> + AV_PIX_FMT_YUV420P9, AV_PIX_FMT_YUV422P9, AV_PIX_FMT_YUV444P9,
>> + AV_PIX_FMT_YUV420P10, AV_PIX_FMT_YUV422P10, AV_PIX_FMT_YUV444P10,
>> + AV_PIX_FMT_YUV420P12, AV_PIX_FMT_YUV422P12, AV_PIX_FMT_YUV444P12, AV_PIX_FMT_YUV440P12,
>> + AV_PIX_FMT_YUV420P14, AV_PIX_FMT_YUV422P14, AV_PIX_FMT_YUV444P14,
>> + AV_PIX_FMT_YUV420P16, AV_PIX_FMT_YUV422P16, AV_PIX_FMT_YUV444P16,
>> + AV_PIX_FMT_YUVA420P9, AV_PIX_FMT_YUVA422P9, AV_PIX_FMT_YUVA444P9,
>> + AV_PIX_FMT_YUVA420P10, AV_PIX_FMT_YUVA422P10, AV_PIX_FMT_YUVA444P10,
>> + AV_PIX_FMT_YUVA420P16, AV_PIX_FMT_YUVA422P16, AV_PIX_FMT_YUVA444P16,
>> + AV_PIX_FMT_GBRP, AV_PIX_FMT_GBRP9, AV_PIX_FMT_GBRP10,
>> + AV_PIX_FMT_GBRP12, AV_PIX_FMT_GBRP14, AV_PIX_FMT_GBRP16,
>> + AV_PIX_FMT_GBRAP, AV_PIX_FMT_GBRAP10, AV_PIX_FMT_GBRAP12, AV_PIX_FMT_GBRAP16,
>> + AV_PIX_FMT_GRAY8, AV_PIX_FMT_GRAY9, AV_PIX_FMT_GRAY10, AV_PIX_FMT_GRAY12, AV_PIX_FMT_GRAY14, AV_PIX_FMT_GRAY16,
>> + AV_PIX_FMT_NONE
>> + };
>> +
>> + return ff_set_common_formats(ctx, ff_make_format_list(pix_fmts));
>> +}
>> +
>> +static int config_input(AVFilterLink *inlink)
>> +{
>> + AVFilterContext *ctx = inlink->dst;
>> + GuidedContext *s = ctx->priv;
>> + const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(inlink->format);
>> +
>> + if (ctx->inputs[0]->w != ctx->inputs[1]->w ||
>> + ctx->inputs[0]->h != ctx->inputs[1]->h) {
>> + av_log(ctx, AV_LOG_ERROR, "Width and height of input videos must be same.\n");
>> + return AVERROR(EINVAL);
>> + }
>> + if (ctx->inputs[0]->format != ctx->inputs[1]->format) {
>> + av_log(ctx, AV_LOG_ERROR, "Inputs must be of same pixel format.\n");
>> + return AVERROR(EINVAL);
>> + }
>> +
>> +
>> + s->depth = desc->comp[0].depth;
>> + s->width = ctx->inputs[0]->w;
>> + s->height = ctx->inputs[0]->h;
>> +
>> + s->planewidth[1] = s->planewidth[2] = AV_CEIL_RSHIFT(inlink->w, desc->log2_chroma_w);
>> + s->planewidth[0] = s->planewidth[3] = inlink->w;
>> + s->planeheight[1] = s->planeheight[2] = AV_CEIL_RSHIFT(inlink->h, desc->log2_chroma_h);
>> + s->planeheight[0] = s->planeheight[3] = inlink->h;
>> +
>> + s->nb_planes = av_pix_fmt_count_planes(inlink->format);
>> + s->box_slice = box_slice;
>> + return 0;
>> +}
>> +
>> +#define GUIDED(type, name) \
>> +static int guided_##name(AVFilterContext *ctx, GuidedContext *s, \
>> + const uint8_t *ssrc, const uint8_t *ssrcRef, \
>> + uint8_t *ddst, int radius, float eps, int width, int height, \
>> + int src_stride, int src_ref_stride, int dst_stride, \
>> + float maxval) \
>> +{ \
>> + int ret = 0; \
>> + type *dst = (type *)ddst; \
>> + const type *src = (const type *)ssrc; \
>> + const type *srcRef = (const type *)ssrcRef; \
>> + \
>> + ThreadData t; \
>> + const int nb_threads = ff_filter_get_nb_threads(ctx); \
>> + float *I; \
>> + float *II; \
>> + float *P; \
>> + float *IP; \
>> + float *meanI; \
>> + float *meanII; \
>> + float *meanP; \
>> + float *meanIP; \
>> + float *A; \
>> + float *B; \
>> + float *meanA; \
>> + float *meanB; \
>> + \
>> + I = av_calloc(width * height, sizeof(float)); \
>> + II = av_calloc(width * height, sizeof(float)); \
>> + P = av_calloc(width * height, sizeof(float)); \
>> + IP = av_calloc(width * height, sizeof(float)); \
>> + meanI = av_calloc(width * height, sizeof(float)); \
>> + meanII = av_calloc(width * height, sizeof(float)); \
>> + meanP = av_calloc(width * height, sizeof(float)); \
>> + meanIP = av_calloc(width * height, sizeof(float)); \
>> + \
>> + A = av_calloc(width * height, sizeof(float)); \
>> + B = av_calloc(width * height, sizeof(float)); \
>> + meanA = av_calloc(width * height, sizeof(float)); \
>> + meanB = av_calloc(width * height, sizeof(float)); \
>> + \
>> + if (!I || !II || !P || !IP || !meanI || !meanII || !meanP || \
>> + !meanIP || !A || !B || !meanA || !meanB){ \
>> + ret = AVERROR(ENOMEM); \
>> + goto end; \
>> + } \
>> + for (int i = 0;i < height;i++) { \
>> + for (int j = 0;j < width;j++) { \
>> + int x = i * width + j; \
>> + I[x] = src[i * src_stride + j] / maxval; \
>> + II[x] = I[x] * I[x]; \
>> + P[x] = srcRef[i * src_ref_stride + j] / maxval; \
>> + IP[x] = I[x] * P[x]; \
>> + } \
>> + } \
>> + \
>> + t.width = width; \
>> + t.height = height; \
>> + t.srcStride = width; \
>> + t.dstStride = width; \
>> + t.src = I; \
>> + t.dst = meanI; \
>> + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \
>> + t.src = II; \
>> + t.dst = meanII; \
>> + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \
>> + t.src = P; \
>> + t.dst = meanP; \
>> + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \
>> + t.src = IP; \
>> + t.dst = meanIP; \
>> + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \
>> + \
>> + for (int i = 0;i < height;i++) { \
>> + for (int j = 0;j < width;j++) { \
>> + int x = i * width + j; \
>> + float varI = meanII[x] - (meanI[x] * meanI[x]); \
>> + float covIP = meanIP[x] - (meanI[x] * meanP[x]); \
>> + A[x] = covIP / (varI + eps); \
>> + B[x] = meanP[x] - A[x] * meanI[x]; \
>> + } \
>> + } \
>> + \
>> + t.src = A; \
>> + t.dst = meanA; \
>> + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \
>> + t.src = B; \
>> + t.dst = meanB; \
>> + ctx->internal->execute(ctx, s->box_slice, &t, NULL, FFMIN(height, nb_threads)); \
>> + \
>> + for (int i = 0;i < height;i++) { \
>> + for (int j = 0;j < width;j++) { \
>> + int x = i * width + j; \
>> + dst[i * dst_stride + j] = meanA[x] * src[i * src_stride + j] + \
>> + meanB[x] * maxval; \
>> + } \
>> + } \
>> +end: \
>> + av_freep(&I); \
>> + av_freep(&II); \
>> + av_freep(&P); \
>> + av_freep(&IP); \
>> + av_freep(&meanI); \
>> + av_freep(&meanII); \
>> + av_freep(&meanP); \
>> + av_freep(&meanIP); \
>> + av_freep(&A); \
>> + av_freep(&B); \
>> + av_freep(&meanA); \
>> + av_freep(&meanB); \
>> + return ret; \
>> +}
>> +
>> +GUIDED(uint8_t, byte)
>> +GUIDED(uint16_t, word)
>> +
>> +static int process_frame(FFFrameSync *fs)
>> +{
>> + AVFilterContext *ctx = fs->parent;
>> + GuidedContext *s = fs->opaque;
>> + AVFilterLink *outlink = ctx->outputs[0];
>> + AVFrame *out_frame = NULL, *main_frame = NULL, *ref_frame = NULL;
>> + int ret;
>> +
>> + ret = ff_framesync_dualinput_get(fs, &main_frame, &ref_frame);
>> + if (ret < 0)
>> + return ret;
>> +
>> + out_frame = ff_get_video_buffer(outlink, outlink->w, outlink->h);
>> + if (!out_frame) {
>> + av_frame_free(&main_frame);
>> + av_frame_free(&ref_frame);
>> + return AVERROR(ENOMEM);
>> + }
>> + av_frame_copy_props(out_frame, main_frame);
>> +
>> + if (ctx->is_disabled || !ref_frame) {
>> + av_frame_copy_props(ref_frame, main_frame);
>> + }
>> +
>> + for (int plane = 0; plane < s->nb_planes; plane++) {
>> + if (!(s->planes & (1 << plane))) {
>> + av_image_copy_plane(out_frame->data[plane], out_frame->linesize[plane],
>> + main_frame->data[plane], main_frame->linesize[plane],
>> + s->planewidth[plane] * ((s->depth + 7) / 8), s->planeheight[plane]);
>> + continue;
>> + }
>> + if (s->depth <= 8)
>> + guided_byte(ctx, s, main_frame->data[plane], ref_frame->data[plane], out_frame->data[plane], s->radius, s->eps,
>> + s->planewidth[plane], s->planeheight[plane],
>> + main_frame->linesize[plane], ref_frame->linesize[plane], out_frame->linesize[plane], (1 << s->depth) - 1.f);
>> + else
>> + guided_word(ctx, s, main_frame->data[plane], ref_frame->data[plane], out_frame->data[plane], s->radius, s->eps,
>> + s->planewidth[plane], s->planeheight[plane],
>> + main_frame->linesize[plane] / 2, ref_frame->linesize[plane] / 2, out_frame->linesize[plane] / 2, (1 << s->depth) - 1.f);
>> + }
>> +
>> + return ff_filter_frame(outlink, out_frame);
>> +}
>> +
>> +
>> +static int config_output(AVFilterLink *outlink)
>> +{
>> + AVFilterContext *ctx = outlink->src;
>> +
>> + GuidedContext *s = ctx->priv;
>> + AVFilterLink *mainlink = ctx->inputs[0];
>> + FFFrameSyncIn *in;
>> + int ret;
>> +
>> +
>> + outlink->w = mainlink->w;
>> + outlink->h = mainlink->h;
>> + outlink->time_base = mainlink->time_base;
>> + outlink->sample_aspect_ratio = mainlink->sample_aspect_ratio;
>> + outlink->frame_rate = mainlink->frame_rate;
>> + if ((ret = ff_framesync_init(&s->fs, ctx, 2)) < 0)
>> + return ret;
>> +
>> + outlink->time_base = s->fs.time_base;
>> +
>> + in = s->fs.in;
>> + in[0].time_base = mainlink->time_base;
>> + in[1].time_base = ctx->inputs[1]->time_base;
>> + in[0].sync = 2;
>> + in[0].before = EXT_INFINITY;
>> + in[0].after = EXT_INFINITY;
>> + in[1].sync = 1;
>> + in[1].before = EXT_INFINITY;
>> + in[1].after = EXT_INFINITY;
>> + s->fs.opaque = s;
>> + s->fs.on_event = process_frame;
>> +
>> + return ff_framesync_configure(&s->fs);
>> +}
>> +
>> +static int activate(AVFilterContext *ctx)
>> +{
>> + GuidedContext *s = ctx->priv;
>> + return ff_framesync_activate(&s->fs);
>> +}
>> +
>> +static av_cold int init(AVFilterContext *ctx)
>> +{
>> + return 0;
>> +}
>> +
>> +static av_cold void uninit(AVFilterContext *ctx)
>> +{
>> + GuidedContext *s = ctx->priv;
>> + ff_framesync_uninit(&s->fs);
>> + return;
>> +}
>> +
>> +
>> +static int process_command(AVFilterContext *ctx,
>> + const char *cmd,
>> + const char *arg,
>> + char *res,
>> + int res_len,
>> + int flags)
>> +{
>> + int ret = ff_filter_process_command(ctx, cmd, arg, res, res_len, flags);
>> +
>> + if (ret < 0)
>> + return ret;
>> +
>> + return 0;
>> +}
>> +
>> +static const AVFilterPad guided_inputs[] = {
>> + {
>> + .name = "main",
>> + .type = AVMEDIA_TYPE_VIDEO,
>> + },{
>> + .name = "reference",
>> + .type = AVMEDIA_TYPE_VIDEO,
>> + .config_props = config_input,
>> + },
>> + { NULL }
>> +};
>> +
>> +static const AVFilterPad guided_outputs[] = {
>> + {
>> + .name = "default",
>> + .type = AVMEDIA_TYPE_VIDEO,
>> + .config_props = config_output,
>> + },
>> + { NULL }
>> +};
>> +
>> +AVFilter ff_vf_guided = {
>> + .name = "guided",
>> + .description = NULL_IF_CONFIG_SMALL("Apply Guided filter."),
>> + .init = init,
>> + .uninit = uninit,
>> + .query_formats = query_formats,
>> + .priv_size = sizeof(GuidedContext),
>> + .priv_class = &guided_class,
>> + .activate = activate,
>> + .inputs = guided_inputs,
>> + .outputs = guided_outputs,
>> + .flags = AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC | AVFILTER_FLAG_SLICE_THREADS,
>> + .process_command = process_command,
>> +};
>> --
>> 1.9.1
>>
>> _______________________________________________
>> ffmpeg-devel mailing list
>> ffmpeg-devel at ffmpeg.org
>> https://ffmpeg.org/mailman/listinfo/ffmpeg-devel
>>
>> To unsubscribe, visit link above, or email
>> ffmpeg-devel-request at ffmpeg.org with subject "unsubscribe".
>>
>
> Looks ok to me, waiting for other reviewers,
Pushed
Thanks
Steven Liu
More information about the ffmpeg-devel
mailing list